<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Solution Design Recommendation on Databricks in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/solution-design-recommendation-on-databricks/m-p/131975#M49306</link>
    <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/184483"&gt;@tyhatwar785&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;1. Should metadata and file download be separate jobs/notebooks or combined?&lt;BR /&gt;Keep them in separate notebooks but orchestrate them under a single Databricks Job.&lt;BR /&gt;for better error handling, and retries .&lt;/P&gt;&lt;P&gt;2. Cluster recommendations&lt;BR /&gt;start with a general-purpose cluster( Standard_DS4_v2 (28 GB memory, 8 vCPU) ) with autoscaling enabled&lt;/P&gt;&lt;P&gt;3. Parallelism&lt;BR /&gt;If all processing is inside Databricks&lt;/P&gt;&lt;P&gt;4. Best practices&lt;/P&gt;&lt;P&gt;Retries: Use Databricks Job-level retries and add custom retry logic using UDF&lt;/P&gt;&lt;P&gt;Error Handling: Use Python’s try/except with structured logging (logging library) for better observability.&lt;/P&gt;&lt;P&gt;Monitoring: Integrate with Databricks Lakehouse Monitoring or send metrics/logs&lt;/P&gt;</description>
    <pubDate>Mon, 15 Sep 2025 12:20:08 GMT</pubDate>
    <dc:creator>nikhilmohod-nm</dc:creator>
    <dc:date>2025-09-15T12:20:08Z</dc:date>
    <item>
      <title>Solution Design Recommendation on Databricks</title>
      <link>https://community.databricks.com/t5/data-engineering/solution-design-recommendation-on-databricks/m-p/131935#M49295</link>
      <description>&lt;P&gt;Hi Team,&lt;/P&gt;&lt;P&gt;We need to design a pipeline in Databricks to:&lt;/P&gt;&lt;P&gt;1. Call a metadata API (returns XML per keyword), parse, and consolidate into a combined JSON.&lt;/P&gt;&lt;P&gt;2. Use this metadata to generate dynamic links for a second API, download ZIPs, unzip, and extract specific HTML files into ADLS.&lt;/P&gt;&lt;P&gt;Looking for suggestions on: Solution design – should metadata and file download be separate jobs/notebooks or combined?&lt;/P&gt;&lt;P&gt;Cluster recommendations – what type/size of cluster is suitable for this workload?&lt;/P&gt;&lt;P&gt;Parallelism – should we use Python async (aiohttp) or Spark parallelism for faster execution?&lt;/P&gt;&lt;P&gt;Best practices – retries, error handling, checkpointing for flaky APIs. Would appreciate guidance on how to design this efficiently.&lt;/P&gt;&lt;P&gt;Thanks!&lt;/P&gt;</description>
      <pubDate>Mon, 15 Sep 2025 07:12:24 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/solution-design-recommendation-on-databricks/m-p/131935#M49295</guid>
      <dc:creator>tyhatwar785</dc:creator>
      <dc:date>2025-09-15T07:12:24Z</dc:date>
    </item>
    <item>
      <title>Re: Solution Design Recommendation on Databricks</title>
      <link>https://community.databricks.com/t5/data-engineering/solution-design-recommendation-on-databricks/m-p/131975#M49306</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/184483"&gt;@tyhatwar785&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;1. Should metadata and file download be separate jobs/notebooks or combined?&lt;BR /&gt;Keep them in separate notebooks but orchestrate them under a single Databricks Job.&lt;BR /&gt;for better error handling, and retries .&lt;/P&gt;&lt;P&gt;2. Cluster recommendations&lt;BR /&gt;start with a general-purpose cluster( Standard_DS4_v2 (28 GB memory, 8 vCPU) ) with autoscaling enabled&lt;/P&gt;&lt;P&gt;3. Parallelism&lt;BR /&gt;If all processing is inside Databricks&lt;/P&gt;&lt;P&gt;4. Best practices&lt;/P&gt;&lt;P&gt;Retries: Use Databricks Job-level retries and add custom retry logic using UDF&lt;/P&gt;&lt;P&gt;Error Handling: Use Python’s try/except with structured logging (logging library) for better observability.&lt;/P&gt;&lt;P&gt;Monitoring: Integrate with Databricks Lakehouse Monitoring or send metrics/logs&lt;/P&gt;</description>
      <pubDate>Mon, 15 Sep 2025 12:20:08 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/solution-design-recommendation-on-databricks/m-p/131975#M49306</guid>
      <dc:creator>nikhilmohod-nm</dc:creator>
      <dc:date>2025-09-15T12:20:08Z</dc:date>
    </item>
  </channel>
</rss>

