<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Slow running Spark job issue - due to the unknown spark stages created by Databircks Compute cluster in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/slow-running-spark-job-issue-due-to-the-unknown-spark-stages/m-p/121118#M46342</link>
    <description>&lt;P&gt;Hi Team,&lt;/P&gt;&lt;P&gt;Recently we migrated the spark jobs from self hosted spark(YARN) Cluster to Databricks.&lt;/P&gt;&lt;P&gt;Currently we are using the Databricks workflows with Job_Compute clusters and the Job Type - Spark JAR type execution, so when we run the job in databricks, what we obsererved is its creating the extra job stages like mentioned in the below image. the problem here which is also taking a singnficant time which is causing the delaying of the total job runtime.&amp;nbsp;&lt;BR /&gt;Databricks Run time: 16.1&lt;BR /&gt;instance type -&amp;nbsp;&lt;SPAN class=""&gt;&lt;SPAN&gt;Standard_E16ds_v4&lt;/SPAN&gt;&lt;/SPAN&gt;&lt;BR /&gt;Can you please add your suggestions.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="databricks_new_stages.png" style="width: 400px;"&gt;&lt;img src="https://community.databricks.com/t5/image/serverpage/image-id/17386i11DE724E11328A53/image-size/medium?v=v2&amp;amp;px=400" role="button" title="databricks_new_stages.png" alt="databricks_new_stages.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt; &lt;/P&gt;</description>
    <pubDate>Fri, 06 Jun 2025 05:13:38 GMT</pubDate>
    <dc:creator>anil_reddaboina</dc:creator>
    <dc:date>2025-06-06T05:13:38Z</dc:date>
    <item>
      <title>Slow running Spark job issue - due to the unknown spark stages created by Databircks Compute cluster</title>
      <link>https://community.databricks.com/t5/data-engineering/slow-running-spark-job-issue-due-to-the-unknown-spark-stages/m-p/121118#M46342</link>
      <description>&lt;P&gt;Hi Team,&lt;/P&gt;&lt;P&gt;Recently we migrated the spark jobs from self hosted spark(YARN) Cluster to Databricks.&lt;/P&gt;&lt;P&gt;Currently we are using the Databricks workflows with Job_Compute clusters and the Job Type - Spark JAR type execution, so when we run the job in databricks, what we obsererved is its creating the extra job stages like mentioned in the below image. the problem here which is also taking a singnficant time which is causing the delaying of the total job runtime.&amp;nbsp;&lt;BR /&gt;Databricks Run time: 16.1&lt;BR /&gt;instance type -&amp;nbsp;&lt;SPAN class=""&gt;&lt;SPAN&gt;Standard_E16ds_v4&lt;/SPAN&gt;&lt;/SPAN&gt;&lt;BR /&gt;Can you please add your suggestions.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="databricks_new_stages.png" style="width: 400px;"&gt;&lt;img src="https://community.databricks.com/t5/image/serverpage/image-id/17386i11DE724E11328A53/image-size/medium?v=v2&amp;amp;px=400" role="button" title="databricks_new_stages.png" alt="databricks_new_stages.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt; &lt;/P&gt;</description>
      <pubDate>Fri, 06 Jun 2025 05:13:38 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/slow-running-spark-job-issue-due-to-the-unknown-spark-stages/m-p/121118#M46342</guid>
      <dc:creator>anil_reddaboina</dc:creator>
      <dc:date>2025-06-06T05:13:38Z</dc:date>
    </item>
    <item>
      <title>Re: Slow running Spark job issue - due to the unknown spark stages created by Databircks Compute clu</title>
      <link>https://community.databricks.com/t5/data-engineering/slow-running-spark-job-issue-due-to-the-unknown-spark-stages/m-p/121131#M46346</link>
      <description>&lt;P&gt;Hi Anil,&lt;/P&gt;&lt;P&gt;How are you doing today?, As per my understanding, When you move Spark jobs from a self-hosted YARN cluster to Databricks and run them using Spark JARs on job compute clusters, it's normal to see a few extra stages added in the job execution plan. These stages are usually due to Databricks’ built-in features like adaptive query execution (AQE), automatic optimizations, or internal tracking. While these help in performance tuning, they can sometimes increase the total runtime if not tuned well. I’d suggest trying to disable AQE temporarily (spark.sql.adaptive.enabled to false) and reviewing the job stages in the Spark UI to see what’s taking time. Also, double-check if broadcast joins or data skew might be causing shuffle delays. Using compute pools can also reduce cold-start delays if you're launching new clusters for each run. A bit of tuning here can make a big difference — happy to help further if you share a specific job plan or logs!&lt;/P&gt;&lt;P&gt;Regards,&lt;/P&gt;&lt;P&gt;Brahma&lt;/P&gt;</description>
      <pubDate>Fri, 06 Jun 2025 11:24:30 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/slow-running-spark-job-issue-due-to-the-unknown-spark-stages/m-p/121131#M46346</guid>
      <dc:creator>Brahmareddy</dc:creator>
      <dc:date>2025-06-06T11:24:30Z</dc:date>
    </item>
    <item>
      <title>Re: Slow running Spark job issue - due to the unknown spark stages created by Databircks Compute clu</title>
      <link>https://community.databricks.com/t5/data-engineering/slow-running-spark-job-issue-due-to-the-unknown-spark-stages/m-p/121136#M46348</link>
      <description>&lt;P&gt;Hey Brahma,&lt;BR /&gt;Thanks for your reply. As a first step I will disable AQE config and test it.&amp;nbsp;&lt;BR /&gt;&lt;BR /&gt;We are using the node pools with job_compute cluster type so that its not spinning up a new cluster for each Job.&amp;nbsp;&lt;/P&gt;&lt;P&gt;I'm configuring the below two configs also, do you think these configs cause any side effects&amp;nbsp;&lt;/P&gt;&lt;DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;"spark.databricks.io.cache.enabled"&lt;/SPAN&gt;&lt;SPAN&gt;: &lt;/SPAN&gt;&lt;SPAN&gt;"true"&lt;/SPAN&gt;&lt;SPAN&gt;,&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;"spark.databricks.io.cache.maxDiskUsage"&lt;/SPAN&gt;&lt;SPAN&gt;: &lt;/SPAN&gt;&lt;SPAN&gt;"50g"&lt;/SPAN&gt;&lt;SPAN&gt;,&lt;/SPAN&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks,&lt;/P&gt;&lt;P&gt;Anil&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 06 Jun 2025 12:33:43 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/slow-running-spark-job-issue-due-to-the-unknown-spark-stages/m-p/121136#M46348</guid>
      <dc:creator>anil_reddaboina</dc:creator>
      <dc:date>2025-06-06T12:33:43Z</dc:date>
    </item>
  </channel>
</rss>

