<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic How to identify the goal of a specific Spark job? in Get Started Discussions</title>
    <link>https://community.databricks.com/t5/get-started-discussions/how-to-identify-the-goal-of-a-specific-spark-job/m-p/108105#M9704</link>
    <description>&lt;P&gt;I'm analyzing the performance of a DBR/Spark request. In this case, the cluster is created using a custom image, and then we run a job on it.&lt;/P&gt;&lt;P&gt;I've dived into the "Spark UI" part of the DBR interface, and identified 3 jobs that appear to account for an outsized amount of execution time: `write at WriteIntoDeltaCommand.scala:85`, `collect at GenerateSymlinkManifest.scala:295`, and `execute at DeltaOptimizedWriterExec.scala:130`. While the UI lets me dig into more detail, it doesn't seem to specify anywhere what the *purpose* of each job is. Is there anywhere I can look to find out why Spark decided it needed to execute this particular 3 jobs?&lt;/P&gt;</description>
    <pubDate>Fri, 31 Jan 2025 15:33:54 GMT</pubDate>
    <dc:creator>mrstevegross</dc:creator>
    <dc:date>2025-01-31T15:33:54Z</dc:date>
    <item>
      <title>How to identify the goal of a specific Spark job?</title>
      <link>https://community.databricks.com/t5/get-started-discussions/how-to-identify-the-goal-of-a-specific-spark-job/m-p/108105#M9704</link>
      <description>&lt;P&gt;I'm analyzing the performance of a DBR/Spark request. In this case, the cluster is created using a custom image, and then we run a job on it.&lt;/P&gt;&lt;P&gt;I've dived into the "Spark UI" part of the DBR interface, and identified 3 jobs that appear to account for an outsized amount of execution time: `write at WriteIntoDeltaCommand.scala:85`, `collect at GenerateSymlinkManifest.scala:295`, and `execute at DeltaOptimizedWriterExec.scala:130`. While the UI lets me dig into more detail, it doesn't seem to specify anywhere what the *purpose* of each job is. Is there anywhere I can look to find out why Spark decided it needed to execute this particular 3 jobs?&lt;/P&gt;</description>
      <pubDate>Fri, 31 Jan 2025 15:33:54 GMT</pubDate>
      <guid>https://community.databricks.com/t5/get-started-discussions/how-to-identify-the-goal-of-a-specific-spark-job/m-p/108105#M9704</guid>
      <dc:creator>mrstevegross</dc:creator>
      <dc:date>2025-01-31T15:33:54Z</dc:date>
    </item>
    <item>
      <title>Re: How to identify the goal of a specific Spark job?</title>
      <link>https://community.databricks.com/t5/get-started-discussions/how-to-identify-the-goal-of-a-specific-spark-job/m-p/108171#M9705</link>
      <description>&lt;P&gt;The spark jobs are decided based on your spark code. You can look at the spark plan to understand what operations each spark job/stage is executing&lt;/P&gt;</description>
      <pubDate>Fri, 31 Jan 2025 18:37:25 GMT</pubDate>
      <guid>https://community.databricks.com/t5/get-started-discussions/how-to-identify-the-goal-of-a-specific-spark-job/m-p/108171#M9705</guid>
      <dc:creator>Lakshay</dc:creator>
      <dc:date>2025-01-31T18:37:25Z</dc:date>
    </item>
  </channel>
</rss>

