<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Jobs in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/jobs/m-p/146887#M52720</link>
    <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/203176"&gt;@ramsai&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;do you need this =&amp;gt; Event log (from cluster details)&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="saurabh18cs_0-1770289795055.png" style="width: 400px;"&gt;&lt;img src="https://community.databricks.com/t5/image/serverpage/image-id/23648iE3854140F72AFDFD/image-size/medium?v=v2&amp;amp;px=400" role="button" title="saurabh18cs_0-1770289795055.png" alt="saurabh18cs_0-1770289795055.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Thu, 05 Feb 2026 11:10:28 GMT</pubDate>
    <dc:creator>saurabh18cs</dc:creator>
    <dc:date>2026-02-05T11:10:28Z</dc:date>
    <item>
      <title>Jobs</title>
      <link>https://community.databricks.com/t5/data-engineering/jobs/m-p/146824#M52707</link>
      <description>&lt;P&gt;&lt;EM&gt;Is there a way to find out how many workers or cores are being utilized in a job cluster? If so, could you please explain how to check this?&lt;/EM&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 04 Feb 2026 15:47:48 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/jobs/m-p/146824#M52707</guid>
      <dc:creator>ramsai</dc:creator>
      <dc:date>2026-02-04T15:47:48Z</dc:date>
    </item>
    <item>
      <title>Re: Jobs</title>
      <link>https://community.databricks.com/t5/data-engineering/jobs/m-p/146887#M52720</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/203176"&gt;@ramsai&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;do you need this =&amp;gt; Event log (from cluster details)&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="saurabh18cs_0-1770289795055.png" style="width: 400px;"&gt;&lt;img src="https://community.databricks.com/t5/image/serverpage/image-id/23648iE3854140F72AFDFD/image-size/medium?v=v2&amp;amp;px=400" role="button" title="saurabh18cs_0-1770289795055.png" alt="saurabh18cs_0-1770289795055.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 05 Feb 2026 11:10:28 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/jobs/m-p/146887#M52720</guid>
      <dc:creator>saurabh18cs</dc:creator>
      <dc:date>2026-02-05T11:10:28Z</dc:date>
    </item>
    <item>
      <title>Re: Jobs</title>
      <link>https://community.databricks.com/t5/data-engineering/jobs/m-p/150105#M53242</link>
      <description>&lt;P&gt;Hi &lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/203176"&gt;@ramsai&lt;/a&gt;,&lt;/P&gt;
&lt;P&gt;Great question! There are several ways to check how many workers and cores are being utilized in a Databricks job cluster. I will walk through each option from simplest to most advanced.&lt;/P&gt;
&lt;P&gt;&lt;BR /&gt;OPTION 1: CLUSTER METRICS TAB (QUICKEST WAY)&lt;/P&gt;
&lt;P&gt;While your job is running (or after it completes, for up to 30 days), you can view resource utilization directly in the UI:&lt;/P&gt;
&lt;P&gt;1. Go to your job run and click on the task that ran.&lt;BR /&gt;2. Click the cluster link to open the cluster details page.&lt;BR /&gt;3. Select the "Metrics" tab.&lt;/P&gt;
&lt;P&gt;Here you will see near-real-time charts (data collected every minute) including:&lt;/P&gt;
&lt;P&gt;- CPU utilization broken down by mode (user, system, idle, iowait)&lt;BR /&gt;- Memory usage (used, free, buffer, cached)&lt;BR /&gt;- Network bytes sent and received&lt;BR /&gt;- Filesystem space&lt;/P&gt;
&lt;P&gt;You can also switch to "Spark Metrics" from the dropdown to see:&lt;/P&gt;
&lt;P&gt;- Active tasks (which tells you how many cores are actively doing work)&lt;BR /&gt;- Task completion and failure rates&lt;BR /&gt;- Shuffle read/write bytes&lt;/P&gt;
&lt;P&gt;To drill into individual workers, use the "All nodes" dropdown to inspect each node separately. This is helpful for spotting if one worker is overloaded while others are idle.&lt;/P&gt;
&lt;P&gt;Documentation: &lt;A href="https://docs.databricks.com/aws/en/compute/cluster-metrics" target="_blank"&gt;https://docs.databricks.com/aws/en/compute/cluster-metrics&lt;/A&gt;&lt;/P&gt;
&lt;P&gt;&lt;BR /&gt;OPTION 2: SPARK UI (EXECUTOR-LEVEL DETAIL)&lt;/P&gt;
&lt;P&gt;From the cluster details page, click the "Spark UI" tab. This gives you the standard Apache Spark web UI where you can:&lt;/P&gt;
&lt;P&gt;- See the Executors tab, which shows each worker (executor) and its core count, memory, active tasks, and GC time&lt;BR /&gt;- Check the Stages tab to see how tasks are distributed across cores&lt;BR /&gt;- Identify data skew or uneven core usage across workers&lt;/P&gt;
&lt;P&gt;This is the most granular way to see exactly how many cores each worker has and whether they are all being utilized.&lt;/P&gt;
&lt;P&gt;Documentation: &lt;A href="https://docs.databricks.com/aws/en/compute/clusters-manage" target="_blank"&gt;https://docs.databricks.com/aws/en/compute/clusters-manage&lt;/A&gt;&lt;/P&gt;
&lt;P&gt;&lt;BR /&gt;OPTION 3: CLUSTER DETAILS PAGE (WORKER COUNT)&lt;/P&gt;
&lt;P&gt;On the compute details page for your running cluster, Databricks shows the number of currently allocated workers. If you have autoscaling enabled, you can compare the allocated count against your configured min and max to see if the cluster has scaled up or down.&lt;/P&gt;
&lt;P&gt;Documentation: &lt;A href="https://docs.databricks.com/aws/en/compute/configure" target="_blank"&gt;https://docs.databricks.com/aws/en/compute/configure&lt;/A&gt;&lt;/P&gt;
&lt;P&gt;&lt;BR /&gt;OPTION 4: SYSTEM TABLES (PROGRAMMATIC / HISTORICAL ANALYSIS)&lt;/P&gt;
&lt;P&gt;For a programmatic and historical approach, Databricks provides system tables that let you query utilization data using SQL. This is especially useful for analyzing past job runs.&lt;/P&gt;
&lt;P&gt;Key tables:&lt;/P&gt;
&lt;P&gt;1. system.compute.node_timeline - Minute-by-minute CPU, memory, network, and disk metrics per node. Columns include cpu_user_percent, cpu_system_percent, cpu_wait_percent, and mem_used_percent.&lt;/P&gt;
&lt;P&gt;2. system.compute.clusters - Configuration info including worker_count, min_autoscale_workers, max_autoscale_workers, driver_node_type, and worker_node_type.&lt;/P&gt;
&lt;P&gt;3. system.compute.node_types - Hardware specs per instance type, including core_count (number of vCPUs) and memory_mb.&lt;/P&gt;
&lt;P&gt;Example query to find total cores and CPU utilization for a job cluster:&lt;/P&gt;
&lt;P&gt;SELECT&lt;BR /&gt;c.cluster_name,&lt;BR /&gt;c.worker_count,&lt;BR /&gt;nt.core_count AS cores_per_worker,&lt;BR /&gt;c.worker_count * nt.core_count AS total_worker_cores,&lt;BR /&gt;AVG(n.cpu_user_percent + n.cpu_system_percent) AS avg_cpu_utilization&lt;BR /&gt;FROM system.compute.clusters c&lt;BR /&gt;JOIN system.compute.node_types nt&lt;BR /&gt;ON c.worker_node_type = nt.node_type&lt;BR /&gt;LEFT JOIN system.compute.node_timeline n&lt;BR /&gt;ON c.cluster_id = n.cluster_id&lt;BR /&gt;WHERE c.cluster_name = '&amp;lt;your_job_cluster_name&amp;gt;'&lt;BR /&gt;GROUP BY c.cluster_name, c.worker_count, nt.core_count&lt;/P&gt;
&lt;P&gt;Note: These system tables only include records for all-purpose and job clusters (not serverless or SQL warehouses). Nodes running less than 10 minutes may not appear.&lt;/P&gt;
&lt;P&gt;Documentation: &lt;A href="https://docs.databricks.com/en/admin/system-tables/compute.html" target="_blank"&gt;https://docs.databricks.com/en/admin/system-tables/compute.html&lt;/A&gt;&lt;/P&gt;
&lt;P&gt;&lt;BR /&gt;QUICK SUMMARY&lt;/P&gt;
&lt;P&gt;- For a quick visual check while a job runs: use the Metrics tab on the cluster details page.&lt;BR /&gt;- For detailed per-executor core usage: use the Spark UI Executors tab.&lt;BR /&gt;- For historical or programmatic analysis: query the system.compute tables.&lt;BR /&gt;- To find the core count for your instance type: query system.compute.node_types or check your cloud provider docs.&lt;/P&gt;
&lt;P&gt;Hope this helps! Let me know if you have any follow-up questions.&lt;/P&gt;
&lt;P&gt;* This reply used an agent system I built to research and draft this response based on the wide set of documentation I have available and previous memory. I personally review the draft for any obvious issues and for monitoring system reliability and update it when I detect any drift, but there is still a small chance that something is inaccurate, especially if you are experimenting with brand new features.&lt;/P&gt;</description>
      <pubDate>Sun, 08 Mar 2026 02:16:54 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/jobs/m-p/150105#M53242</guid>
      <dc:creator>SteveOstrowski</dc:creator>
      <dc:date>2026-03-08T02:16:54Z</dc:date>
    </item>
  </channel>
</rss>

