<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Displaying spark job process in dashboard in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/displaying-spark-job-process-in-dashboard/m-p/22421#M15346</link>
    <description>&lt;P&gt;In databricks is there a way to display the spark job process in a dashboard? I have a simple dashboard that displays a table, but the main spark job behind it takes 15 minutes to run. Is there a way to show the spark job progress bar in a dashboard?&lt;/P&gt;</description>
    <pubDate>Fri, 18 Jun 2021 20:09:30 GMT</pubDate>
    <dc:creator>User16765131552</dc:creator>
    <dc:date>2021-06-18T20:09:30Z</dc:date>
    <item>
      <title>Displaying spark job process in dashboard</title>
      <link>https://community.databricks.com/t5/data-engineering/displaying-spark-job-process-in-dashboard/m-p/22421#M15346</link>
      <description>&lt;P&gt;In databricks is there a way to display the spark job process in a dashboard? I have a simple dashboard that displays a table, but the main spark job behind it takes 15 minutes to run. Is there a way to show the spark job progress bar in a dashboard?&lt;/P&gt;</description>
      <pubDate>Fri, 18 Jun 2021 20:09:30 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/displaying-spark-job-process-in-dashboard/m-p/22421#M15346</guid>
      <dc:creator>User16765131552</dc:creator>
      <dc:date>2021-06-18T20:09:30Z</dc:date>
    </item>
    <item>
      <title>Re: Displaying spark job process in dashboard</title>
      <link>https://community.databricks.com/t5/data-engineering/displaying-spark-job-process-in-dashboard/m-p/22422#M15347</link>
      <description>&lt;P&gt;The best way to do so would be to collect data about the job run using the &lt;A href="https://docs.databricks.com/dev-tools/api/latest/jobs.html#runs-get" alt="https://docs.databricks.com/dev-tools/api/latest/jobs.html#runs-get" target="_blank"&gt;REST API&lt;/A&gt; (runs get endpoint). This endpoint provides as much metadata as possible. You may need to use other endpoints to get the job or run ids in order to get the correct information. &lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Anyway you can pull the data about the job while it was running. Save the data to a Delta table and make the visualization you would like. It tracks the start time and end time of a job while you can do the difference between the current and start time to get a progress bar. &lt;/P&gt;</description>
      <pubDate>Fri, 18 Jun 2021 20:41:41 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/displaying-spark-job-process-in-dashboard/m-p/22422#M15347</guid>
      <dc:creator>Ryan_Chynoweth</dc:creator>
      <dc:date>2021-06-18T20:41:41Z</dc:date>
    </item>
  </channel>
</rss>

