<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Using a cluster of type SINGLE_USER to run parallel python tasks in one job in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/using-a-cluster-of-type-single-user-to-run-parallel-python-tasks/m-p/139973#M51324</link>
    <description>&lt;P&gt;Not sure to understand completely but If you are running parallel tasks with each task executed in a given notebook with same variable names, answer is no. The scope of those variables is kind of the spark session or notebook, not the cluster.&lt;/P&gt;&lt;P&gt;To share "data" at cluster level you can use Cluster-Scoped Environment Variables,&amp;nbsp;Global Temp Views, Databricks Secrets for confidential data or even Shared files.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Fri, 21 Nov 2025 21:00:33 GMT</pubDate>
    <dc:creator>Coffee77</dc:creator>
    <dc:date>2025-11-21T21:00:33Z</dc:date>
    <item>
      <title>Using a cluster of type SINGLE_USER to run parallel python tasks in one job</title>
      <link>https://community.databricks.com/t5/data-engineering/using-a-cluster-of-type-single-user-to-run-parallel-python-tasks/m-p/139950#M51322</link>
      <description>&lt;P&gt;Hi,&amp;nbsp;&lt;/P&gt;&lt;P&gt;I have set up a job of multiple spark python tasks running in parallel. I have only set up one job cluster, single node, data security mode SINGLE_USER, using Databricks Runtime version&amp;nbsp;&lt;SPAN&gt;14.3.x-scala2.12. &lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;These parallel spark python tasks share some similar variable names, but they are not technically global variables, everything is defined under one main function per file.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Will the python tasks somehow share these variables since I am using the same clusters? Can this ever happen using Databricks cluster?&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 21 Nov 2025 17:28:37 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/using-a-cluster-of-type-single-user-to-run-parallel-python-tasks/m-p/139950#M51322</guid>
      <dc:creator>oye</dc:creator>
      <dc:date>2025-11-21T17:28:37Z</dc:date>
    </item>
    <item>
      <title>Re: Using a cluster of type SINGLE_USER to run parallel python tasks in one job</title>
      <link>https://community.databricks.com/t5/data-engineering/using-a-cluster-of-type-single-user-to-run-parallel-python-tasks/m-p/139973#M51324</link>
      <description>&lt;P&gt;Not sure to understand completely but If you are running parallel tasks with each task executed in a given notebook with same variable names, answer is no. The scope of those variables is kind of the spark session or notebook, not the cluster.&lt;/P&gt;&lt;P&gt;To share "data" at cluster level you can use Cluster-Scoped Environment Variables,&amp;nbsp;Global Temp Views, Databricks Secrets for confidential data or even Shared files.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 21 Nov 2025 21:00:33 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/using-a-cluster-of-type-single-user-to-run-parallel-python-tasks/m-p/139973#M51324</guid>
      <dc:creator>Coffee77</dc:creator>
      <dc:date>2025-11-21T21:00:33Z</dc:date>
    </item>
    <item>
      <title>Re: Using a cluster of type SINGLE_USER to run parallel python tasks in one job</title>
      <link>https://community.databricks.com/t5/data-engineering/using-a-cluster-of-type-single-user-to-run-parallel-python-tasks/m-p/140061#M51339</link>
      <description>&lt;P&gt;Hi thanks for replying!&lt;/P&gt;&lt;P&gt;In my case, it would be running parallel tasks of type spark python tasks in a lakeflow job. This is a screenshot of the setup:&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="oye_0-1763974540051.png" style="width: 400px;"&gt;&lt;img src="https://community.databricks.com/t5/image/serverpage/image-id/21921iC3C0BB0DE84CFB7F/image-size/medium?v=v2&amp;amp;px=400" role="button" title="oye_0-1763974540051.png" alt="oye_0-1763974540051.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;Aside from the fact that the tasks will share the same resource and thus might run slower, I wonder if there could be any other problem from cluster sharing.&lt;/P&gt;&lt;P&gt;But going from what you said, then there should not be any problem for my setup.&lt;/P&gt;</description>
      <pubDate>Mon, 24 Nov 2025 08:57:56 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/using-a-cluster-of-type-single-user-to-run-parallel-python-tasks/m-p/140061#M51339</guid>
      <dc:creator>oye</dc:creator>
      <dc:date>2025-11-24T08:57:56Z</dc:date>
    </item>
    <item>
      <title>Re: Using a cluster of type SINGLE_USER to run parallel python tasks in one job</title>
      <link>https://community.databricks.com/t5/data-engineering/using-a-cluster-of-type-single-user-to-run-parallel-python-tasks/m-p/140069#M51342</link>
      <description>&lt;P&gt;In my case, we've some jobs configured in a similar way and not issues so far. We are indeed leveraging usage of global temp views at cluster level to improve performance &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt;&lt;/P&gt;</description>
      <pubDate>Mon, 24 Nov 2025 09:25:21 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/using-a-cluster-of-type-single-user-to-run-parallel-python-tasks/m-p/140069#M51342</guid>
      <dc:creator>Coffee77</dc:creator>
      <dc:date>2025-11-24T09:25:21Z</dc:date>
    </item>
    <item>
      <title>Re: Using a cluster of type SINGLE_USER to run parallel python tasks in one job</title>
      <link>https://community.databricks.com/t5/data-engineering/using-a-cluster-of-type-single-user-to-run-parallel-python-tasks/m-p/140155#M51347</link>
      <description>&lt;P&gt;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/152073"&gt;@oye&lt;/a&gt;&amp;nbsp;- The variables scope is local to the individual task and do no interfere with other tasks even if the underlying cluster is same. In fact, the issue is normally other way round where if we have to share the variable across tasks - Then the solutions mentioned by&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/179536"&gt;@Coffee77&lt;/a&gt;&amp;nbsp;- Global Temp view or Cluster-scoped env vars.&lt;/P&gt;</description>
      <pubDate>Mon, 24 Nov 2025 11:49:56 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/using-a-cluster-of-type-single-user-to-run-parallel-python-tasks/m-p/140155#M51347</guid>
      <dc:creator>Raman_Unifeye</dc:creator>
      <dc:date>2025-11-24T11:49:56Z</dc:date>
    </item>
  </channel>
</rss>

