<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Dynamic Jobs community Edition in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/dynamic-jobs-community-edition/m-p/134031#M49992</link>
    <description>&lt;P&gt;Hi&amp;nbsp;&lt;A href="https://community.databricks.com/t5/user/viewprofilepage/user-id/188930" target="_blank"&gt;@Ritesh-Dhumne&lt;/A&gt;&amp;nbsp;,&lt;/P&gt;&lt;P&gt;I'm assuming that you mistakenly named Free Edition as Community since you're using volumes.&lt;/P&gt;&lt;P&gt;I’m not sure if I’ve understood your approach correctly, but at first glance it seems incorrect - you can’t pass a DataFrame between tasks. What you can do is load all the files from the volume into a bronze table in Notebook1. You can use the special&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;&lt;STRONG&gt;_metadata&lt;/STRONG&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;column to add information about the file_path from which each particular row originates. Here’s an example of how to use it:&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="szymon_dybczak_0-1759826052216.png" style="width: 400px;"&gt;&lt;img src="https://community.databricks.com/t5/image/serverpage/image-id/20520i2A416FCE8F8F729C/image-size/medium?v=v2&amp;amp;px=400" role="button" title="szymon_dybczak_0-1759826052216.png" alt="szymon_dybczak_0-1759826052216.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Then, in Notebook2, you can apply your transformations based on this bronze table. You can count nulls, handle dirty data, and benefit from the fact that you can relate all these issues to a particular file, since this information is added to the bronze table through the _metadata special column.&lt;/P&gt;&lt;P&gt;From what I see you're in a learning process so I won't introduce the concept of autoloader which is pretty handy for ingestion of files &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt;&lt;/P&gt;</description>
    <pubDate>Tue, 07 Oct 2025 08:36:02 GMT</pubDate>
    <dc:creator>szymon_dybczak</dc:creator>
    <dc:date>2025-10-07T08:36:02Z</dc:date>
    <item>
      <title>Dynamic Jobs community Edition</title>
      <link>https://community.databricks.com/t5/data-engineering/dynamic-jobs-community-edition/m-p/133900#M49955</link>
      <description>&lt;P&gt;Hello&amp;nbsp;&amp;nbsp;I tried this,&lt;/P&gt;&lt;P&gt;Notebook 1 :&lt;/P&gt;&lt;P&gt;dbutils.jobs.taskValues.set(key = "my_key", value = "hi From Notebook1")&lt;/P&gt;&lt;P&gt;Notebook2:&lt;/P&gt;&lt;P&gt;X = dbutils.jobs.taskValues.get(taskKey="01", key="my_key", debugValue = "Fail")&lt;/P&gt;&lt;P&gt;print(X)&lt;/P&gt;&lt;P&gt;&amp;nbsp;Here I get "Fail" as output, its not fetching my_key&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Mon, 06 Oct 2025 07:08:52 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/dynamic-jobs-community-edition/m-p/133900#M49955</guid>
      <dc:creator>Ritesh-Dhumne</dc:creator>
      <dc:date>2025-10-06T07:08:52Z</dc:date>
    </item>
    <item>
      <title>Re: Dynamic Jobs community Edition</title>
      <link>https://community.databricks.com/t5/data-engineering/dynamic-jobs-community-edition/m-p/133904#M49958</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/188930"&gt;@Ritesh-Dhumne&lt;/a&gt;&amp;nbsp;&lt;SPAN&gt;f you run both notebooks manually or outside a Databricks job with multiple tasks,&amp;nbsp;&lt;/SPAN&gt;taskValues&lt;SPAN&gt;&amp;nbsp;will not work as expected. you should d&lt;/SPAN&gt;efine your job with multiple tasks in Databricks Workflows.&lt;/P&gt;</description>
      <pubDate>Mon, 06 Oct 2025 07:35:58 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/dynamic-jobs-community-edition/m-p/133904#M49958</guid>
      <dc:creator>saurabh18cs</dc:creator>
      <dc:date>2025-10-06T07:35:58Z</dc:date>
    </item>
    <item>
      <title>Re: Dynamic Jobs community Edition</title>
      <link>https://community.databricks.com/t5/data-engineering/dynamic-jobs-community-edition/m-p/133918#M49961</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/188930"&gt;@Ritesh-Dhumne&lt;/a&gt;&amp;nbsp;,&lt;/P&gt;&lt;P&gt;Folllow my steps. I created 2 notebooks:&lt;/P&gt;&lt;P&gt;- first one called Notebook1 with followign content&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="szymon_dybczak_0-1759737959023.png" style="width: 400px;"&gt;&lt;img src="https://community.databricks.com/t5/image/serverpage/image-id/20483i1002D982D73C74B6/image-size/medium?v=v2&amp;amp;px=400" role="button" title="szymon_dybczak_0-1759737959023.png" alt="szymon_dybczak_0-1759737959023.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;- second one called Notebook2 with following content that will read value defined in Notebook1&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="szymon_dybczak_1-1759738000891.png" style="width: 400px;"&gt;&lt;img src="https://community.databricks.com/t5/image/serverpage/image-id/20484i4ED4B4D3E6455691/image-size/medium?v=v2&amp;amp;px=400" role="button" title="szymon_dybczak_1-1759738000891.png" alt="szymon_dybczak_1-1759738000891.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Here's my definiton of workflow that is using those 2 notebooks:&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="szymon_dybczak_2-1759738044356.png" style="width: 400px;"&gt;&lt;img src="https://community.databricks.com/t5/image/serverpage/image-id/20485i6A98D2B6F4FEBED0/image-size/medium?v=v2&amp;amp;px=400" role="button" title="szymon_dybczak_2-1759738044356.png" alt="szymon_dybczak_2-1759738044356.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;Pay attention to my taskKey in get method is named the same as the task in a workflow (Notebook1):&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="szymon_dybczak_4-1759738096570.png" style="width: 400px;"&gt;&lt;img src="https://community.databricks.com/t5/image/serverpage/image-id/20487i68520CCD52B30FEC/image-size/medium?v=v2&amp;amp;px=400" role="button" title="szymon_dybczak_4-1759738096570.png" alt="szymon_dybczak_4-1759738096570.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="szymon_dybczak_3-1759738090797.png" style="width: 400px;"&gt;&lt;img src="https://community.databricks.com/t5/image/serverpage/image-id/20486iD07A6BA96305EC5A/image-size/medium?v=v2&amp;amp;px=400" role="button" title="szymon_dybczak_3-1759738090797.png" alt="szymon_dybczak_3-1759738090797.png" /&gt;&lt;/span&gt;&lt;BR /&gt;&lt;BR /&gt;When I ran workflow it works as expected:&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="szymon_dybczak_5-1759738194034.png" style="width: 400px;"&gt;&lt;img src="https://community.databricks.com/t5/image/serverpage/image-id/20488i1EA93D1A8163E436/image-size/medium?v=v2&amp;amp;px=400" role="button" title="szymon_dybczak_5-1759738194034.png" alt="szymon_dybczak_5-1759738194034.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Mon, 06 Oct 2025 08:09:57 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/dynamic-jobs-community-edition/m-p/133918#M49961</guid>
      <dc:creator>szymon_dybczak</dc:creator>
      <dc:date>2025-10-06T08:09:57Z</dc:date>
    </item>
    <item>
      <title>Re: Dynamic Jobs community Edition</title>
      <link>https://community.databricks.com/t5/data-engineering/dynamic-jobs-community-edition/m-p/133937#M49965</link>
      <description>&lt;P&gt;Could u Provide me the code and flow.&lt;/P&gt;</description>
      <pubDate>Mon, 06 Oct 2025 09:32:30 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/dynamic-jobs-community-edition/m-p/133937#M49965</guid>
      <dc:creator>Ritesh-Dhumne</dc:creator>
      <dc:date>2025-10-06T09:32:30Z</dc:date>
    </item>
    <item>
      <title>Re: Dynamic Jobs community Edition</title>
      <link>https://community.databricks.com/t5/data-engineering/dynamic-jobs-community-edition/m-p/133943#M49968</link>
      <description>&lt;P&gt;hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/110502"&gt;@szymon_dybczak&lt;/a&gt;&amp;nbsp;, Could u help me in a scenario like I'm trying to build a pipeline where notebook1 captures the file name and format of the file in the catalog . Notebook2 will take the filename and format from notebook1 and perform Basic transformations.&lt;/P&gt;</description>
      <pubDate>Mon, 06 Oct 2025 10:09:17 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/dynamic-jobs-community-edition/m-p/133943#M49968</guid>
      <dc:creator>Ritesh-Dhumne</dc:creator>
      <dc:date>2025-10-06T10:09:17Z</dc:date>
    </item>
    <item>
      <title>Re: Dynamic Jobs community Edition</title>
      <link>https://community.databricks.com/t5/data-engineering/dynamic-jobs-community-edition/m-p/133946#M49970</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/188930"&gt;@Ritesh-Dhumne&lt;/a&gt;&amp;nbsp;,&lt;/P&gt;&lt;P&gt;Sure, but could you describe what you need help with? What's the problem? &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt;&lt;/P&gt;</description>
      <pubDate>Mon, 06 Oct 2025 10:59:14 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/dynamic-jobs-community-edition/m-p/133946#M49970</guid>
      <dc:creator>szymon_dybczak</dc:creator>
      <dc:date>2025-10-06T10:59:14Z</dc:date>
    </item>
    <item>
      <title>Re: Dynamic Jobs community Edition</title>
      <link>https://community.databricks.com/t5/data-engineering/dynamic-jobs-community-edition/m-p/134005#M49985</link>
      <description>&lt;P&gt;I wanted to extract all files in the volume I have uploaded , in notebook 1 and then in notebook 2 perform basic transformation , also I want to store the null , dirty records seperately and a clean dataframe seperately for all the files .In Community Edition&lt;/P&gt;</description>
      <pubDate>Tue, 07 Oct 2025 05:02:56 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/dynamic-jobs-community-edition/m-p/134005#M49985</guid>
      <dc:creator>Ritesh-Dhumne</dc:creator>
      <dc:date>2025-10-07T05:02:56Z</dc:date>
    </item>
    <item>
      <title>Re: Dynamic Jobs community Edition</title>
      <link>https://community.databricks.com/t5/data-engineering/dynamic-jobs-community-edition/m-p/134031#M49992</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;A href="https://community.databricks.com/t5/user/viewprofilepage/user-id/188930" target="_blank"&gt;@Ritesh-Dhumne&lt;/A&gt;&amp;nbsp;,&lt;/P&gt;&lt;P&gt;I'm assuming that you mistakenly named Free Edition as Community since you're using volumes.&lt;/P&gt;&lt;P&gt;I’m not sure if I’ve understood your approach correctly, but at first glance it seems incorrect - you can’t pass a DataFrame between tasks. What you can do is load all the files from the volume into a bronze table in Notebook1. You can use the special&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;&lt;STRONG&gt;_metadata&lt;/STRONG&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;column to add information about the file_path from which each particular row originates. Here’s an example of how to use it:&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="szymon_dybczak_0-1759826052216.png" style="width: 400px;"&gt;&lt;img src="https://community.databricks.com/t5/image/serverpage/image-id/20520i2A416FCE8F8F729C/image-size/medium?v=v2&amp;amp;px=400" role="button" title="szymon_dybczak_0-1759826052216.png" alt="szymon_dybczak_0-1759826052216.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Then, in Notebook2, you can apply your transformations based on this bronze table. You can count nulls, handle dirty data, and benefit from the fact that you can relate all these issues to a particular file, since this information is added to the bronze table through the _metadata special column.&lt;/P&gt;&lt;P&gt;From what I see you're in a learning process so I won't introduce the concept of autoloader which is pretty handy for ingestion of files &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 07 Oct 2025 08:36:02 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/dynamic-jobs-community-edition/m-p/134031#M49992</guid>
      <dc:creator>szymon_dybczak</dc:creator>
      <dc:date>2025-10-07T08:36:02Z</dc:date>
    </item>
  </channel>
</rss>

