<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: batch process autoloader in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/batch-process-autoloader/m-p/124411#M47179</link>
    <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/65591"&gt;@seefoods&lt;/a&gt;&amp;nbsp;,&lt;BR /&gt;If it works, you can mark my answer as a solution so that if someone has the same problem, it will be easier to find an answer.&lt;/P&gt;</description>
    <pubDate>Tue, 08 Jul 2025 08:07:29 GMT</pubDate>
    <dc:creator>MariuszK</dc:creator>
    <dc:date>2025-07-08T08:07:29Z</dc:date>
    <item>
      <title>batch process autoloader</title>
      <link>https://community.databricks.com/t5/data-engineering/batch-process-autoloader/m-p/123869#M47076</link>
      <description>&lt;P&gt;My job continue to running after is finished susccessfully this i my case, i enable useNotification&amp;nbsp;&lt;BR /&gt;&lt;BR /&gt;&lt;/P&gt;&lt;DIV&gt;&lt;PRE&gt;&lt;SPAN&gt;if &lt;/SPAN&gt;&lt;SPAN&gt;self&lt;/SPAN&gt;.autoloader_config.use_autoloader:&lt;BR /&gt;    logger_file_ingestion.info(&lt;SPAN&gt;"debut d'ecriture en mode streaming"&lt;/SPAN&gt;)&lt;BR /&gt;&lt;BR /&gt;    &lt;SPAN&gt;if &lt;/SPAN&gt;&lt;SPAN&gt;self&lt;/SPAN&gt;.write_mode.value.lower() == &lt;SPAN&gt;"append"&lt;/SPAN&gt;:&lt;BR /&gt;        logger_file_ingestion.info(&lt;SPAN&gt;"ecriture en mode %s"&lt;/SPAN&gt;, &lt;SPAN&gt;self&lt;/SPAN&gt;.write_mode.value)&lt;BR /&gt;&lt;BR /&gt;        &lt;SPAN&gt;# Création de la configuration de base du stream&lt;BR /&gt;&lt;/SPAN&gt;        stream_writer = (df.writeStream&lt;BR /&gt;                         .format(&lt;SPAN&gt;"delta"&lt;/SPAN&gt;)&lt;BR /&gt;                         .outputMode(&lt;SPAN&gt;"append"&lt;/SPAN&gt;)&lt;BR /&gt;                         .option(&lt;SPAN&gt;"checkpointLocation"&lt;/SPAN&gt;, &lt;SPAN&gt;self&lt;/SPAN&gt;.checkpoint_location)&lt;BR /&gt;                         .option(&lt;SPAN&gt;"mergeSchema"&lt;/SPAN&gt;, &lt;SPAN&gt;"true"&lt;/SPAN&gt;)&lt;BR /&gt;                         .trigger(&lt;SPAN&gt;availableNow&lt;/SPAN&gt;=&lt;SPAN&gt;True&lt;/SPAN&gt;))&lt;BR /&gt;&lt;BR /&gt;        &lt;SPAN&gt;# Ajout des partitions si nécessaire&lt;BR /&gt;&lt;/SPAN&gt;        &lt;SPAN&gt;if &lt;/SPAN&gt;(&lt;SPAN&gt;self&lt;/SPAN&gt;.source_name.lower() == &lt;SPAN&gt;"name"&lt;/SPAN&gt;) &lt;SPAN&gt;and &lt;/SPAN&gt;(&lt;SPAN&gt;self&lt;/SPAN&gt;.file_format.lower() == &lt;SPAN&gt;"parquet"&lt;/SPAN&gt;&lt;span class="lia-unicode-emoji" title=":disappointed_face:"&gt;😞&lt;/span&gt;&lt;BR /&gt;            stream_writer = stream_writer.partitionBy(&lt;SPAN&gt;"year"&lt;/SPAN&gt;, &lt;SPAN&gt;"day"&lt;/SPAN&gt;, &lt;SPAN&gt;"month"&lt;/SPAN&gt;)&lt;BR /&gt;        &lt;SPAN&gt;elif &lt;/SPAN&gt;(&lt;SPAN&gt;self&lt;/SPAN&gt;.source_name.lower() == &lt;SPAN&gt;"test"&lt;/SPAN&gt;) &lt;SPAN&gt;and &lt;/SPAN&gt;(&lt;SPAN&gt;self&lt;/SPAN&gt;.file_format.lower() == &lt;SPAN&gt;"parquet"&lt;/SPAN&gt;&lt;span class="lia-unicode-emoji" title=":disappointed_face:"&gt;😞&lt;/span&gt;&lt;BR /&gt;            stream_writer = stream_writer.partitionBy(&lt;SPAN&gt;"day"&lt;/SPAN&gt;, &lt;SPAN&gt;"month"&lt;/SPAN&gt;, &lt;SPAN&gt;"year&lt;/SPAN&gt;&lt;SPAN&gt;"&lt;/SPAN&gt;)&lt;BR /&gt;&lt;BR /&gt;        &lt;SPAN&gt;# Lancement du stream et capture de la référence&lt;BR /&gt;&lt;/SPAN&gt;        stream_writer.toTable("bronze")&lt;BR /&gt;       &lt;/PRE&gt;&lt;/DIV&gt;</description>
      <pubDate>Thu, 03 Jul 2025 12:03:04 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/batch-process-autoloader/m-p/123869#M47076</guid>
      <dc:creator>seefoods</dc:creator>
      <dc:date>2025-07-03T12:03:04Z</dc:date>
    </item>
    <item>
      <title>Re: batch process autoloader</title>
      <link>https://community.databricks.com/t5/data-engineering/batch-process-autoloader/m-p/123991#M47106</link>
      <description>&lt;P&gt;Did you try use&amp;nbsp;.awaitTermination() on stream writing?&lt;/P&gt;</description>
      <pubDate>Fri, 04 Jul 2025 09:18:59 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/batch-process-autoloader/m-p/123991#M47106</guid>
      <dc:creator>MariuszK</dc:creator>
      <dc:date>2025-07-04T09:18:59Z</dc:date>
    </item>
    <item>
      <title>Re: batch process autoloader</title>
      <link>https://community.databricks.com/t5/data-engineering/batch-process-autoloader/m-p/124401#M47177</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/131065"&gt;@MariuszK&lt;/a&gt;,&amp;nbsp;&lt;BR /&gt;Thanks it works now. i have use&amp;nbsp;&lt;SPAN&gt;awaitTermination()&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 08 Jul 2025 07:09:08 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/batch-process-autoloader/m-p/124401#M47177</guid>
      <dc:creator>seefoods</dc:creator>
      <dc:date>2025-07-08T07:09:08Z</dc:date>
    </item>
    <item>
      <title>Re: batch process autoloader</title>
      <link>https://community.databricks.com/t5/data-engineering/batch-process-autoloader/m-p/124411#M47179</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/65591"&gt;@seefoods&lt;/a&gt;&amp;nbsp;,&lt;BR /&gt;If it works, you can mark my answer as a solution so that if someone has the same problem, it will be easier to find an answer.&lt;/P&gt;</description>
      <pubDate>Tue, 08 Jul 2025 08:07:29 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/batch-process-autoloader/m-p/124411#M47179</guid>
      <dc:creator>MariuszK</dc:creator>
      <dc:date>2025-07-08T08:07:29Z</dc:date>
    </item>
  </channel>
</rss>

