<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: ignoreDeletes' option with Delta Live Table streaming source in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/ignoredeletes-option-with-delta-live-table-streaming-source/m-p/19026#M12701</link>
    <description>&lt;P&gt;I haven't heard back, but the response above was copy and pasted from here: &lt;A href="https://docs.databricks.com/delta/delta-streaming.html" alt="https://docs.databricks.com/delta/delta-streaming.html" target="_blank"&gt;Table streaming reads and writes | Databricks on AWS&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;We decided to just move these tables to a true structured stream. We hope that DLT can support simple appends later on.&lt;/P&gt;</description>
    <pubDate>Fri, 10 Jun 2022 20:10:10 GMT</pubDate>
    <dc:creator>Zachary_Higgins</dc:creator>
    <dc:date>2022-06-10T20:10:10Z</dc:date>
    <item>
      <title>ignoreDeletes' option with Delta Live Table streaming source</title>
      <link>https://community.databricks.com/t5/data-engineering/ignoredeletes-option-with-delta-live-table-streaming-source/m-p/19023#M12698</link>
      <description>&lt;P&gt;We have a delta streaming source in our delta live table pipelines that may have data deleted from time to time. &lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;The error message is pretty self explanatory:&lt;/P&gt;&lt;P&gt;&lt;B&gt;...from streaming source at version 191. This is currently not supported. If you'd like to ignore deletes, set the option 'ignoreDeletes' to 'true'.&lt;/B&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;What's not clear is how to set this option. This is what we have now but it's not producing the desired results. The desired result being new data is read and deletes are ignored.&lt;/P&gt;&lt;PRE&gt;&lt;CODE&gt;SET pipelines.ignoreDeletes = true;
CREATE OR REFRESH STREAMING LIVE TABLE...&lt;/CODE&gt;&lt;/PRE&gt;&lt;P&gt;How should this option be set in a delta live table?&lt;/P&gt;</description>
      <pubDate>Wed, 01 Jun 2022 20:56:25 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/ignoredeletes-option-with-delta-live-table-streaming-source/m-p/19023#M12698</guid>
      <dc:creator>Zachary_Higgins</dc:creator>
      <dc:date>2022-06-01T20:56:25Z</dc:date>
    </item>
    <item>
      <title>Re: ignoreDeletes' option with Delta Live Table streaming source</title>
      <link>https://community.databricks.com/t5/data-engineering/ignoredeletes-option-with-delta-live-table-streaming-source/m-p/19025#M12700</link>
      <description>&lt;P&gt;Hi - Thanks for the response. Does your suggestion work with Delta live tables when you try it? This seems to produce the same error message when I use the code below:&lt;/P&gt;&lt;PRE&gt;&lt;CODE&gt;@dlt.table(
   ...
    })
&amp;nbsp;
def table_fnc():
    return spark.readStream.format("delta").option("ignoreDeletes", "true").table("tablename")&lt;/CODE&gt;&lt;/PRE&gt;&lt;P&gt;I'm not worried about duplicates. I just want to stream out the tables current state and append it to a sink in my DLT pipeline. As far as I know, DLT can't just append data from a source unless it's streamed in... &lt;/P&gt;</description>
      <pubDate>Thu, 02 Jun 2022 18:00:58 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/ignoredeletes-option-with-delta-live-table-streaming-source/m-p/19025#M12700</guid>
      <dc:creator>Zachary_Higgins</dc:creator>
      <dc:date>2022-06-02T18:00:58Z</dc:date>
    </item>
    <item>
      <title>Re: ignoreDeletes' option with Delta Live Table streaming source</title>
      <link>https://community.databricks.com/t5/data-engineering/ignoredeletes-option-with-delta-live-table-streaming-source/m-p/19026#M12701</link>
      <description>&lt;P&gt;I haven't heard back, but the response above was copy and pasted from here: &lt;A href="https://docs.databricks.com/delta/delta-streaming.html" alt="https://docs.databricks.com/delta/delta-streaming.html" target="_blank"&gt;Table streaming reads and writes | Databricks on AWS&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;We decided to just move these tables to a true structured stream. We hope that DLT can support simple appends later on.&lt;/P&gt;</description>
      <pubDate>Fri, 10 Jun 2022 20:10:10 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/ignoredeletes-option-with-delta-live-table-streaming-source/m-p/19026#M12701</guid>
      <dc:creator>Zachary_Higgins</dc:creator>
      <dc:date>2022-06-10T20:10:10Z</dc:date>
    </item>
    <item>
      <title>Re: ignoreDeletes' option with Delta Live Table streaming source</title>
      <link>https://community.databricks.com/t5/data-engineering/ignoredeletes-option-with-delta-live-table-streaming-source/m-p/19027#M12702</link>
      <description>&lt;P&gt;@Kaniz Fatma​&amp;nbsp;Hi Kaniz, can we please circle around to this?  Like @Zachary Higgins​&amp;nbsp;, I am unsure how to set the ignoreDeletes or ignoreChanges spark.sql configuration for my Delta Live Table Pipeline defined in SQL.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;</description>
      <pubDate>Mon, 01 Aug 2022 03:00:30 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/ignoredeletes-option-with-delta-live-table-streaming-source/m-p/19027#M12702</guid>
      <dc:creator>JohnA</dc:creator>
      <dc:date>2022-08-01T03:00:30Z</dc:date>
    </item>
    <item>
      <title>Re: ignoreDeletes' option with Delta Live Table streaming source</title>
      <link>https://community.databricks.com/t5/data-engineering/ignoredeletes-option-with-delta-live-table-streaming-source/m-p/19028#M12703</link>
      <description>&lt;P&gt;Databricks, please provide an answer to this. It seems like there is no documentation on how delta live tables support table updates. The ignoreChanges is bound to spark.readstream method which is not made to dlt.read_stream&lt;/P&gt;</description>
      <pubDate>Fri, 11 Nov 2022 12:34:51 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/ignoredeletes-option-with-delta-live-table-streaming-source/m-p/19028#M12703</guid>
      <dc:creator>7effrey</dc:creator>
      <dc:date>2022-11-11T12:34:51Z</dc:date>
    </item>
    <item>
      <title>Re: ignoreDeletes' option with Delta Live Table streaming source</title>
      <link>https://community.databricks.com/t5/data-engineering/ignoredeletes-option-with-delta-live-table-streaming-source/m-p/19029#M12704</link>
      <description>&lt;P&gt;I'd am looking at this as well and would like to understand my options here.&lt;/P&gt;</description>
      <pubDate>Sun, 18 Dec 2022 21:31:37 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/ignoredeletes-option-with-delta-live-table-streaming-source/m-p/19029#M12704</guid>
      <dc:creator>Michael42</dc:creator>
      <dc:date>2022-12-18T21:31:37Z</dc:date>
    </item>
    <item>
      <title>Re: ignoreDeletes' option with Delta Live Table streaming source</title>
      <link>https://community.databricks.com/t5/data-engineering/ignoredeletes-option-with-delta-live-table-streaming-source/m-p/19030#M12705</link>
      <description>&lt;P&gt;@Kaniz Fatma​&amp;nbsp;- Has Databricks found a way to prune unwanted records from a source without requiring the entire sink table be recalculated with DLT?&lt;/P&gt;</description>
      <pubDate>Thu, 02 Feb 2023 19:48:28 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/ignoredeletes-option-with-delta-live-table-streaming-source/m-p/19030#M12705</guid>
      <dc:creator>Zachary_Higgins</dc:creator>
      <dc:date>2023-02-02T19:48:28Z</dc:date>
    </item>
    <item>
      <title>Re: ignoreDeletes' option with Delta Live Table streaming source</title>
      <link>https://community.databricks.com/t5/data-engineering/ignoredeletes-option-with-delta-live-table-streaming-source/m-p/113254#M44482</link>
      <description>&lt;P&gt;We had to delete some records in the destination tables created by the DLT pipeline today and hit this error. What resolved it was using spark.readStream and then those options could be set. Can't set them on `&lt;SPAN&gt;&lt;SPAN class=""&gt;dlt.read_stream`&lt;/SPAN&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;PRE&gt;&lt;SPAN&gt;&lt;SPAN class=""&gt;df&lt;/SPAN&gt; &lt;SPAN class=""&gt;=&lt;/SPAN&gt; &lt;SPAN class=""&gt;dlt&lt;/SPAN&gt;&lt;SPAN class=""&gt;.&lt;/SPAN&gt;&lt;SPAN class=""&gt;read_stream&lt;/SPAN&gt;&lt;SPAN class=""&gt;(&lt;/SPAN&gt;&lt;SPAN class=""&gt;"&lt;/SPAN&gt;&lt;SPAN class=""&gt;request_params&lt;/SPAN&gt;&lt;SPAN class=""&gt;"&lt;/SPAN&gt;&lt;SPAN class=""&gt;)&lt;/SPAN&gt;&lt;/SPAN&gt;
&lt;SPAN&gt;    &lt;SPAN class=""&gt;# df = dlt.read_stream("request_params")&lt;/SPAN&gt;&lt;/SPAN&gt;
&lt;SPAN&gt;    &lt;SPAN class=""&gt;df&lt;/SPAN&gt; &lt;SPAN class=""&gt;=&lt;/SPAN&gt; &lt;SPAN class=""&gt;spark&lt;/SPAN&gt;&lt;SPAN class=""&gt;.&lt;/SPAN&gt;&lt;SPAN class=""&gt;readStream&lt;/SPAN&gt;&lt;SPAN class=""&gt;.&lt;/SPAN&gt;&lt;SPAN class=""&gt;option&lt;/SPAN&gt;&lt;SPAN class=""&gt;(&lt;/SPAN&gt;&lt;SPAN class=""&gt;"&lt;/SPAN&gt;&lt;SPAN class=""&gt;skipChangeCommits&lt;/SPAN&gt;&lt;SPAN class=""&gt;"&lt;/SPAN&gt;&lt;SPAN class=""&gt;,&lt;/SPAN&gt; &lt;SPAN class=""&gt;"&lt;/SPAN&gt;&lt;SPAN class=""&gt;true&lt;/SPAN&gt;&lt;SPAN class=""&gt;"&lt;/SPAN&gt;&lt;SPAN class=""&gt;).&lt;/SPAN&gt;&lt;SPAN class=""&gt;option&lt;/SPAN&gt;&lt;SPAN class=""&gt;(&lt;/SPAN&gt;&lt;SPAN class=""&gt;"&lt;/SPAN&gt;&lt;SPAN class=""&gt;ignoreDeletes&lt;/SPAN&gt;&lt;SPAN class=""&gt;"&lt;/SPAN&gt;&lt;SPAN class=""&gt;,&lt;/SPAN&gt; &lt;SPAN class=""&gt;"&lt;/SPAN&gt;&lt;SPAN class=""&gt;true&lt;/SPAN&gt;&lt;SPAN class=""&gt;"&lt;/SPAN&gt;&lt;SPAN class=""&gt;).&lt;/SPAN&gt;&lt;SPAN class=""&gt;table&lt;/SPAN&gt;&lt;SPAN class=""&gt;(&lt;/SPAN&gt;&lt;SPAN class=""&gt;"&lt;/SPAN&gt;&lt;SPAN class=""&gt;operations.api_logs.request_params&lt;/SPAN&gt;&lt;SPAN class=""&gt;"&lt;/SPAN&gt;&lt;SPAN class=""&gt;)&lt;/SPAN&gt;&lt;/SPAN&gt;
&lt;/PRE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 21 Mar 2025 04:38:03 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/ignoredeletes-option-with-delta-live-table-streaming-source/m-p/113254#M44482</guid>
      <dc:creator>yegorski</dc:creator>
      <dc:date>2025-03-21T04:38:03Z</dc:date>
    </item>
    <item>
      <title>Re: ignoreDeletes' option with Delta Live Table streaming source</title>
      <link>https://community.databricks.com/t5/data-engineering/ignoredeletes-option-with-delta-live-table-streaming-source/m-p/113255#M44483</link>
      <description>&lt;P&gt;&lt;A href="https://docs.databricks.com/aws/en/structured-streaming/delta-lake" target="_blank"&gt;https://docs.databricks.com/aws/en/structured-streaming/delta-lake&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&lt;A href="https://docs.databricks.com/aws/en/dlt/python-ref" target="_blank"&gt;https://docs.databricks.com/aws/en/dlt/python-ref&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 21 Mar 2025 04:39:30 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/ignoredeletes-option-with-delta-live-table-streaming-source/m-p/113255#M44483</guid>
      <dc:creator>yegorski</dc:creator>
      <dc:date>2025-03-21T04:39:30Z</dc:date>
    </item>
    <item>
      <title>Re: ignoreDeletes' option with Delta Live Table streaming source</title>
      <link>https://community.databricks.com/t5/data-engineering/ignoredeletes-option-with-delta-live-table-streaming-source/m-p/122935#M46907</link>
      <description>&lt;P&gt;I had this same issue whilst doing some POC work. I was able to overcome it as follows:&lt;/P&gt;&lt;OL&gt;&lt;LI&gt;&lt;SPAN&gt;Navigate to Workflows | &lt;/SPAN&gt;&lt;SPAN&gt;&lt;SPAN class=""&gt;Jobs &amp;amp; pipelines.&lt;/SPAN&gt;&lt;/SPAN&gt;&lt;/LI&gt;&lt;LI&gt;&lt;SPAN&gt;&lt;SPAN class=""&gt;&lt;SPAN class=""&gt;Select your pipeline&lt;/SPAN&gt;.&lt;/SPAN&gt;&lt;/SPAN&gt;&lt;/LI&gt;&lt;LI&gt;&lt;SPAN&gt;&lt;SPAN class=""&gt;Click the drop-down next to the Start button.&lt;/SPAN&gt;&lt;/SPAN&gt;&lt;/LI&gt;&lt;LI&gt;&lt;SPAN&gt;&lt;SPAN class=""&gt;Choose&amp;nbsp;"Full refresh all".&lt;/SPAN&gt;&lt;/SPAN&gt;&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;&lt;SPAN&gt;&lt;SPAN class=""&gt;That resets it all and fixes the problem caused by deleting rows (in my case, all the rows) from the bronze table (which broke the silver table load in the pipeline, but it's fine after full refresh operation). It might not be these best solution for everyone's needs because it'll force a full reload from scratch and maybe you don't even have all your historic source files/messages available anymore.&lt;/SPAN&gt;&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Thu, 26 Jun 2025 10:59:09 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/ignoredeletes-option-with-delta-live-table-streaming-source/m-p/122935#M46907</guid>
      <dc:creator>IanB_Argento</dc:creator>
      <dc:date>2025-06-26T10:59:09Z</dc:date>
    </item>
  </channel>
</rss>

