<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Delta live tables running count output mode? in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/delta-live-tables-running-count-output-mode/m-p/62038#M31883</link>
    <description>&lt;P&gt;Hi, Kaniz&lt;/P&gt;&lt;P&gt;Could you please more details and example on how to configure the outputmode? From public available document table_properties configuration for DLT,&amp;nbsp; does not have the option for outputMode.&amp;nbsp; I have also found that sometimes the DLT "decided" to use complete mode instead of append mode, which results in downstream workflow error such as: "streaming tables may only use append-only streaming sources".&amp;nbsp; Please clarify. Thank you!&lt;/P&gt;</description>
    <pubDate>Tue, 27 Feb 2024 01:02:09 GMT</pubDate>
    <dc:creator>harvey-c</dc:creator>
    <dc:date>2024-02-27T01:02:09Z</dc:date>
    <item>
      <title>Delta live tables running count output mode?</title>
      <link>https://community.databricks.com/t5/data-engineering/delta-live-tables-running-count-output-mode/m-p/44022#M27592</link>
      <description>&lt;P&gt;I have a DLT with a table that I want to contain the running aggregation (for the sake of simplicitly let's assume it's a count) for each value of some key column, using a session window. The input table goes back several years and to clean up aggregation state, I want to add a watermark. Doing this, however, appears to output no rows.&lt;/P&gt;&lt;P&gt;I believe this is because in the default append output mode, only expired session windows are emitted. Looking at the delta table's history I see appends only. How do I configure update output mode? Or is there another way to achieve my goal?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="python"&gt;def running_aggregation():    
    return (
        spark.readStream
            .option("withEventTimeOrder", "true")
            .table("LIVE.input_data")
            .withWatermark("created", "365 days") # Watermark in combination with append output mode (don't know how to change for DLT) results in only expired session windows being output..
            .groupBy(session_window("created", "90 days"), "key")            
            .agg(
                count('*')
            )&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 07 Sep 2023 16:16:40 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/delta-live-tables-running-count-output-mode/m-p/44022#M27592</guid>
      <dc:creator>vroste</dc:creator>
      <dc:date>2023-09-07T16:16:40Z</dc:date>
    </item>
    <item>
      <title>Re: Delta live tables running count output mode?</title>
      <link>https://community.databricks.com/t5/data-engineering/delta-live-tables-running-count-output-mode/m-p/62038#M31883</link>
      <description>&lt;P&gt;Hi, Kaniz&lt;/P&gt;&lt;P&gt;Could you please more details and example on how to configure the outputmode? From public available document table_properties configuration for DLT,&amp;nbsp; does not have the option for outputMode.&amp;nbsp; I have also found that sometimes the DLT "decided" to use complete mode instead of append mode, which results in downstream workflow error such as: "streaming tables may only use append-only streaming sources".&amp;nbsp; Please clarify. Thank you!&lt;/P&gt;</description>
      <pubDate>Tue, 27 Feb 2024 01:02:09 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/delta-live-tables-running-count-output-mode/m-p/62038#M31883</guid>
      <dc:creator>harvey-c</dc:creator>
      <dc:date>2024-02-27T01:02:09Z</dc:date>
    </item>
  </channel>
</rss>

