<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Delay rows coming into DLT pipeline in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/delay-rows-coming-into-dlt-pipeline/m-p/71036#M34219</link>
    <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/78121"&gt;@Mathias&lt;/a&gt;,&amp;nbsp;&lt;/P&gt;
&lt;P&gt;I'd say that watermarking might be a good solution for your use case. Please check&amp;nbsp;&lt;A href="https://docs.databricks.com/en/structured-streaming/watermarks.html#control-late-data-threshold-with-multiple-watermark-policy-in-structured-streaming" target="_self"&gt;Control late data threshold with multiple watermark policy in Structured Streaming.&amp;nbsp;&lt;/A&gt;&lt;/P&gt;
&lt;P&gt;If you want to dig-in further there's also: &lt;A href="https://spark.apache.org/docs/latest/structured-streaming-programming-guide.html#handling-late-data-and-watermarking" target="_self"&gt;Spark Structured Streaming Programming Guide - Handling Late Data and Watermarking.&lt;/A&gt;&lt;/P&gt;
&lt;P&gt;There are other ways to achieve what you're aiming for, I think it's more of a design decision.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Wed, 29 May 2024 15:51:31 GMT</pubDate>
    <dc:creator>raphaelblg</dc:creator>
    <dc:date>2024-05-29T15:51:31Z</dc:date>
    <item>
      <title>Delay rows coming into DLT pipeline</title>
      <link>https://community.databricks.com/t5/data-engineering/delay-rows-coming-into-dlt-pipeline/m-p/71015#M34211</link>
      <description>&lt;P&gt;&lt;BR /&gt;Backgroundand requirements: We are reading data from our factory and storing it in a DLT table called telemetry with columns sensorid, timestamp and value. We need to get rows where sensorid is “qrreader-x” and join with some other data from that same table and store elsewhere. The qr code’s are coming in with very low latency, much less than some of the sensors they should be joined with. There is need for a delay to wait for the other data coming in before processing these rows.&lt;/P&gt;&lt;P&gt;Suggestion: Can we create a DLT pipeline that would be run as batch every 5 minutes and only read rows that have timestamp older than x minutes?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/97035"&gt;@Dlt&lt;/a&gt;.table(
comment="A table summarizing counts of the top baby names for New York for 2021."
)
def top_baby_names_2021():
return (
dlt.read("baby_names_prepared")
.filter(expr("Year_Of_Birth == 2021"))
)&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;When looking at the above example code, I assume the pipeline would consider all rows coming in as handled, even though they are filtered out. So is there a way to put the filtered rows back into the pipeline if they are too fresh?&lt;/P&gt;</description>
      <pubDate>Wed, 29 May 2024 12:01:30 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/delay-rows-coming-into-dlt-pipeline/m-p/71015#M34211</guid>
      <dc:creator>Mathias</dc:creator>
      <dc:date>2024-05-29T12:01:30Z</dc:date>
    </item>
    <item>
      <title>Re: Delay rows coming into DLT pipeline</title>
      <link>https://community.databricks.com/t5/data-engineering/delay-rows-coming-into-dlt-pipeline/m-p/71036#M34219</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/78121"&gt;@Mathias&lt;/a&gt;,&amp;nbsp;&lt;/P&gt;
&lt;P&gt;I'd say that watermarking might be a good solution for your use case. Please check&amp;nbsp;&lt;A href="https://docs.databricks.com/en/structured-streaming/watermarks.html#control-late-data-threshold-with-multiple-watermark-policy-in-structured-streaming" target="_self"&gt;Control late data threshold with multiple watermark policy in Structured Streaming.&amp;nbsp;&lt;/A&gt;&lt;/P&gt;
&lt;P&gt;If you want to dig-in further there's also: &lt;A href="https://spark.apache.org/docs/latest/structured-streaming-programming-guide.html#handling-late-data-and-watermarking" target="_self"&gt;Spark Structured Streaming Programming Guide - Handling Late Data and Watermarking.&lt;/A&gt;&lt;/P&gt;
&lt;P&gt;There are other ways to achieve what you're aiming for, I think it's more of a design decision.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 29 May 2024 15:51:31 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/delay-rows-coming-into-dlt-pipeline/m-p/71036#M34219</guid>
      <dc:creator>raphaelblg</dc:creator>
      <dc:date>2024-05-29T15:51:31Z</dc:date>
    </item>
  </channel>
</rss>

