<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Lakeflow pipeline (formerly DLT pipeline) performance progressively degrades on a persistent cluster in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/lakeflow-pipeline-formerly-dlt-pipeline-performance/m-p/126862#M47787</link>
    <description>&lt;P&gt;I have a small (under 20 tables, all streaming) DLT pipeline running in triggered mode, scheduled every 15min during the workday.&amp;nbsp; For development I've set `pipelines.clusterShutdown.delay` to avoid having to start a cluster every update.&lt;/P&gt;&lt;P&gt;I've noticed that the updates' runtimes are progressively worse as the time goes on, ultimately doubling in time after only 2h. It increases progressively even after updates on where there are no updates to any of the tables; each table's update duration is individually low but the overall runtime is high. Eventually we have to let the compute shut down to restart and regain performance.&lt;/P&gt;&lt;P&gt;Cluster metrics show nothing out of ordinary; even though free memory slowly decreases over time there's still enough, and CPU load is way below its limit even at its peak. There's nothing obviously wrong in the logs either.&lt;/P&gt;&lt;P&gt;I'm assuming restarting the cluster periodically is expected somehow, but what if it were a continuous pipeline instead where it would stay up until manually shut down, wouldn't this issue be more prominent?&lt;/P&gt;&lt;P&gt;Is there a way to mitigate this without restarting the cluster several times a day?&lt;/P&gt;</description>
    <pubDate>Tue, 29 Jul 2025 20:57:36 GMT</pubDate>
    <dc:creator>rcostanza</dc:creator>
    <dc:date>2025-07-29T20:57:36Z</dc:date>
    <item>
      <title>Lakeflow pipeline (formerly DLT pipeline) performance progressively degrades on a persistent cluster</title>
      <link>https://community.databricks.com/t5/data-engineering/lakeflow-pipeline-formerly-dlt-pipeline-performance/m-p/126862#M47787</link>
      <description>&lt;P&gt;I have a small (under 20 tables, all streaming) DLT pipeline running in triggered mode, scheduled every 15min during the workday.&amp;nbsp; For development I've set `pipelines.clusterShutdown.delay` to avoid having to start a cluster every update.&lt;/P&gt;&lt;P&gt;I've noticed that the updates' runtimes are progressively worse as the time goes on, ultimately doubling in time after only 2h. It increases progressively even after updates on where there are no updates to any of the tables; each table's update duration is individually low but the overall runtime is high. Eventually we have to let the compute shut down to restart and regain performance.&lt;/P&gt;&lt;P&gt;Cluster metrics show nothing out of ordinary; even though free memory slowly decreases over time there's still enough, and CPU load is way below its limit even at its peak. There's nothing obviously wrong in the logs either.&lt;/P&gt;&lt;P&gt;I'm assuming restarting the cluster periodically is expected somehow, but what if it were a continuous pipeline instead where it would stay up until manually shut down, wouldn't this issue be more prominent?&lt;/P&gt;&lt;P&gt;Is there a way to mitigate this without restarting the cluster several times a day?&lt;/P&gt;</description>
      <pubDate>Tue, 29 Jul 2025 20:57:36 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/lakeflow-pipeline-formerly-dlt-pipeline-performance/m-p/126862#M47787</guid>
      <dc:creator>rcostanza</dc:creator>
      <dc:date>2025-07-29T20:57:36Z</dc:date>
    </item>
    <item>
      <title>Re: Lakeflow pipeline (formerly DLT pipeline) performance progressively degrades on a persistent clu</title>
      <link>https://community.databricks.com/t5/data-engineering/lakeflow-pipeline-formerly-dlt-pipeline-performance/m-p/126877#M47793</link>
      <description>&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;P&gt;You're right to be concerned — this sounds like a classic case of memory or resource leakage over time, which can affect long-running jobs even if metrics look okay on the surface. In triggered DLT (now Lakeflow) pipelines, tasksand state can accumulate in memory, especially with streaming workloads. For continuous pipelines, this degradation would likely be worse. While a restart is the simplest fix, you can mitigate this by optimizing stateful operations (like joins and aggregations), enabling state cleanup settings, and ensuring checkpoint locations aren't bloating. Also consider using autoscaling clusters with auto-shutdown enabled between runs to reset state without manual restarts.&lt;/P&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&amp;nbsp;&lt;/DIV&gt;&lt;/DIV&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;SPAN class=""&gt;&lt;SPAN class=""&gt;Ask Cha&lt;/SPAN&gt;&lt;/SPAN&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;</description>
      <pubDate>Wed, 30 Jul 2025 04:15:42 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/lakeflow-pipeline-formerly-dlt-pipeline-performance/m-p/126877#M47793</guid>
      <dc:creator>jerrygen78</dc:creator>
      <dc:date>2025-07-30T04:15:42Z</dc:date>
    </item>
    <item>
      <title>Re: Lakeflow pipeline (formerly DLT pipeline) performance progressively degrades on a persistent clu</title>
      <link>https://community.databricks.com/t5/data-engineering/lakeflow-pipeline-formerly-dlt-pipeline-performance/m-p/146443#M52655</link>
      <description>&lt;P&gt;I'm facing this exact issue, only with a standard job instead of a DLT pipeline. I can't use serverless or restart the cluster periodically due to things out of my control. Any specific advice on diagnosis and resolving? I don't think it can be checkpoint bloat since cluster restart solves the issue for a time.&lt;/P&gt;</description>
      <pubDate>Mon, 02 Feb 2026 07:09:39 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/lakeflow-pipeline-formerly-dlt-pipeline-performance/m-p/146443#M52655</guid>
      <dc:creator>JargerBiirli</dc:creator>
      <dc:date>2026-02-02T07:09:39Z</dc:date>
    </item>
  </channel>
</rss>

