<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Lakeflow Declarative Pipeline queue in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/lakeflow-declarative-pipeline-queue/m-p/156103#M54356</link>
    <description>&lt;P&gt;&lt;STRONG&gt;What’s going on&lt;/STRONG&gt;&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;The January 2026 release note is correct that the &lt;EM&gt;engine&lt;/EM&gt; for Lakeflow Spark Declarative Pipelines supports &lt;STRONG&gt;queued execution&lt;/STRONG&gt;, but it’s a &lt;STRONG&gt;backend behavior&lt;/STRONG&gt;, not a user-configurable option; there is no UI or DAB field to flip it on or off.&lt;/LI&gt;
&lt;LI&gt;Internally, there has been a “queuing guardrail” and staged rollout work (e.g. discussion on enabling queued update execution and reverting that guardrail only now), so some environments/pipelines still behave as “fail on concurrent StartUpdate” instead of queueing.&lt;/LI&gt;
&lt;LI&gt;Separately, there are &lt;STRONG&gt;open/very recent bugs&lt;/STRONG&gt; where Jobs sends duplicate &lt;CODE&gt;StartUpdate&lt;/CODE&gt; requests or loses the response, causing &lt;CODE&gt;RUN_EXECUTION_ERROR: Pipeline update already in progress&lt;/CODE&gt; even when only one update is actually running; see ES-1635313 and follow-ons, and Nokia BL-16616 / SUP-27441 where the pipeline completes but the job task shows this error.&lt;/LI&gt;
&lt;LI&gt;The &lt;STRONG&gt;Jobs queue&lt;/STRONG&gt; setting (&lt;CODE&gt;queue.enabled: true&lt;/CODE&gt;) you tried in DAB is a &lt;EM&gt;Jobs-level run queue&lt;/EM&gt;, not the DLT/Lakeflow pipeline queuing feature, so toggling that won’t change the pipeline control-plane behavior.&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;Below are &lt;STRONG&gt;3 options&lt;/STRONG&gt;&lt;/P&gt;
&lt;HR /&gt;
&lt;H3&gt;Option 1 – Serialize all updates via one Job (no secondary triggers)&lt;/H3&gt;
&lt;P&gt;Mitigate by ensuring there is &lt;STRONG&gt;exactly one&lt;/STRONG&gt; place that can trigger the pipeline and that it never overlaps runs:&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;Use a &lt;STRONG&gt;single Lakeflow Job&lt;/STRONG&gt; with one pipeline task, and set either:
&lt;UL&gt;
&lt;LI&gt;&lt;CODE&gt;max_concurrent_runs = 1&lt;/CODE&gt;, or&lt;/LI&gt;
&lt;LI&gt;in DAB, on that same job:
&lt;PRE&gt;&lt;CODE class="language-yaml"&gt;jobs:
  my_pipeline_job:
    queue:
      enabled: true
&lt;/CODE&gt;&lt;/PRE&gt;
&lt;/LI&gt;
&lt;/UL&gt;
&lt;/LI&gt;
&lt;LI&gt;Remove/avoid: manual “Start” in the pipeline UI, other Jobs, or ad-hoc API calls that can also call &lt;CODE&gt;StartUpdate&lt;/CODE&gt;.&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;This won’t fix the known Jobs bug cases where one job run issues duplicate &lt;CODE&gt;StartUpdate&lt;/CODE&gt;, but it does remove &lt;EM&gt;most&lt;/EM&gt; genuine concurrency conflicts.&lt;/P&gt;
&lt;HR /&gt;
&lt;H3&gt;Option 2 – Treat as known Jobs / pipeline integration bug and add retries&lt;/H3&gt;
&lt;P&gt;If your pattern is “job task sometimes fails with &lt;CODE&gt;RUN_EXECUTION_ERROR&lt;/CODE&gt; but the underlying pipeline update actually succeeded” (as in ES-1635313 / Nokia cases), treat this as a transient integration bug:&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;In DAB, add a small &lt;CODE&gt;max_retries&lt;/CODE&gt; on the pipeline task itself so the job auto-retries when it sees this specific error.&lt;/LI&gt;
&lt;LI&gt;Operationally treat these failures as “false negatives” for now (confirm via &lt;CODE&gt;go/dlt/debug&lt;/CODE&gt; / event log that only one update ran and completed).&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;This doesn’t give you true queuing semantics, but it makes the symptom operationally tolerable until the backend rollout and Jobs fixes are fully in place.&lt;/P&gt;
&lt;HR /&gt;
&lt;H3&gt;Option 3 – Escalate via Support/ES to get attached to existing incidents &lt;STRONG&gt;(Recommended)&lt;/STRONG&gt;&lt;/H3&gt;
&lt;P&gt;Given you’re still seeing &lt;CODE&gt;RUN_EXECUTION_ERROR&lt;/CODE&gt; post-announcement and there are active incidents (ES-1635313, BL-16616/SUP-27441) specifically around this error and queued/duplicate &lt;CODE&gt;StartUpdate&lt;/CODE&gt; behavior:&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;Open a Support ticket (or ES via Support) with:
&lt;UL&gt;
&lt;LI&gt;Workspace ID, pipeline ID, and a few failing Job run URLs.&lt;/LI&gt;
&lt;LI&gt;Note that this is &lt;STRONG&gt;“Lakeflow Pipelines + Lakeflow Jobs: &lt;CODE&gt;RUN_EXECUTION_ERROR: Pipeline update already in progress&lt;/CODE&gt; despite queued execution being announced; please check against ES-1635313 / BL-16616 behavior.”&lt;/STRONG&gt;&lt;/LI&gt;
&lt;/UL&gt;
&lt;/LI&gt;
&lt;LI&gt;In parallel, apply Option 1 (single orchestrating job, &lt;CODE&gt;max_concurrent_runs&lt;/CODE&gt; or job queue) and, if needed, Option 2 (retries) as immediate mitigations.&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;&lt;STRONG&gt;Recommendation:&lt;/STRONG&gt;&lt;BR /&gt;Use &lt;STRONG&gt;Option 3&lt;/STRONG&gt; as the primary path so engineering can confirm whether your workspace/pipelines are on the queued-execution rollout and attach you to the ongoing fixes, and in the meantime implement &lt;STRONG&gt;Option 1&lt;/STRONG&gt; (single orchestrator + serialization) plus light retries from &lt;STRONG&gt;Option 2&lt;/STRONG&gt; to reduce operational pain.&lt;/P&gt;</description>
    <pubDate>Mon, 04 May 2026 17:48:27 GMT</pubDate>
    <dc:creator>Lu_Wang_ENB_DBX</dc:creator>
    <dc:date>2026-05-04T17:48:27Z</dc:date>
    <item>
      <title>Lakeflow Declarative Pipeline queue</title>
      <link>https://community.databricks.com/t5/data-engineering/lakeflow-declarative-pipeline-queue/m-p/156098#M54353</link>
      <description>&lt;P&gt;In the &lt;A href="https://docs.databricks.com/aws/en/release-notes/dlt/2026#january-2026:~:text=Pipelines%20now%20support%20queued%20execution%20mode%2C%20where%20multiple%20update%20requests%20are%20automatically%20queued%20and%20executed%20sequentially%20instead%20of%20failing%20with%20conflicts.%20This%20simplifies%20operations%20for%20pipelines%20with%20frequent%20update%20triggers%20and%20eliminates%20the%20need%20for%20manual%20retry%20coordination." target="_self"&gt;January 2026 release notes&lt;/A&gt;, it was announced that: "&lt;SPAN&gt;Pipelines now support queued execution mode, where multiple update requests are automatically queued and executed sequentially instead of failing with conflicts. This simplifies operations for pipelines with frequent update triggers and eliminates the need for manual retry coordination."&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;However, I am still seeing concurrent runs fail with "RUN_EXECUTION_ERROR: Pipeline update already in progress." I also don't see an option to apply this queue setting in the UI, nor any documentation for this within DAB. I tried setting this in DAB with `queue: enabled: true` that is set for a job, but this does not work.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Has the pipeline queue been working for anyone else?&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Mon, 04 May 2026 16:54:10 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/lakeflow-declarative-pipeline-queue/m-p/156098#M54353</guid>
      <dc:creator>tt_921</dc:creator>
      <dc:date>2026-05-04T16:54:10Z</dc:date>
    </item>
    <item>
      <title>Re: Lakeflow Declarative Pipeline queue</title>
      <link>https://community.databricks.com/t5/data-engineering/lakeflow-declarative-pipeline-queue/m-p/156103#M54356</link>
      <description>&lt;P&gt;&lt;STRONG&gt;What’s going on&lt;/STRONG&gt;&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;The January 2026 release note is correct that the &lt;EM&gt;engine&lt;/EM&gt; for Lakeflow Spark Declarative Pipelines supports &lt;STRONG&gt;queued execution&lt;/STRONG&gt;, but it’s a &lt;STRONG&gt;backend behavior&lt;/STRONG&gt;, not a user-configurable option; there is no UI or DAB field to flip it on or off.&lt;/LI&gt;
&lt;LI&gt;Internally, there has been a “queuing guardrail” and staged rollout work (e.g. discussion on enabling queued update execution and reverting that guardrail only now), so some environments/pipelines still behave as “fail on concurrent StartUpdate” instead of queueing.&lt;/LI&gt;
&lt;LI&gt;Separately, there are &lt;STRONG&gt;open/very recent bugs&lt;/STRONG&gt; where Jobs sends duplicate &lt;CODE&gt;StartUpdate&lt;/CODE&gt; requests or loses the response, causing &lt;CODE&gt;RUN_EXECUTION_ERROR: Pipeline update already in progress&lt;/CODE&gt; even when only one update is actually running; see ES-1635313 and follow-ons, and Nokia BL-16616 / SUP-27441 where the pipeline completes but the job task shows this error.&lt;/LI&gt;
&lt;LI&gt;The &lt;STRONG&gt;Jobs queue&lt;/STRONG&gt; setting (&lt;CODE&gt;queue.enabled: true&lt;/CODE&gt;) you tried in DAB is a &lt;EM&gt;Jobs-level run queue&lt;/EM&gt;, not the DLT/Lakeflow pipeline queuing feature, so toggling that won’t change the pipeline control-plane behavior.&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;Below are &lt;STRONG&gt;3 options&lt;/STRONG&gt;&lt;/P&gt;
&lt;HR /&gt;
&lt;H3&gt;Option 1 – Serialize all updates via one Job (no secondary triggers)&lt;/H3&gt;
&lt;P&gt;Mitigate by ensuring there is &lt;STRONG&gt;exactly one&lt;/STRONG&gt; place that can trigger the pipeline and that it never overlaps runs:&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;Use a &lt;STRONG&gt;single Lakeflow Job&lt;/STRONG&gt; with one pipeline task, and set either:
&lt;UL&gt;
&lt;LI&gt;&lt;CODE&gt;max_concurrent_runs = 1&lt;/CODE&gt;, or&lt;/LI&gt;
&lt;LI&gt;in DAB, on that same job:
&lt;PRE&gt;&lt;CODE class="language-yaml"&gt;jobs:
  my_pipeline_job:
    queue:
      enabled: true
&lt;/CODE&gt;&lt;/PRE&gt;
&lt;/LI&gt;
&lt;/UL&gt;
&lt;/LI&gt;
&lt;LI&gt;Remove/avoid: manual “Start” in the pipeline UI, other Jobs, or ad-hoc API calls that can also call &lt;CODE&gt;StartUpdate&lt;/CODE&gt;.&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;This won’t fix the known Jobs bug cases where one job run issues duplicate &lt;CODE&gt;StartUpdate&lt;/CODE&gt;, but it does remove &lt;EM&gt;most&lt;/EM&gt; genuine concurrency conflicts.&lt;/P&gt;
&lt;HR /&gt;
&lt;H3&gt;Option 2 – Treat as known Jobs / pipeline integration bug and add retries&lt;/H3&gt;
&lt;P&gt;If your pattern is “job task sometimes fails with &lt;CODE&gt;RUN_EXECUTION_ERROR&lt;/CODE&gt; but the underlying pipeline update actually succeeded” (as in ES-1635313 / Nokia cases), treat this as a transient integration bug:&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;In DAB, add a small &lt;CODE&gt;max_retries&lt;/CODE&gt; on the pipeline task itself so the job auto-retries when it sees this specific error.&lt;/LI&gt;
&lt;LI&gt;Operationally treat these failures as “false negatives” for now (confirm via &lt;CODE&gt;go/dlt/debug&lt;/CODE&gt; / event log that only one update ran and completed).&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;This doesn’t give you true queuing semantics, but it makes the symptom operationally tolerable until the backend rollout and Jobs fixes are fully in place.&lt;/P&gt;
&lt;HR /&gt;
&lt;H3&gt;Option 3 – Escalate via Support/ES to get attached to existing incidents &lt;STRONG&gt;(Recommended)&lt;/STRONG&gt;&lt;/H3&gt;
&lt;P&gt;Given you’re still seeing &lt;CODE&gt;RUN_EXECUTION_ERROR&lt;/CODE&gt; post-announcement and there are active incidents (ES-1635313, BL-16616/SUP-27441) specifically around this error and queued/duplicate &lt;CODE&gt;StartUpdate&lt;/CODE&gt; behavior:&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;Open a Support ticket (or ES via Support) with:
&lt;UL&gt;
&lt;LI&gt;Workspace ID, pipeline ID, and a few failing Job run URLs.&lt;/LI&gt;
&lt;LI&gt;Note that this is &lt;STRONG&gt;“Lakeflow Pipelines + Lakeflow Jobs: &lt;CODE&gt;RUN_EXECUTION_ERROR: Pipeline update already in progress&lt;/CODE&gt; despite queued execution being announced; please check against ES-1635313 / BL-16616 behavior.”&lt;/STRONG&gt;&lt;/LI&gt;
&lt;/UL&gt;
&lt;/LI&gt;
&lt;LI&gt;In parallel, apply Option 1 (single orchestrating job, &lt;CODE&gt;max_concurrent_runs&lt;/CODE&gt; or job queue) and, if needed, Option 2 (retries) as immediate mitigations.&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;&lt;STRONG&gt;Recommendation:&lt;/STRONG&gt;&lt;BR /&gt;Use &lt;STRONG&gt;Option 3&lt;/STRONG&gt; as the primary path so engineering can confirm whether your workspace/pipelines are on the queued-execution rollout and attach you to the ongoing fixes, and in the meantime implement &lt;STRONG&gt;Option 1&lt;/STRONG&gt; (single orchestrator + serialization) plus light retries from &lt;STRONG&gt;Option 2&lt;/STRONG&gt; to reduce operational pain.&lt;/P&gt;</description>
      <pubDate>Mon, 04 May 2026 17:48:27 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/lakeflow-declarative-pipeline-queue/m-p/156103#M54356</guid>
      <dc:creator>Lu_Wang_ENB_DBX</dc:creator>
      <dc:date>2026-05-04T17:48:27Z</dc:date>
    </item>
    <item>
      <title>Re: Lakeflow Declarative Pipeline queue</title>
      <link>https://community.databricks.com/t5/data-engineering/lakeflow-declarative-pipeline-queue/m-p/156110#M54357</link>
      <description>&lt;P&gt;Thank you very much for the detailed response! We unfortunately can't proceed with &lt;STRONG&gt;option 1&lt;/STRONG&gt;, as we do require multiple places that can trigger the pipeline (an API call to the parent job, and a direct API call to the pipeline itself). This is due to the specific configurable options available in a pipeline API call vs a job API call, namely `full_refresh_selection` to fully refresh specific tables.&amp;nbsp;&lt;/P&gt;&lt;P&gt;We do have queue enabled at the job level and a small `max_retries` on the pipeline.&lt;/P&gt;&lt;P&gt;For now it seems we will need to open a ticket and wait until the pipeline execution queue is fully rolled out.&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Mon, 04 May 2026 18:39:04 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/lakeflow-declarative-pipeline-queue/m-p/156110#M54357</guid>
      <dc:creator>tt_921</dc:creator>
      <dc:date>2026-05-04T18:39:04Z</dc:date>
    </item>
  </channel>
</rss>

