<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Spark verison errors in &amp;quot;Build an ETL pipeline with Lakeflow Spark Declarative Pipelines&amp;quot; in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/spark-verison-errors-in-quot-build-an-etl-pipeline-with-lakeflow/m-p/142633#M51987</link>
    <description>&lt;P&gt;I'm trying to define a job for a pipeline using the Asset Bundle Python SDK. I created the pipeline first (using the SDK) and i'm now trying to add the Job. The DAB validates and deploys successfully, but when I run the Job i get an error:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;UNAUTHORIZED_ERROR: User &amp;lt;some-guid&amp;gt; does not have Run permissions on pipeline None.&lt;/DIV&gt;&lt;DIV class=""&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV class=""&gt;How can I define the job to link to the already existing pipeline (which is already running in Continuous mode)?&lt;/DIV&gt;&lt;DIV class=""&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV class=""&gt;The DAB code is as follows:&lt;/DIV&gt;&lt;DIV class=""&gt;&amp;nbsp;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;LI-CODE lang="python"&gt;my_pipeline = Pipeline(
    name = "My Pipeline",
    catalog = "mycatalog",
    schema = "default",
    continuous=True,
    clusters = [
        PipelineCluster(
            ...
        )
    ],
    libraries = [
        PipelineLibrary(
            file=FileLibrary(path="src/my_sdp.py")
        )
    ]
)

my_task = Task(
    task_key="My_pipeline_task",
    pipeline_task=PipelineTask(
        pipeline_id=str(my_pipeline.id)
    )
)

my_job = Job(
    name="My Pipeline Job",
    tasks=[
        my_task
    ]
)&lt;/LI-CODE&gt;</description>
    <pubDate>Mon, 29 Dec 2025 11:26:32 GMT</pubDate>
    <dc:creator>liquibricks</dc:creator>
    <dc:date>2025-12-29T11:26:32Z</dc:date>
    <item>
      <title>Spark verison errors in "Build an ETL pipeline with Lakeflow Spark Declarative Pipelines"</title>
      <link>https://community.databricks.com/t5/data-engineering/spark-verison-errors-in-quot-build-an-etl-pipeline-with-lakeflow/m-p/142633#M51987</link>
      <description>&lt;P&gt;I'm trying to define a job for a pipeline using the Asset Bundle Python SDK. I created the pipeline first (using the SDK) and i'm now trying to add the Job. The DAB validates and deploys successfully, but when I run the Job i get an error:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;UNAUTHORIZED_ERROR: User &amp;lt;some-guid&amp;gt; does not have Run permissions on pipeline None.&lt;/DIV&gt;&lt;DIV class=""&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV class=""&gt;How can I define the job to link to the already existing pipeline (which is already running in Continuous mode)?&lt;/DIV&gt;&lt;DIV class=""&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV class=""&gt;The DAB code is as follows:&lt;/DIV&gt;&lt;DIV class=""&gt;&amp;nbsp;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;LI-CODE lang="python"&gt;my_pipeline = Pipeline(
    name = "My Pipeline",
    catalog = "mycatalog",
    schema = "default",
    continuous=True,
    clusters = [
        PipelineCluster(
            ...
        )
    ],
    libraries = [
        PipelineLibrary(
            file=FileLibrary(path="src/my_sdp.py")
        )
    ]
)

my_task = Task(
    task_key="My_pipeline_task",
    pipeline_task=PipelineTask(
        pipeline_id=str(my_pipeline.id)
    )
)

my_job = Job(
    name="My Pipeline Job",
    tasks=[
        my_task
    ]
)&lt;/LI-CODE&gt;</description>
      <pubDate>Mon, 29 Dec 2025 11:26:32 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/spark-verison-errors-in-quot-build-an-etl-pipeline-with-lakeflow/m-p/142633#M51987</guid>
      <dc:creator>liquibricks</dc:creator>
      <dc:date>2025-12-29T11:26:32Z</dc:date>
    </item>
    <item>
      <title>Re: Spark verison errors in "Build an ETL pipeline with Lakeflow Spark Declarative Pipelines&amp;qu</title>
      <link>https://community.databricks.com/t5/data-engineering/spark-verison-errors-in-quot-build-an-etl-pipeline-with-lakeflow/m-p/142636#M51989</link>
      <description>&lt;P&gt;Hi, if you've already created the pipeline you don't need to create it again via the DAB, just get the pipeline id from the UI and pass that into your job. Also your syntax for the task and the job should be more like this:&lt;/P&gt;
&lt;LI-CODE lang="markup"&gt;  jobs:
    my_pipeline_job:
      name: my-pipeline-job
      tasks:
        - task_key: my-pipeline-task
          pipeline_task:
            pipeline_id: [pass your the pipeline of existing pipeline here]&lt;/LI-CODE&gt;</description>
      <pubDate>Mon, 29 Dec 2025 11:51:32 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/spark-verison-errors-in-quot-build-an-etl-pipeline-with-lakeflow/m-p/142636#M51989</guid>
      <dc:creator>emma_s</dc:creator>
      <dc:date>2025-12-29T11:51:32Z</dc:date>
    </item>
    <item>
      <title>Re: Spark verison errors in "Build an ETL pipeline with Lakeflow Spark Declarative Pipelines&amp;qu</title>
      <link>https://community.databricks.com/t5/data-engineering/spark-verison-errors-in-quot-build-an-etl-pipeline-with-lakeflow/m-p/142640#M51990</link>
      <description>&lt;P&gt;The error happens because my_pipeline.id does not exist when the Asset Bundle is defined. Resource IDs are only created after deployment, so your job is effectively created with pipeline_id = None. When the job runs, Databricks tries to run a pipeline with ID None, which results in the “Run permissions on pipeline None” error.&lt;/P&gt;&lt;P&gt;In Databricks Asset Bundles, you must link resources symbolically, not by accessing their IDs directly in Python.&lt;/P&gt;&lt;P&gt;To fix this, reference the pipeline using the bundle resource reference syntax:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="python"&gt;my_task = Task(
    task_key="My_pipeline_task",
    pipeline_task=PipelineTask(
        pipeline_id="${resources.pipelines.my_pipeline.id}"
    )
)&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Here, my_pipeline is the Python variable name used when defining the Pipeline resource. Databricks resolves this reference to the actual pipeline ID at deploy time.&lt;/P&gt;&lt;P&gt;Your job definition can then remain unchanged.&lt;/P&gt;&lt;P&gt;One important note: because your pipeline is running in continuous mode, triggering it from a job will restart the pipeline each time the job runs. If you don’t need scheduled restarts or orchestration with other tasks, you may not need a job at all, just deploying the pipeline is sufficient.&lt;/P&gt;&lt;P&gt;Key takeaway: never use .id directly in Asset Bundle code. Always use ${resources.&amp;lt;type&amp;gt;.&amp;lt;name&amp;gt;.id} to link bundle-managed resources.&lt;/P&gt;</description>
      <pubDate>Mon, 29 Dec 2025 12:54:58 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/spark-verison-errors-in-quot-build-an-etl-pipeline-with-lakeflow/m-p/142640#M51990</guid>
      <dc:creator>ethanop</dc:creator>
      <dc:date>2025-12-29T12:54:58Z</dc:date>
    </item>
    <item>
      <title>Re: Spark verison errors in "Build an ETL pipeline with Lakeflow Spark Declarative Pipelines&amp;qu</title>
      <link>https://community.databricks.com/t5/data-engineering/spark-verison-errors-in-quot-build-an-etl-pipeline-with-lakeflow/m-p/142655#M51994</link>
      <description>&lt;P&gt;This happens because the job is not actually linked to the deployed pipeline and the pipeline id is None at runtime. When using Asset Bundles, the pipeline id is only resolved after deployment, so referencing my_pipeline.id in code does not work. Instead, the job must reference the pipeline using the bundle resource reference, not a Python variable. You should define the pipeline and job as bundle resources and set the pipeline task pipeline id to the bundle reference for that pipeline. Also ensure that the job owner has Run permission on the pipeline. Once the job correctly references the deployed pipeline resource and permissions are in place, the unauthorized error will be resolved.&lt;/P&gt;</description>
      <pubDate>Mon, 29 Dec 2025 16:05:19 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/spark-verison-errors-in-quot-build-an-etl-pipeline-with-lakeflow/m-p/142655#M51994</guid>
      <dc:creator>mukul1409</dc:creator>
      <dc:date>2025-12-29T16:05:19Z</dc:date>
    </item>
  </channel>
</rss>

