<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Yml file replacing job cluster with all-purpose cluster when running a workflow in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/yml-file-replacing-job-cluster-with-all-purpose-cluster-when/m-p/72333#M34548</link>
    <description>&lt;P&gt;I don't know if you've cut off your yaml snippet, but your snippet doesn't show your job cluster with key &lt;EM&gt;job-cluster.&amp;nbsp;&lt;/EM&gt;Just to validate, your job cluster is also defined in your workflow yaml?&lt;BR /&gt;&lt;BR /&gt;Edit: Looking it it again and knowing the defaults, it looks like you're pointing to job_cluster_key "job-cluster". The default is &lt;STRONG&gt;job_cluster&lt;/STRONG&gt; (with an underscore instead of a regular hyphen). Could this be your issue?&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Tue, 11 Jun 2024 13:35:37 GMT</pubDate>
    <dc:creator>jacovangelder</dc:creator>
    <dc:date>2024-06-11T13:35:37Z</dc:date>
    <item>
      <title>Yml file replacing job cluster with all-purpose cluster when running a workflow</title>
      <link>https://community.databricks.com/t5/data-engineering/yml-file-replacing-job-cluster-with-all-purpose-cluster-when/m-p/72248#M34529</link>
      <description>&lt;P&gt;Hi all,&lt;/P&gt;&lt;P&gt;I have been having some trouble running a workflow that consists of 3 tasks that run sequentially. Task1 runs on an all-purpose cluster and kicks off Task2 that needs to run on a job cluster. Task2 kicks off Task3 which also uses a job cluster.&lt;/P&gt;&lt;P&gt;We have identified that Task2 is running on an all-purpose cluster instead of a job cluster despite configuring the task to run using a job cluster in the yaml file for the asset bundle. This task is dependent on another task which does use the all-purpose cluster as specified in the yaml file. We tried modifying the yaml file but when running a &lt;I&gt;databricks bundle validate, it looks like the task is being overwritten to use the all-purpose cluster despite explicitly indicating it to use the job cluster. Renaming the task names is being picked up by the validate command.&lt;/I&gt;&lt;/P&gt;&lt;P&gt;Here is a snippet of the yaml file:&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;tasks:
  - task_key: Task1
    existing_cluster_id: all-purpose-cluster-id
    notebook_task:
      notebook_path: ../src/Task1.py
      base_parameters:
        catalog: ${var.catalog}
        target: ${var.target}

  - task_key: Task2
    job_cluster_key: job-cluster
    depends_on:
      - task_key: Task1
    notebook_task:
      notebook_path: ../src/Task2.py
      base_parameters:
        catalog: ${var.catalog}
        target: ${var.target}

  - task_key: Task3
    job_cluster_key: job-cluster
    depends_on:
      - task_key: Task2
    notebook_task:
      notebook_path: ../src/Task3.py
      base_parameters:
        catalog: ${var.catalog}
        target: ${var.target}&lt;/LI-CODE&gt;&lt;P&gt;After running &lt;I&gt;databricks bundle validate, this is the output:&lt;/I&gt;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;"tasks": [
          {
            "existing_cluster_id": "all-purpose-cluster-id",
            "notebook_task": {
              "base_parameters": {
                "catalog": "catalog",
                "target": "target"
              },
              "notebook_path": "/Users/user/.bundle/folder/dev/files/src/Task1"
            },
            "task_key": "Task1"
          },
          {
            "depends_on": [
              {
                "task_key": "Task1"
              }
            ],
            "existing_cluster_id": "all-purpose-cluster-id",
            "notebook_task": {
              "base_parameters": {
                "catalog": "catalog",
                "target": "target"
              },
              "notebook_path": "/Users/user/.bundle/folder/dev/files/src/Task2"
            },
            "task_key": "Task2"
          },
          {
            "depends_on": [
              {
                "task_key": "Task2"
              }
            ],
            "existing_cluster_id": "all-purpose-cluster-id",
            "notebook_task": {
              "base_parameters": {
                "catalog": "catalog",
                "target": "target"
              },
              "notebook_path": "/Users/user/.bundle/folder/dev/files/src/Task3"
            },
            "task_key": "Task3"
          }
        ]&lt;/LI-CODE&gt;&lt;P&gt;As you can see, the all-purpose cluster id is replacing the job-cluster key for Task2 and Task3. The strangest part of all of this is that I'm the only one on the team that is experiencing this issue. Everyone else seems to be able to run the workflow without any issues. Any ideas on how to resolve this issue?&lt;/P&gt;&lt;P&gt;Thank you in advanced!&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Mon, 10 Jun 2024 16:51:30 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/yml-file-replacing-job-cluster-with-all-purpose-cluster-when/m-p/72248#M34529</guid>
      <dc:creator>dataengutility</dc:creator>
      <dc:date>2024-06-10T16:51:30Z</dc:date>
    </item>
    <item>
      <title>Re: Yml file replacing job cluster with all-purpose cluster when running a workflow</title>
      <link>https://community.databricks.com/t5/data-engineering/yml-file-replacing-job-cluster-with-all-purpose-cluster-when/m-p/72333#M34548</link>
      <description>&lt;P&gt;I don't know if you've cut off your yaml snippet, but your snippet doesn't show your job cluster with key &lt;EM&gt;job-cluster.&amp;nbsp;&lt;/EM&gt;Just to validate, your job cluster is also defined in your workflow yaml?&lt;BR /&gt;&lt;BR /&gt;Edit: Looking it it again and knowing the defaults, it looks like you're pointing to job_cluster_key "job-cluster". The default is &lt;STRONG&gt;job_cluster&lt;/STRONG&gt; (with an underscore instead of a regular hyphen). Could this be your issue?&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Tue, 11 Jun 2024 13:35:37 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/yml-file-replacing-job-cluster-with-all-purpose-cluster-when/m-p/72333#M34548</guid>
      <dc:creator>jacovangelder</dc:creator>
      <dc:date>2024-06-11T13:35:37Z</dc:date>
    </item>
    <item>
      <title>Re: Yml file replacing job cluster with all-purpose cluster when running a workflow</title>
      <link>https://community.databricks.com/t5/data-engineering/yml-file-replacing-job-cluster-with-all-purpose-cluster-when/m-p/72403#M34563</link>
      <description>&lt;P&gt;Hi, thank you for your response! Yes, I did cut off the yaml snippet to the problem area since the entire yaml is quite a large file. We do define the job-cluster in the workflow yaml.&amp;nbsp;&lt;SPAN&gt;"job-cluster" is just a pseudonym. Sorry for the confusing snippet. Here is the portion where the job cluster is defined:&lt;/SPAN&gt;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;job_clusters:
        - job_cluster_key: job-cluster
          new_cluster:
            spark_version: 12.2.x-scala2.12
            node_type_id: m5d.4xlarge
            driver_node_type_id: m5d.4xlarge
            data_security_mode: SINGLE_USER
            runtime_engine: PHOTON
            autoscale:
              min_workers: 2
              max_workers: 12
            aws_attributes:
              instance_profile_arn: [removed]
              zone_id: auto
              first_on_demand: 1&lt;/LI-CODE&gt;</description>
      <pubDate>Tue, 11 Jun 2024 18:44:04 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/yml-file-replacing-job-cluster-with-all-purpose-cluster-when/m-p/72403#M34563</guid>
      <dc:creator>dataengutility</dc:creator>
      <dc:date>2024-06-11T18:44:04Z</dc:date>
    </item>
    <item>
      <title>Re: Yml file replacing job cluster with all-purpose cluster when running a workflow</title>
      <link>https://community.databricks.com/t5/data-engineering/yml-file-replacing-job-cluster-with-all-purpose-cluster-when/m-p/72408#M34564</link>
      <description>&lt;P&gt;That should work just fine. Just tested it on my end. As long as your &lt;STRONG&gt;job_cluster_key&lt;/STRONG&gt; value matches with the one in your task, it should work.&amp;nbsp;&lt;BR /&gt;&lt;BR /&gt;Perhaps you can try to throw away your bundle folder (and perhaps your workflows too) in your workspace and then deploy again. Do keep in mind that run history is purged because of this. It could be that the Terraform state is somehow messed up because of previous faulty deployments.&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 12 Jun 2024 07:15:12 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/yml-file-replacing-job-cluster-with-all-purpose-cluster-when/m-p/72408#M34564</guid>
      <dc:creator>jacovangelder</dc:creator>
      <dc:date>2024-06-12T07:15:12Z</dc:date>
    </item>
    <item>
      <title>Re: Yml file replacing job cluster with all-purpose cluster when running a workflow</title>
      <link>https://community.databricks.com/t5/data-engineering/yml-file-replacing-job-cluster-with-all-purpose-cluster-when/m-p/74679#M34775</link>
      <description>&lt;P&gt;My issue is resolved. I had to upgrade my CLI version from v0.215 to v0.221 and everything works fine now. Thank you for your help!&lt;/P&gt;</description>
      <pubDate>Mon, 17 Jun 2024 16:57:13 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/yml-file-replacing-job-cluster-with-all-purpose-cluster-when/m-p/74679#M34775</guid>
      <dc:creator>dataengutility</dc:creator>
      <dc:date>2024-06-17T16:57:13Z</dc:date>
    </item>
    <item>
      <title>Re: Yml file replacing job cluster with all-purpose cluster when running a workflow</title>
      <link>https://community.databricks.com/t5/data-engineering/yml-file-replacing-job-cluster-with-all-purpose-cluster-when/m-p/148955#M52996</link>
      <description>&lt;P&gt;I am facing same issue , i see 2 clusters one ALL purpose which is default and job compute, in my YML I have configured to use only job compute, not sure how All Purpose is getting referenced ? I do not have any CLI issue as i am using latest ,&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Sat, 21 Feb 2026 13:24:53 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/yml-file-replacing-job-cluster-with-all-purpose-cluster-when/m-p/148955#M52996</guid>
      <dc:creator>vsharma</dc:creator>
      <dc:date>2026-02-21T13:24:53Z</dc:date>
    </item>
  </channel>
</rss>

