<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Multi Task Job  creation through Pulumi in Get Started Discussions</title>
    <link>https://community.databricks.com/t5/get-started-discussions/multi-task-job-creation-through-pulumi/m-p/55030#M1956</link>
    <description>&lt;P&gt;Hello &lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/9"&gt;@Retired_mod&lt;/a&gt;, thanks for your answer, but the problem keeps the same. I had already tested with different cluster configurations, single-node and multi-node, including those cluster configurations which worked with single task jobs, but the error does not change, it is always about the new cluster size.&lt;/P&gt;&lt;P&gt;According to&amp;nbsp; documentation here:&amp;nbsp;&lt;A href="https://www.pulumi.com/registry/packages/databricks/api-docs/job/#jobnewcluster" target="_blank"&gt;https://www.pulumi.com/registry/packages/databricks/api-docs/job/#jobnewcluster&lt;/A&gt;&amp;nbsp;I understand that I need to set the cluster specifications in the parameter&amp;nbsp;&lt;STRONG&gt;job_clusters&lt;/STRONG&gt;, not in&amp;nbsp;&lt;STRONG&gt;new_cluster&lt;/STRONG&gt; as with single task jobs.&lt;/P&gt;</description>
    <pubDate>Mon, 11 Dec 2023 08:33:28 GMT</pubDate>
    <dc:creator>Borkadd</dc:creator>
    <dc:date>2023-12-11T08:33:28Z</dc:date>
    <item>
      <title>Multi Task Job  creation through Pulumi</title>
      <link>https://community.databricks.com/t5/get-started-discussions/multi-task-job-creation-through-pulumi/m-p/54955#M1948</link>
      <description>&lt;P&gt;I am trying to create a multi-task Databricks Job in Azure Cloud with its own cluster.&lt;/P&gt;&lt;P&gt;Although I was able to create a single task job without any issues, the code to deploy the multi-task job fails due to the following cluster validation error:&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;error: 1 error occurred:
        * cannot create job: Cluster validation error: Missing required field: settings.cluster_spec.new_cluster.size&lt;/LI-CODE&gt;&lt;P&gt;he code to create the Job is the following:&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;job = Job(
            resource_name = f"{job_name}-job",
            args=JobArgs(
                name = f"{job_name}-job",
                job_clusters=[
                    JobJobClusterArgs(
                        job_cluster_key="pulumiTest-basic-cluster",
                        new_cluster=JobJobClusterNewClusterArgs(
                            spark_version="13.3.x-scala2.12",
                            cluster_name="",
                            num_workers=0,
                            node_type_id="Standard_DS3_v2",
                            enable_elastic_disk=True,
                            runtime_engine="STANDARD",
                            spark_conf={
                                # f"fs.azure.account.key.{self.storage_account_name}.dfs.core.windows.net": "{{secrets/pulumiTest-secret-scope/puluTest-storage-access-token}}"
                                "spark.master": "local[*,4]",
                                "spark.databricks.cluster.profile": "singleNode"
                            },
                            custom_tags={
                                "ResourceClass": "SingleNode"
                            },
                            data_security_mode="LEGACY_SINGLE_USER_STANDARD"
                        )
                    )
                ],
                computes=[
                    JobComputeArgs(
                        compute_key="landing_task",
                        spec=JobComputeSpecArgs(kind="spark_python_task")
                    ),
                    JobComputeArgs(
                        compute_key="staging_task",
                        spec=JobComputeSpecArgs(kind="spark_python_task")
                    ),
                    JobComputeArgs(
                        compute_key="refined_task",
                        spec=JobComputeSpecArgs(kind="spark_python_task")
                    )
                ],
                tasks = [
                    JobTaskArgs(
                        task_key="landing_task",
                        job_cluster_key="pulumiTest-basic-cluster",
                        spark_python_task=JobSparkPythonTaskArgs(
                            python_file="/pipelineExample/landing.py",
                            source="GIT"
                        ),
                        run_if="ALL_SUCCESS",
                        libraries=[
                            JobLibraryArgs(
                                whl=whl_path
                            )
                        ]
                    ),
                    JobTaskArgs(
                        task_key="staging_task",
                        job_cluster_key="pulumiTest-basic-cluster",
                        spark_python_task=JobSparkPythonTaskArgs(
                            python_file="/pipelineExample/staging.py",
                            source="GIT"
                        ),
                        depends_ons=[
                            JobTaskDependsOnArgs(
                                task_key="landing_task"
                            )
                        ],
                        run_if="ALL_SUCCESS",
                        libraries=[
                            JobLibraryArgs(
                                whl=whl_path
                            )
                        ]
                    ),
                    JobTaskArgs(
                        task_key="refined_task",
                        job_cluster_key="pulumiTest-basic-cluster",
                        spark_python_task=JobSparkPythonTaskArgs(
                            python_file="/pipelineExample/refined.py",
                            source="GIT"
                        ),
                        depends_ons=[
                            JobTaskDependsOnArgs(
                                task_key="staging_task"
                            )
                        ],
                        run_if="ALL_SUCCESS",
                        libraries=[
                            JobLibraryArgs(
                                whl=whl_path
                            )
                        ]
                    )
                ],
                git_source=JobGitSourceArgs(
                    url=git_url,
                    provider="gitHub",
                    branch="main"
                ),
                format="MULTI_TASK"
            )
        )
)
pulumi.export('Job URL', job.url)&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;Does anyone know where the problem could be?&lt;/P&gt;</description>
      <pubDate>Fri, 08 Dec 2023 22:09:27 GMT</pubDate>
      <guid>https://community.databricks.com/t5/get-started-discussions/multi-task-job-creation-through-pulumi/m-p/54955#M1948</guid>
      <dc:creator>Borkadd</dc:creator>
      <dc:date>2023-12-08T22:09:27Z</dc:date>
    </item>
    <item>
      <title>Re: Multi Task Job  creation through Pulumi</title>
      <link>https://community.databricks.com/t5/get-started-discussions/multi-task-job-creation-through-pulumi/m-p/55030#M1956</link>
      <description>&lt;P&gt;Hello &lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/9"&gt;@Retired_mod&lt;/a&gt;, thanks for your answer, but the problem keeps the same. I had already tested with different cluster configurations, single-node and multi-node, including those cluster configurations which worked with single task jobs, but the error does not change, it is always about the new cluster size.&lt;/P&gt;&lt;P&gt;According to&amp;nbsp; documentation here:&amp;nbsp;&lt;A href="https://www.pulumi.com/registry/packages/databricks/api-docs/job/#jobnewcluster" target="_blank"&gt;https://www.pulumi.com/registry/packages/databricks/api-docs/job/#jobnewcluster&lt;/A&gt;&amp;nbsp;I understand that I need to set the cluster specifications in the parameter&amp;nbsp;&lt;STRONG&gt;job_clusters&lt;/STRONG&gt;, not in&amp;nbsp;&lt;STRONG&gt;new_cluster&lt;/STRONG&gt; as with single task jobs.&lt;/P&gt;</description>
      <pubDate>Mon, 11 Dec 2023 08:33:28 GMT</pubDate>
      <guid>https://community.databricks.com/t5/get-started-discussions/multi-task-job-creation-through-pulumi/m-p/55030#M1956</guid>
      <dc:creator>Borkadd</dc:creator>
      <dc:date>2023-12-11T08:33:28Z</dc:date>
    </item>
  </channel>
</rss>

