<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Policy for DLT in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/policy-for-dlt/m-p/120976#M46295</link>
    <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/99143"&gt;@ankit001mittal&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;The error you're encountering is because Delta Live Tables (DLT) has specific requirements and automatically manages certain cluster configurations, including the Spark version. DLT pipelines are designed to use optimized Spark versions that are compatible with the DLT runtime, and allowing users to specify custom Spark versions can lead to compatibility issues.&lt;BR /&gt;Here's how to fix your cluster policy for DLT pipelines:&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Remove the spark_version constraint from your policy:&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;{&lt;BR /&gt;"spark_conf.spark.databricks.cluster.profile": {&lt;BR /&gt;"type": "forbidden",&lt;BR /&gt;"hidden": true&lt;BR /&gt;},&lt;BR /&gt;"node_type_id": {&lt;BR /&gt;"type": "unlimited",&lt;BR /&gt;"defaultValue": "Standard_DS3_v2",&lt;BR /&gt;"isOptional": true&lt;BR /&gt;},&lt;BR /&gt;"num_workers": {&lt;BR /&gt;"type": "unlimited",&lt;BR /&gt;"defaultValue": 4,&lt;BR /&gt;"isOptional": true&lt;BR /&gt;},&lt;BR /&gt;"azure_attributes.availability": {&lt;BR /&gt;"type": "unlimited",&lt;BR /&gt;"defaultValue": "SPOT_WITH_FALLBACK_AZURE"&lt;BR /&gt;},&lt;BR /&gt;"azure_attributes.spot_bid_max_price": {&lt;BR /&gt;"type": "fixed",&lt;BR /&gt;"value": 100,&lt;BR /&gt;"hidden": true&lt;BR /&gt;},&lt;BR /&gt;"instance_pool_id": {&lt;BR /&gt;"type": "forbidden",&lt;BR /&gt;"hidden": true&lt;BR /&gt;},&lt;BR /&gt;"driver_instance_pool_id": {&lt;BR /&gt;"type": "forbidden",&lt;BR /&gt;"hidden": true&lt;BR /&gt;},&lt;BR /&gt;"cluster_type": {&lt;BR /&gt;"type": "fixed",&lt;BR /&gt;"value": "dlt"&lt;BR /&gt;}&lt;BR /&gt;}&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;STRONG&gt;Why this happens:&lt;/STRONG&gt;&lt;BR /&gt;1. DLT Runtime Management: DLT automatically selects and manages the appropriate Spark version based on the DLT runtime version and channel (current/preview) you're using&lt;BR /&gt;2. Compatibility: DLT includes specific optimizations and features that require particular Spark versions&lt;BR /&gt;3. Automatic Updates: DLT handles Spark version updates as part of its managed service approach&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;STRONG&gt;Alternative approaches if you need version control:&lt;/STRONG&gt;&lt;BR /&gt;1. Use DLT Runtime Channels: Instead of specifying Spark versions, you can control which DLT runtime channel your pipeline uses (current vs preview) in the pipeline configuration&lt;BR /&gt;2. Separate Policies: Consider having separate cluster policies - one for DLT pipelines (without spark_version) and another for regular clusters (with spark_version constraints)&lt;BR /&gt;3. Pipeline-Level Configuration: Set any specific runtime requirements at the pipeline level rather than the cluster policy level&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Thu, 05 Jun 2025 03:26:07 GMT</pubDate>
    <dc:creator>lingareddy_Alva</dc:creator>
    <dc:date>2025-06-05T03:26:07Z</dc:date>
    <item>
      <title>Policy for DLT</title>
      <link>https://community.databricks.com/t5/data-engineering/policy-for-dlt/m-p/120967#M46291</link>
      <description>&lt;P&gt;Hi,&lt;BR /&gt;I am trying to define a policy for our DLT pipelines and I would like to provide a specific spark version like in the below example:&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;{
  "spark_conf.spark.databricks.cluster.profile": {
    "type": "forbidden",
    "hidden": true
  },
  "spark_version": {
    "type": "allowlist",
    "values": [
      "14.3.x-scala2.12"
    ]
  },
  "node_type_id": {
    "type": "unlimited",
    "defaultValue": "Standard_DS3_v2",
    "isOptional": true
  },
  "num_workers": {
    "type": "unlimited",
    "defaultValue": 4,
    "isOptional": true
  },
  "azure_attributes.availability": {
    "type": "unlimited",
    "defaultValue": "SPOT_WITH_FALLBACK_AZURE"
  },
  "azure_attributes.spot_bid_max_price": {
    "type": "fixed",
    "value": 100,
    "hidden": true
  },
  "instance_pool_id": {
    "type": "forbidden",
    "hidden": true
  },
  "driver_instance_pool_id": {
    "type": "forbidden",
    "hidden": true
  },
  "cluster_type": {
    "type": "fixed",
    "value": "dlt"
  }
}&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;But I am getting this error in my pipeline:&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;INVALID_PARAMETER_VALUE&lt;/STRONG&gt;: [DLT ERROR CODE: INVALID_CLUSTER_SETTING.CLIENT_ERROR] The cluster policy specified in the pipeline settings is not compatible with Delta Live Tables. Remove 'spark_version’ from your cluster policy.&lt;BR /&gt;&lt;BR /&gt;Could you please help me with it?&lt;/P&gt;</description>
      <pubDate>Wed, 04 Jun 2025 20:52:30 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/policy-for-dlt/m-p/120967#M46291</guid>
      <dc:creator>ankit001mittal</dc:creator>
      <dc:date>2025-06-04T20:52:30Z</dc:date>
    </item>
    <item>
      <title>Re: Policy for DLT</title>
      <link>https://community.databricks.com/t5/data-engineering/policy-for-dlt/m-p/120976#M46295</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/99143"&gt;@ankit001mittal&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;The error you're encountering is because Delta Live Tables (DLT) has specific requirements and automatically manages certain cluster configurations, including the Spark version. DLT pipelines are designed to use optimized Spark versions that are compatible with the DLT runtime, and allowing users to specify custom Spark versions can lead to compatibility issues.&lt;BR /&gt;Here's how to fix your cluster policy for DLT pipelines:&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Remove the spark_version constraint from your policy:&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;{&lt;BR /&gt;"spark_conf.spark.databricks.cluster.profile": {&lt;BR /&gt;"type": "forbidden",&lt;BR /&gt;"hidden": true&lt;BR /&gt;},&lt;BR /&gt;"node_type_id": {&lt;BR /&gt;"type": "unlimited",&lt;BR /&gt;"defaultValue": "Standard_DS3_v2",&lt;BR /&gt;"isOptional": true&lt;BR /&gt;},&lt;BR /&gt;"num_workers": {&lt;BR /&gt;"type": "unlimited",&lt;BR /&gt;"defaultValue": 4,&lt;BR /&gt;"isOptional": true&lt;BR /&gt;},&lt;BR /&gt;"azure_attributes.availability": {&lt;BR /&gt;"type": "unlimited",&lt;BR /&gt;"defaultValue": "SPOT_WITH_FALLBACK_AZURE"&lt;BR /&gt;},&lt;BR /&gt;"azure_attributes.spot_bid_max_price": {&lt;BR /&gt;"type": "fixed",&lt;BR /&gt;"value": 100,&lt;BR /&gt;"hidden": true&lt;BR /&gt;},&lt;BR /&gt;"instance_pool_id": {&lt;BR /&gt;"type": "forbidden",&lt;BR /&gt;"hidden": true&lt;BR /&gt;},&lt;BR /&gt;"driver_instance_pool_id": {&lt;BR /&gt;"type": "forbidden",&lt;BR /&gt;"hidden": true&lt;BR /&gt;},&lt;BR /&gt;"cluster_type": {&lt;BR /&gt;"type": "fixed",&lt;BR /&gt;"value": "dlt"&lt;BR /&gt;}&lt;BR /&gt;}&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;STRONG&gt;Why this happens:&lt;/STRONG&gt;&lt;BR /&gt;1. DLT Runtime Management: DLT automatically selects and manages the appropriate Spark version based on the DLT runtime version and channel (current/preview) you're using&lt;BR /&gt;2. Compatibility: DLT includes specific optimizations and features that require particular Spark versions&lt;BR /&gt;3. Automatic Updates: DLT handles Spark version updates as part of its managed service approach&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;STRONG&gt;Alternative approaches if you need version control:&lt;/STRONG&gt;&lt;BR /&gt;1. Use DLT Runtime Channels: Instead of specifying Spark versions, you can control which DLT runtime channel your pipeline uses (current vs preview) in the pipeline configuration&lt;BR /&gt;2. Separate Policies: Consider having separate cluster policies - one for DLT pipelines (without spark_version) and another for regular clusters (with spark_version constraints)&lt;BR /&gt;3. Pipeline-Level Configuration: Set any specific runtime requirements at the pipeline level rather than the cluster policy level&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 05 Jun 2025 03:26:07 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/policy-for-dlt/m-p/120976#M46295</guid>
      <dc:creator>lingareddy_Alva</dc:creator>
      <dc:date>2025-06-05T03:26:07Z</dc:date>
    </item>
  </channel>
</rss>

