<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Add Spark Configurations Serverless Compute in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/add-spark-configurations-serverless-compute/m-p/125323#M47419</link>
    <description>&lt;P&gt;Thank you for your reply. Hopefully they add this functionality in the future.&lt;/P&gt;</description>
    <pubDate>Tue, 15 Jul 2025 15:00:09 GMT</pubDate>
    <dc:creator>mac_delvalle</dc:creator>
    <dc:date>2025-07-15T15:00:09Z</dc:date>
    <item>
      <title>Add Spark Configurations Serverless Compute</title>
      <link>https://community.databricks.com/t5/data-engineering/add-spark-configurations-serverless-compute/m-p/123763#M47064</link>
      <description>&lt;P&gt;Hi everyone,&lt;/P&gt;&lt;P&gt;We’re in the process of migrating from all-purpose clusters to serverless compute in Databricks. On our all-purpose clusters, we’ve been setting specific Spark configurations (e.g., via the cluster’s advanced options). However, we’ve noticed that serverless compute doesn’t expose the same “Advanced Options” UI for Spark config overrides.&lt;/P&gt;&lt;P&gt;Is there a recommended way to apply custom Spark configurations when using serverless compute? Ideally, we’d like to apply these settings at cluster start or job submission time.&lt;/P&gt;&lt;P&gt;Any guidance or best practices would be greatly appreciated!&lt;/P&gt;&lt;P&gt;Below are the configs we are setting:&lt;/P&gt;&lt;P&gt;Overwrite partitions:&amp;nbsp;&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;spark.sql.sources.partitionOverwriteMode dynamic&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;Remove default files written (_started, _SUCCESS, _committed):&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;mapreduce.fileoutputcommitter.marksuccessfuljobs false&lt;/LI&gt;&lt;LI&gt;parquet.enable.summary-metadata false&lt;/LI&gt;&lt;LI&gt;spark.sql.sources.commitProtocolClass org.apache.spark.sql.execution.datasources.SQLHadoopMapReduceCommitProtocol&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;Thanks,&lt;BR /&gt;Mac&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 02 Jul 2025 18:20:46 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/add-spark-configurations-serverless-compute/m-p/123763#M47064</guid>
      <dc:creator>mac_delvalle</dc:creator>
      <dc:date>2025-07-02T18:20:46Z</dc:date>
    </item>
    <item>
      <title>Re: Add Spark Configurations Serverless Compute</title>
      <link>https://community.databricks.com/t5/data-engineering/add-spark-configurations-serverless-compute/m-p/123772#M47065</link>
      <description>&lt;P&gt;Hi &lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/172437"&gt;@mac_delvalle&lt;/a&gt;&amp;nbsp;,&lt;/P&gt;&lt;P&gt;I'm afraid that when it comes to serverless compute your options are kind of limited.&lt;SPAN&gt;Severless compute does not support setting most Spark properties for notebooks or job.&lt;/SPAN&gt;&amp;nbsp;According to following documentation entry: "Severless compute does not support setting most Spark properties for notebooks or job"&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;A href="https://learn.microsoft.com/en-us/azure/databricks/spark/conf#configure-spark-properties-for-serverless-notebooks-and-jobs" target="_blank" rel="noopener"&gt;https://learn.microsoft.com/en-us/azure/databricks/spark/conf#configure-spark-properties-for-serverless-notebooks-and-jobs&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 02 Jul 2025 18:34:23 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/add-spark-configurations-serverless-compute/m-p/123772#M47065</guid>
      <dc:creator>szymon_dybczak</dc:creator>
      <dc:date>2025-07-02T18:34:23Z</dc:date>
    </item>
    <item>
      <title>Re: Add Spark Configurations Serverless Compute</title>
      <link>https://community.databricks.com/t5/data-engineering/add-spark-configurations-serverless-compute/m-p/125323#M47419</link>
      <description>&lt;P&gt;Thank you for your reply. Hopefully they add this functionality in the future.&lt;/P&gt;</description>
      <pubDate>Tue, 15 Jul 2025 15:00:09 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/add-spark-configurations-serverless-compute/m-p/125323#M47419</guid>
      <dc:creator>mac_delvalle</dc:creator>
      <dc:date>2025-07-15T15:00:09Z</dc:date>
    </item>
    <item>
      <title>Re: Add Spark Configurations Serverless Compute</title>
      <link>https://community.databricks.com/t5/data-engineering/add-spark-configurations-serverless-compute/m-p/125324#M47420</link>
      <description>&lt;P&gt;I think you will not be able to set spark configurations in cluster in serverless. But you can put this in notebook.&lt;/P&gt;&lt;LI-CODE lang="python"&gt;spark.conf.set(
  "spark.sql.sources.partitionOverwriteMode", "dynamic"
)&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Tue, 15 Jul 2025 15:09:36 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/add-spark-configurations-serverless-compute/m-p/125324#M47420</guid>
      <dc:creator>nayan_wylde</dc:creator>
      <dc:date>2025-07-15T15:09:36Z</dc:date>
    </item>
    <item>
      <title>Re: Add Spark Configurations Serverless Compute</title>
      <link>https://community.databricks.com/t5/data-engineering/add-spark-configurations-serverless-compute/m-p/125328#M47421</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/172437"&gt;@mac_delvalle&lt;/a&gt;&amp;nbsp;,&lt;/P&gt;&lt;P&gt;If any of the answers were helpful to you, consider marking it as an accepted solution. This way, the next person with a similar question will be able to find the solution to their problem more quickly.&lt;/P&gt;</description>
      <pubDate>Tue, 15 Jul 2025 15:32:46 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/add-spark-configurations-serverless-compute/m-p/125328#M47421</guid>
      <dc:creator>szymon_dybczak</dc:creator>
      <dc:date>2025-07-15T15:32:46Z</dc:date>
    </item>
  </channel>
</rss>

