09-05-2023 06:53 AM
Dear,
In the current setup, we are using dbt as a modeling tool for our data lakehouse.
For a specific use case, we want to use the insert_overwrite strategy, where dbt will replace all data for a specific partition:
Databricks configurations | dbt Developer Hub (getdbt.com)
Hereby the specific dbt model configuration:
{{ config( materialized='incremental', partition_by=['zcFiscalYearPeriod'], file_format='delta', incremental_strategy='insert_overwrite', on_schema_change='sync_all_columns' ) }} |
When dbt is execution the needed query’s, we retrieve the following message:
Qry:
/* {"app": "dbt", "dbt_version": "1.5.6", "dbt_databricks_version": "1.5.5", "databricks_sql_connector_version": "2.7.0", "profile_name": "user", "target_name": "default", "node_id": "model.data_platform.prep_InventoryStockDay_test"} */ set spark.sql.sources.partitionOverwriteMode = DYNAMIC
|
Msg:
Configuration spark.sql.sources.partitionOverwriteMode is not available. |
Maybe important to note that we are connecting via a sql endpoint, to a unity-catalog enabled cluster.
We already contacted the DBT-team and everything looks good on dbt side, maybe it’s a setting in databricks we need to change?
09-05-2023 07:14 AM
the sql endpoint could be the culprit. dynamic partition overwrite is in public preview on workspace so it could be that the sql endpoints are not yet supported.
What happens if you use replaceWhere instead of insertoverwrite?
09-06-2023 01:14 AM
Hi Werners1,
thank you for the response, using the replaceWhere statement gives us the same message
09-06-2023 01:31 AM
Weird, because replaceWhere does not require the parameter to be set.
Or do you still try to set it to true, because that is not a requirement for replaceWhere
10-13-2023 04:02 AM
Did you ever get a solution or answer on this error? Apparently partition overwrites are no longer a preview feature, they were released in version Databricks SQL version 2023.40 (see link below). My SQL warehouse is using 2023.40 but I am still getting the same "spark.sql.sources.partitionOverwriteMode is not available" error as you via dbt.
https://docs.databricks.com/en/sql/release-notes/index.html
Block schema overwrite is available when using dynamic partition overwrites.
10-16-2023 12:58 AM
Hi, it seems that this is not supported via sql warehousing (ref. dbt code)
01-24-2024 04:09 AM
Hi!
I have same issue with insert_overwrite on Databricks with SQL Warehouse. Do you have any solution or updates? Or is it still not supported by Databricks?
01-24-2024 05:19 AM
Hi @nad__ , This is still not available in sql warehouse
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.
If there isn’t a group near you, start one and help create a community that brings people together.
Request a New Group