ā08-02-2025 02:12 PM
Hello community! š Iām new to Databricks and currently working on a project structured in Bronze / Silver / Gold layers using Delta Lake and Change Data Feed.
I recently added 3 new columns to a table and initially applied these changes via PySpark SQL commands within our generic job files that handle streaming between layers. I later realized this might not be the best approach, so I dropped and re-added the columns using a Databricks notebook instead.
However, now my job is failing without a specific error. The job run shows the following message:
Has anyone encountered something similar? What might cause this issue or how can I further troubleshoot it?
Any guidance would be appreciated. Thanks!
ā08-06-2025 05:04 AM
Hi @leticialima__ ,
The failure is likely due to a non-additive schema change, such as dropping and re-adding columns. To handle such changes, you can set the schemaTrackingLocation option in your readStream query.
Also, ensure that column mapping is enabled on the table.
ā08-14-2025 05:25 AM
ā08-02-2025 07:45 PM
Hello @leticialima__
Good day
Can you please share the error observed on the driver log.
is it : [Errno 13] Permission denied or No such file or directory? Please let me know the error on the driver log.
THank you.
ā08-04-2025 05:04 AM - edited ā08-04-2025 05:30 AM
Hello, I gave up trying to solve it on Saturday, and today when I checked the failure log of the scheduled job, a new error appeared. the link in the error: https://docs.databricks.com/aws/en/delta/column-mapping
ā
ā08-06-2025 05:04 AM
Hi @leticialima__ ,
The failure is likely due to a non-additive schema change, such as dropping and re-adding columns. To handle such changes, you can set the schemaTrackingLocation option in your readStream query.
Also, ensure that column mapping is enabled on the table.
ā08-14-2025 05:25 AM
Thanks!! It worked!
Passionate about hosting events and connecting people? Help us grow a vibrant local communityāsign up today to get started!
Sign Up Now