cancel
Showing results forย 
Search instead forย 
Did you mean:ย 
Data Engineering
cancel
Showing results forย 
Search instead forย 
Did you mean:ย 

Delta table version protocol

noname123
New Contributor III

I do:

df.write.format("delta").mode("append").partitionBy("timestamp").option("mergeSchema", "true").save(destination)

If table doesn't exist, it creates new table with "minReaderVersion":3,"minWriterVersion":7.
Yesterday it was creating table with "minReaderVersion":1,"minWriterVersion":2. I changed nothing and now versions are different. Setting delta protocol version on cluster does nothing.
This causes breaking change to me because some tables are used by Azure Data Factory and it's not compatible.
1 ACCEPTED SOLUTION

Accepted Solutions

noname123
New Contributor III

Thanks for help.

Issue was caused by "Auto-Enable Deletion Vectors" setting. 

View solution in original post

2 REPLIES 2

Kaniz
Community Manager
Community Manager

Hi @noname123, I understand the issue youโ€™re facing with Delta Lake protocol versions.

Letโ€™s break it down:

  1. Delta Lake Protocol Versions:

  2. Your Scenario:

    • Youโ€™ve noticed that the protocol versions have changed unexpectedly.
    • Previously, the table was created with โ€œminReaderVersionโ€:1 and โ€œminWriterVersionโ€:2.
    • Now, itโ€™s creating tables with โ€œminReaderVersionโ€:3 and โ€œminWriterVersionโ€:7.
    • This discrepancy is causing compatibility issues with Azure Data Factory.
  3. Options and Considerations:

  4. Next Steps:

    • Evaluate the impact of the protocol version change on your existing workflows.
    • If possible, coordinate with other users or teams to ensure compatibility.
    • Consider upgrading Azure Data Factory to a version that supports the new protocol version.
    • If necessary, explore alternative solutions or workarounds based on your specific requirements.

Remember that protocol version changes should be approached carefully, especially in production environments. Itโ€™s essential to balance the benefits of new features with potential disruptions to existing processes. ๐Ÿ› ๏ธ

 

noname123
New Contributor III

Thanks for help.

Issue was caused by "Auto-Enable Deletion Vectors" setting. 

Welcome to Databricks Community: Lets learn, network and celebrate together

Join our fast-growing data practitioner and expert community of 80K+ members, ready to discover, help and collaborate together while making meaningful connections. 

Click here to register and join today! 

Engage in exciting technical discussions, join a group with your peers and meet our Featured Members.