I have a large delta table that I would like to back up and I am wondering what is the best practice for backing it up.
The goal is so that if there is any accidental corruption or data loss either at the Azure blob storage level or within Databricks itself I can restore the data.
Is using the Azure blob "Point-in-time" restore features appropriate? On paper, it sounds like it has all the features I require. However, what is the downstream effect of using it on a delta table and will weekly OPTIMIZE cause rewrites of the data and blow out the costs?
In other Azure/Databricks documentation, there was mention of using Deep Clone for data replication.
Any thoughts appreciated.