Deleting data programmatically from databricks live delta tables
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
06-14-2024 06:12 AM
Hello all,
I am relatively new in data engineering and working on a project requiring me to programmatically delete data from delta live tables. However, I found that simply stopping the streaming job and deleting rows from the delta tables caused the stream to fail once I restarted it. The only solution seems to create a new checkpoint for the stream to write to after the deletion or to delete all the entries in the parquet files. Are these the correct solutions to this problem? Which solution do people employ in such cases? Whenever I need to delete data, will I need to create a new checkpoint location or possibly parse billions of parquet records and delete their entries?
Thanks !
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
06-14-2024 12:43 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
06-19-2024 10:34 AM
Hi @shan_chandra Full refreshing brings back the deleted data since it exists in the pubsub source.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
06-20-2024 08:58 AM
@aozero - restart the pipeline with the new checkpoint.
![](/skins/images/B38AF44D4BD6CE643D2A527BE673CCF6/responsive_peak/images/icon_anonymous_message.png)
![](/skins/images/B38AF44D4BD6CE643D2A527BE673CCF6/responsive_peak/images/icon_anonymous_message.png)