Hi all,
I am currently running a job that will upsert a table by reading from delta change data feed from my silver table. Here is the relevent snippet of code:
rds_changes = spark.read.format("delta") \
.option("readChangeFeed", "true") \
.option("startingVersion", 0) \
.table("main.default.gold_table") \
.where(f"_commit_timestamp >= '{(datetime.now() - timedelta(hours=1)).strftime('%Y-%m-%d %H:%M:%S')}'")
Here is the error returned
com.databricks.sql.transaction.tahoe.DeltaFileNotFoundException: No file found in the directory: s3://databricks-workspace-stack-70da1-metastore-bucket/60ed403c-0a54-4f42-8b8a-73b8cea1bdc3/tables/6d4a9b3d-f88b-436e-be1b-09852f605f4c/_delta_log.
I have done the following:
- Verify that the delta log folder is not deleted by accessing S3 directly
- Able to query the table directly and perform `DESCRIBE HISTORY gold_table` on it without any issue
Anyone has any idea why this happen when I am running the job which was working fine previously without any changes