Hi all,
I am currently running a job that will upsert a table by reading from delta change data feed from my silver table. Here is the relevent snippet of code:
 
 
rds_changes = spark.read.format("delta") \
  .option("readChangeFeed", "true") \
  .option("startingVersion", 0) \
  .table("main.default.gold_table") \
  .where(f"_commit_timestamp >= '{(datetime.now() - timedelta(hours=1)).strftime('%Y-%m-%d %H:%M:%S')}'")
 
 
Here is the error returned
 
 
com.databricks.sql.transaction.tahoe.DeltaFileNotFoundException: No file found in the directory: s3://databricks-workspace-stack-70da1-metastore-bucket/60ed403c-0a54-4f42-8b8a-73b8cea1bdc3/tables/6d4a9b3d-f88b-436e-be1b-09852f605f4c/_delta_log.
 
 
 I have done the following:
- Verify that the delta log folder is not deleted by accessing S3 directly
 - Able to query the table directly and perform `DESCRIBE HISTORY gold_table` on it without any issue
 
Anyone has any idea why this happen when I am running the job which was working fine previously without any changes