Hi All,
I am getting this error, when I am reading the streaming source
Full load - It loads
Incremental load - am facing this error
This is the piece of code that am using
def gim_suppliers_ln():
logger.info("Starting __cn_gim_suppliers")
overall_start_time = time.time()
try:
# Log the start time of the readStream execution
read_stream_start_time = time.time()
logger.info(f"readStream execution started at {read_stream_start_time}")
spark.conf.set("spark.sql.files.ignoreMissingFiles","true")
spark._jvm.com.databricks.sql.transaction.tahoe.DeltaLog.clearCache()
spark.conf.set("spark.databricks.delta.retentionDurationCheck.enabled", "false")
spark.conf.set("spark.databricks.delta.changeDataFeed.timestampOutOfRange.enabled", "true")
# Read from the source streaming table
df= spark.readStream \
.format("delta") \
.option("readChangeFeed", "true") \
.option("startingTimestamp", starting_timestamp) \
.option("endingTimestamp", ending_timestamp) \
.load(primary_delta_table_path_url)
# Log the end time of the readStream execution
read_stream_end_time = time.time()
logger.info(f"readStream execution completed at {read_stream_end_time}")
logger.info(f"readStream execution took {read_stream_end_time - read_stream_start_time} seconds")
Any help would be appreciated