Hi all,
I have a Delta Table as a Spark Streaming source.
This table contains signals on row level -> each signal is one append to the source table that creates a new version in the delta transaction history.
I am not really sure now how Spark streaming works if I define
spark
.readStream
.format("delta")
.option("startingVersion", "latest")
.option("maxFilesPerTrigger ", 100)
Are those the last 100 transactions from the delta transaction history? Or the last 100 parquet files?
Best regards
Michael