I have faced a problem about error listing of _delta_log, when the spark read stream with delta format in GCS.
This is the full log of the issue:
org.apache.spark.sql.streaming.StreamingQueryException: Failed to get result: java.io.IOException: Error listing gs://<bucket_name>/bronze-layer/<database_name>/<table_name>/_delta_log/. reason=Connection closed prematurely: bytesRead = 199316, Content-Length = 1010876 with message : java.io.IOException: Error listing gs://<bucket_name>/bronze-layer/<database_name>/<table_name>/_delta_log/. reason=Connection closed prematurely: bytesRead = 199316, Content-Length = 1010876
Has anyone encountered this issue? I only see it occasionally and not often., I ensure that the databricks environment grants the necessary permissions to access the GCS bucket and set log retention interval 7 days. Currently, this table is really large and have many log files in _delta_log.
On this issue, how do I solve it, and determine the root cause?