Hey guys I'm trying to find the estimate for an ingestion pipeline that uses autoloader on an S3 bucket every 2 minutes.
I found the pricing for s3 bucket api consumption but I am not certain what api calls will autoloader make.
Talking to chatGPT it told me i will have a List call per execution and then a get call per file per execution, but i am not certain of this answer
spark.readStream
.format("cloudFiles")
.option("cloudFiles.format", "json")
.schema(json_schema_bronze)
.load("s3://path")
dfBronze.writeStream \
.format("delta") \
.option("checkpointLocation", checkpoint_dir_path_bronze) \
.outputMode("append") \
.trigger(processingTime="120 second") \
.table(bronze_table)