Hi @Michaล ,
I'm assuming you're talking about streaming table here. I think what you're trying to achieve is not possible. It's not limitation of lakeflow declarative pipelines per se but rather how spark structured streaming works. All rows will be processed on first run and you can influence micro-batch size with options such as maxBytesPerTrigger or maxFilesPerTrigger.
So, you can't set it up like this:
- load only 10 files and shutdown
- then on the next load consume another 10 files and shutdown
What you can do is to setup size of a micro-batch to i.e 10 files per tirgger:
- all available data will be consumed and at most 10 file will be an input for micro-batch
One more note here. In the case of a declarative pipeline, the remaining question for verification is whether in declarative pipelines you can set properties such as maxFilesPerTrigger or maxBytesPerTrigger. Since it is a declarative framework, it is possible that in this case the options are limited