Yup that's exactly my current plan but there's an issue here. I will explain:
I have 1 million rows in my table
Let's say I give row# 1 to row# 500000 to 1st notebook
and
I give row# 500001 to row# 1 Million to 2nd Notebook.
What if the data is such that the first half (row# 1 to row# 500000) takes 1/10th the time for processing when compared to 2nd half (row# 500001 to row# 1 Million) ?
You see how this can mean Notebook 1 will finish way before Notebook 2. Ideally both Notebooks should run for equal time (around the same time) to finish my activity the fastest..
Hence predetermining the datasets for each notebook is not efficient. The notebooks should dynamically ingest a new batch(of 300 rows) as soon as it finishes the current batch. But my problem is both notebooks might end up ingesting the same batch .
Let me know if that makes sense ๐