01-24-2022 11:34 AM
I created a Databricks job with multiple tasks. Is there a way to pass variable values from one task to another. For example, if I have tasks A and B as Databricks notebooks. Can I create a variable (e.g. x) in notebook A and later use that value in notebook B?
01-25-2022 04:31 AM
01-25-2022 04:31 AM
01-25-2022 07:17 AM
Thank you, Werner! From my understanding, if I put a notebook in %run command, I won't have it as a separate task in the job. However, I'd also like to keep the ability to see it in the task view for transparency. Is there a way to achieve that?
01-25-2022 07:23 AM
hm I don´t think so as actually the notebooks will run within the same task.
01-25-2022 07:26 AM
you could also consider using an orchestration tool like Data Factory (Azure) or Glue (AWS). there you can inject and use parameters from notebooks.
The job scheduling of databricks also has the possibility to add parameters, but I do not know if you can dynamically determine the input (based on another task).
01-25-2022 07:43 AM
Thank you! We're on Azure. Will explore Data Factory!
Join our fast-growing data practitioner and expert community of 80K+ members, ready to discover, help and collaborate together while making meaningful connections.
Click here to register and join today!
Engage in exciting technical discussions, join a group with your peers and meet our Featured Members.