Hi,

I also use azue data factory to run databricks notebook as with data factory you can nice handle many data flow scenarios depend on task success/failure/completion/timeout etc. -- Can't we implement this type of scenario in Databricks using Multiple tasks job.

Thank you