Hello,
We package our Spark jobs + utilities in a custom package to be used in wheel tasks in Databricks. In my opinion, having several versions of this job (say "production" and "dev") run on the same cluster against different versions of this custom package is a completely valid requirement to facilitate a somewhat resource-friendly CI/CD workflow.
Alas, Databricks does not allow this since wheel libraries end up being installed cluster-wide and only one version of the same library is allowed at a time. To make matter more inconvenient - the cluster needs to be re-started to uninstall a library.
Since we cannot be the only team facing this issue my question is: how to circumvent this shortcoming. Rolling everything into one script - ugly. Notebooks - not an option either.
Thank you,David