Starting yesterday afternoon, my job clusters across different workstations started throwing an error when importing from pypi the MLFlow library upon cluster initiation and startup.
I'm using an Azure Databricks automated job cluster (details below) and installing MLFlow (mlflow==1.26.1) as one of several dependent libraries via pypi. I additionally tried changing the MLFlow version, which did not change the result, and tried not specifying a version at all, which also did not work.
These jobs and clusters were working the previous day. Any troubleshoot suggestions is much appreciated.
Cluster details:
Driver: Standard_DS5_v2
Workers: Standard_DS5_v2 ยท 8 workers ยท
7.3 LTS (includes Apache Spark 3.0.1, Scala 2.12)
Error message:
Run result unavailable: job failed with error message
Library installation failed for library due to user error for pypi {
package: "mlflow==1.26.1"
}
. Error messages:
Library installation attempted on the driver node of cluster 0208-140630-58jkle3z and failed. Please refer to the following error message to fix the library or contact Databricks support. Error Code: DRIVER_LIBRARY_INSTALLATION_FAILURE. Error Message: org.apache.spark.SparkException: Process List(/databricks/python/bin/pip, install, mlflow==1.26.1, --disable-pip-version-check) exited with code 1. ERROR: Command errored out with exit status 1:
command: /databricks/python3/bin/python3.7 /databricks/python3/lib/python3.7/site-packages/pip/_vendor/pep517/_in_process.py get_requires_for_build_wheel /tmp/tmpiooih4q6
cwd: /tmp/pip-install-ffemy0b0/alembic
Complete output (16 lines):
Traceback (most recent call last):
File "/databricks/python3/lib/python3.7/sit ...
***WARNING: message truncated. Skipped 1338 bytes of output**