โ08-06-2025 08:01 AM
Hello!
I have following problem: All my job runs fail when the job uses a library. Even the most basic job (print a string) and the most basic library package (no secondary dependencies, the script does not even import/use the library) fails with `Failed to reach the driver`:
* All my libraries are python wheels
* I use `spark_python_task` (but also tested `python_wheel_task` with same error)
* If I use serverless (same script, same .whl), all works fine
* If I remove the package from the job's library section, all works fine (as said: I don't even import/use the library)
* I also used a different python wheel package and created a wheel using https://docs.databricks.com/aws/en/jobs/how-to/use-python-wheels-in-workflows#step-6-run-the-job-and...
* It doesn't matter if I configure the job via yml in asset bundle or "manually" in the UI
The logs of the clusters don't really help me.
I really appreciate your ideas - thank you!
The script:
The job yml:
โ08-12-2025 12:08 AM
I think I found a (the?) solution. The cluster tried to connect to the legacy Hive Catalog, so I
* set the default catalog for the workspace to the proper catalog
* disabled "Legacy access"
These steps solve my `DriverError`. This log4j error message gave the hint: `Caused by: com.zaxxer.hikari.pool.HikariPool$PoolInitializationException: Failed to initialize pool: Could not connect to address=(host=consolidated-westeuropec2-prod-metastore-3.mysql.database.azure.com)(port=3306)(type=master) : Socket fail to connect to host:consolidated-westeuropec2-prod-metastore-3.mysql.database.azure.com, port:3306. Connect timed out`
I still don't know why installing the wheel made such problems, but I consider my problem now as solved.
โ08-06-2025 08:04 AM
Maybe worth mentioning: If I install the library in a python notebook using
%pip install /Workspace/Shared/code/my_package-0.1-py3-none-any.whl
all works fine.
โ08-07-2025 09:23 PM
My current workaround (I'm surprised that this works) is to install the library via "pypi" (actually: an internal pypi mirror using artifactory) instead of via the .whl file. I would still be interested in a reason of and a solution to the problem, though.
Thanks!
โ08-12-2025 12:08 AM
I think I found a (the?) solution. The cluster tried to connect to the legacy Hive Catalog, so I
* set the default catalog for the workspace to the proper catalog
* disabled "Legacy access"
These steps solve my `DriverError`. This log4j error message gave the hint: `Caused by: com.zaxxer.hikari.pool.HikariPool$PoolInitializationException: Failed to initialize pool: Could not connect to address=(host=consolidated-westeuropec2-prod-metastore-3.mysql.database.azure.com)(port=3306)(type=master) : Socket fail to connect to host:consolidated-westeuropec2-prod-metastore-3.mysql.database.azure.com, port:3306. Connect timed out`
I still don't know why installing the wheel made such problems, but I consider my problem now as solved.
8 hours ago
Hi, can you please share detailed soluton, did you include any config in databricks.yml? I'm also having same issue.