cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

External dependency on serverless job from Airflow is not working on s3 path and workspace

manojpatil04
New Contributor III

I am working on use case where we have to run python script from serverless job through Airflow. when we are trying to trigger serverless job and passing external dependency as wheel from s3 path or workspace path it is not working, but on volume it is working.

Is serverless job currently not supporting dependencies from s3 path or workspace.

"environment_key": "Test_environment" }],
                "environments": [
                {
                "environment_key": "Test_environment",
                "spec":
                {
                "client": "1",
                "dependencies":["wheel file path from s3 or workspace"]
                }
                }],
5 REPLIES 5

Walter_C
Databricks Employee
Databricks Employee

It seems that is should be supported, are you using the following format for the URI: { "whl": "s3://my-bucket/library.whl" }?

I am using these external dependencies as below,

"environment_key""Test_environment",
                "spec":
                {
                "client""1",
                "dependencies":["wheel file path from s3 or workspace"]
                }
                }],
for volumes it is working, but for S3 and workspace it is not working, we don't need to specifically add "whl" in the dependency.

Walter_C
Databricks Employee
Databricks Employee

And does the compute has proper permissions to access the S3?

ColeBerl_37881
New Contributor II

I am having a similar issue. I am on Azure, running a Python Wheel workflow from a whl file stored in a Volume. The workflow runs successfully on a normal job cluster, but fails on serverless with the message `run failed with error message Unexpected error during library installation`. Almost impossible to debug due to the lack of logs when using serverless. 

Walter_C
Databricks Employee
Databricks Employee

As per serverless compute limitations I can see the following: 

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.

If there isn’t a group near you, start one and help create a community that brings people together.

Request a New Group