04-02-2025 10:35 AM
Hi Everyone,
I am currently facing an issue with in our Test Environment where Data bricks is not able to mount with the storage account and we are using the same mount in other environments those are Dev,Preprod and Prod and it works fine there without any errors.
I have validated below settings
Fyi, We are using a secret scope in Data bricks and from there we have been fetching these details through a keyvault using the same SP details across all environments and the secret is updated and in sync.
Required STORAGE BLOB DATA CONTRIBUTOR ROLE is assigned to the SP in Storage account.
Network is not restricted and it got ENABLED for all public networks and for the Cluster.
We have created a new SP and tried with it no luck.
SOFT DELETION is enabled for the storage account for all our storage accounts in other environments and it is working fine.
Can someone please through some light on this issue
Still getting following authentication java lang exception error 480
Thanks
2 weeks ago
This issue in your Test environment, where Databricks fails to mount an Azure Storage account with the error java.lang.Exception: 480, is most likely related to expired credentials or cached authentication tokens, even though the same configuration works in Dev, Preprod, and Prod.
Based on technical documentation and recent Databricks/Microsoft support discussions , here are the most probable causes and solutions:
Expired or Invalid Client Secret
The Service Principal (SP) secret configured in Azure Key Vault might have expired or become invalid in the Test workspace context. Even if the same secret is updated across environments, cached mounts in Databricks can hold old tokens, causing authentication to fail.
Stale or Conflicting Mount
Previous mounts in the Test environment might still be holding old credentials. Databricks does not automatically refresh tokens for existing mounts, so an invalid cached credential could trigger the 480 error.
Configuration or Permission Drift
The SP or role assignment may appear identical but differ slightly due to propagation delays or hidden policy differences between subscriptions. Double-check the IAM scope—ensure the Storage Blob Data Contributor role is applied at the storage account level, not just at the container level.
Soft Delete or Hierarchical Namespace Issues
In rare cases, the soft delete feature can interfere with mounting operations if remnants of a previous mount container still exist and conflict with current tokens. This edge case has been observed when remounting after a cluster or workspace reset.
Cluster Token Expiry
If the Databricks cluster was running for an extended period, its internal access token used to reach Azure AD might have expired. Restarting the cluster refreshes this token and can often fix transient authentication exceptions.
Follow these actions in order:
Recreate Client Secret (if older than 90 days)
In Azure Entra ID, generate a new secret for the Service Principal.
Update the secret in your Key Vault and refresh it in your Databricks Secret Scope.
Unmount and Remount the Path
Run:
dbutils.fs.unmount("/mnt/<your-mount-path>")
dbutils.fs.refreshMounts()
Then remount using updated credentials:
dbutils.fs.mount(
source="abfss://<container>@<storage-account>.dfs.core.windows.net/",
mount_point="/mnt/<your-mount-path>",
extra_configs={
"fs.azure.account.auth.type.<storage-account>.dfs.core.windows.net": "OAuth",
"fs.azure.account.oauth.provider.type.<storage-account>.dfs.core.windows.net": "org.apache.hadoop.fs.azurebfs.oauth2.ClientCredsTokenProvider",
"fs.azure.account.oauth2.client.id.<storage-account>.dfs.core.windows.net": dbutils.secrets.get(scope="<scope>", key="sp-client-id"),
"fs.azure.account.oauth2.client.secret.<storage-account>.dfs.core.windows.net": dbutils.secrets.get(scope="<scope>", key="sp-client-secret"),
"fs.azure.account.oauth2.client.endpoint.<storage-account>.dfs.core.windows.net": "https://login.microsoftonline.com/<tenant-id>/oauth2/token"
}
)
Restart the Databricks Cluster
This refreshes any stale authentication caches.
Validate Connectivity
Test using Azure Storage Explorer with the same SP credentials.
Verify the Databricks workspace subnet (if using managed VNet) can access your storage endpoint.
Avoid Nested Mount Points
Ensure you're not creating a mount inside another mount, which can throw authentication or path-related exceptions.
If the issue persists after these steps, enabling Databricks debug logs (spark.databricks.service.server.logLevel=DEBUG) can help identify whether token acquisition or blob authentication is failing specifically.
In most reported 480 cases from 2025 , unmounting and remounting with a refreshed SP secret and restarting the cluster resolved the issue fully in test or QA environments.
05-01-2025 12:00 AM
Checking.
2 weeks ago
This issue in your Test environment, where Databricks fails to mount an Azure Storage account with the error java.lang.Exception: 480, is most likely related to expired credentials or cached authentication tokens, even though the same configuration works in Dev, Preprod, and Prod.
Based on technical documentation and recent Databricks/Microsoft support discussions , here are the most probable causes and solutions:
Expired or Invalid Client Secret
The Service Principal (SP) secret configured in Azure Key Vault might have expired or become invalid in the Test workspace context. Even if the same secret is updated across environments, cached mounts in Databricks can hold old tokens, causing authentication to fail.
Stale or Conflicting Mount
Previous mounts in the Test environment might still be holding old credentials. Databricks does not automatically refresh tokens for existing mounts, so an invalid cached credential could trigger the 480 error.
Configuration or Permission Drift
The SP or role assignment may appear identical but differ slightly due to propagation delays or hidden policy differences between subscriptions. Double-check the IAM scope—ensure the Storage Blob Data Contributor role is applied at the storage account level, not just at the container level.
Soft Delete or Hierarchical Namespace Issues
In rare cases, the soft delete feature can interfere with mounting operations if remnants of a previous mount container still exist and conflict with current tokens. This edge case has been observed when remounting after a cluster or workspace reset.
Cluster Token Expiry
If the Databricks cluster was running for an extended period, its internal access token used to reach Azure AD might have expired. Restarting the cluster refreshes this token and can often fix transient authentication exceptions.
Follow these actions in order:
Recreate Client Secret (if older than 90 days)
In Azure Entra ID, generate a new secret for the Service Principal.
Update the secret in your Key Vault and refresh it in your Databricks Secret Scope.
Unmount and Remount the Path
Run:
dbutils.fs.unmount("/mnt/<your-mount-path>")
dbutils.fs.refreshMounts()
Then remount using updated credentials:
dbutils.fs.mount(
source="abfss://<container>@<storage-account>.dfs.core.windows.net/",
mount_point="/mnt/<your-mount-path>",
extra_configs={
"fs.azure.account.auth.type.<storage-account>.dfs.core.windows.net": "OAuth",
"fs.azure.account.oauth.provider.type.<storage-account>.dfs.core.windows.net": "org.apache.hadoop.fs.azurebfs.oauth2.ClientCredsTokenProvider",
"fs.azure.account.oauth2.client.id.<storage-account>.dfs.core.windows.net": dbutils.secrets.get(scope="<scope>", key="sp-client-id"),
"fs.azure.account.oauth2.client.secret.<storage-account>.dfs.core.windows.net": dbutils.secrets.get(scope="<scope>", key="sp-client-secret"),
"fs.azure.account.oauth2.client.endpoint.<storage-account>.dfs.core.windows.net": "https://login.microsoftonline.com/<tenant-id>/oauth2/token"
}
)
Restart the Databricks Cluster
This refreshes any stale authentication caches.
Validate Connectivity
Test using Azure Storage Explorer with the same SP credentials.
Verify the Databricks workspace subnet (if using managed VNet) can access your storage endpoint.
Avoid Nested Mount Points
Ensure you're not creating a mount inside another mount, which can throw authentication or path-related exceptions.
If the issue persists after these steps, enabling Databricks debug logs (spark.databricks.service.server.logLevel=DEBUG) can help identify whether token acquisition or blob authentication is failing specifically.
In most reported 480 cases from 2025 , unmounting and remounting with a refreshed SP secret and restarting the cluster resolved the issue fully in test or QA environments.
Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!
Sign Up Now