- 1518 Views
- 3 replies
- 0 kudos
When using spot fleet pools to schedule jobs, driver and worker nodes are provisioned from the spot pools and we are noticing jobs failing with the below exception when there is a driver spot loss. Share best practices around using fleet pools with 1...
- 1518 Views
- 3 replies
- 0 kudos
Latest Reply
In this scenario, the driver node is reclaimed by AWS. Databricks started preview of hybrid pools feature which would allow you to provision driver node from a different pool. We recommend using on-demand pool for driver node to improve reliability i...
2 More Replies
- 1338 Views
- 1 replies
- 0 kudos
I have a bunch of libraries that I want to uninstall. All of them are marked as auto-install.
- 1338 Views
- 1 replies
- 0 kudos
Latest Reply
1) Find the corresponding library definition from an existing cluster using "libraries/cluster-status?cluster_id=<cluster_id>".$ curl -X GET 'https://cust-success.cloud.databricks.com/api/2.0/libraries/cluster-status?cluster_id=1226-232931-cuffs129' ...
- 1750 Views
- 1 replies
- 0 kudos
I had a cluster that I used in the past. I do not see the cluster any longer. I checked with the admin and my team and everyone confirmed that there no user deletion.
- 1750 Views
- 1 replies
- 0 kudos
Latest Reply
If the cluster is unsued for 30 days, Databricks removes the cluster. This is a general clean-up policy. It's possible to whitelist a cluster from this clean-up by Pinning the cluster. https://docs.databricks.com/clusters/clusters-manage.html#pin-a-c...
- 1292 Views
- 1 replies
- 0 kudos
Whenever I restart a Databricks cluster new instances are not launched. This is because Databricks re-uses the instances. However, sometimes it's needed to launch new instances. Some scenarios are to mitigate a bad VM issue or maybe to get a patch fr...
- 1292 Views
- 1 replies
- 0 kudos
Latest Reply
Currently, there is no direct option to restart the cluster with new instances. An easy hack to ensure new instances are launched is to add Cluster tags on your cluster. This will ensure that new instances have to be acquired as it's not possible to ...
- 1022 Views
- 1 replies
- 1 kudos
On the Databricks cluster UI, when I click on the Driver logs, sometimes I see historic logs and sometimes I see logs for the last few hours. Why do we see this inconsistency
- 1022 Views
- 1 replies
- 1 kudos
Latest Reply
This is working per design! This is the expected behavior. When the cluster is in terminated state, the logs are serviced by the Spark History server hosted on the Databricks control plane. When the cluster is up and running the logs are serviced by ...
- 1053 Views
- 1 replies
- 0 kudos
The cluster is Idle and there are no Spark jobs running on the Spark UI. Still I see my cluster is active and not getting terminated.
- 1053 Views
- 1 replies
- 0 kudos
Latest Reply
Databricks cluster is treated as active if there are any spark or non-Spark operations running on the cluster. Even though there are no Spark jobs running on the cluster, it's possible to have some driver-specific application code running marking th...
- 2678 Views
- 1 replies
- 1 kudos
Some of the Jobs are failing in prod with below error message:Can you please check and let us know the reason for this? These are running under pool cluster.Run result unavailable: job failed with error messageUnexpected failure while waiting for the...
- 2678 Views
- 1 replies
- 1 kudos
Latest Reply
@Kunal Gaurav​ , This status code only occurs in one of two conditions:We’re able to request the instances for the cluster but can’t bootstrap them in time We setup the containers on each instance, but can’t start the containers in timethis is an edg...
- 1346 Views
- 1 replies
- 0 kudos
It's governed by Azure subscription limits.
- 1346 Views
- 1 replies
- 0 kudos
Latest Reply
In addition to subscription limits, the total capacity of clusters in each workspace is a function of the masks used for the workspace's enclosing Vnet and the pair of subnets associated with each cluster in the workspace. The masks can be changed if...
- 666 Views
- 1 replies
- 0 kudos
What are the best practices around Z ordering, Should be include as Manu column as Possible in Z order or lesser the better and why?
- 666 Views
- 1 replies
- 0 kudos
Latest Reply
With Z-order and Hilbert curves, the effectiveness of clustering decreases with each column added - so you'd want to zorder only the columns that you's actually use so that it's speed up your workloads.