โ06-02-2025 09:15 PM - edited โ06-02-2025 09:33 PM
Hi Community,
I am using Databricks Serverless compute in notebooks. When I create multiple notebooks and choose Serverless as the compute, I noticed that I can select the same serverless cluster for all of them.
This brings up a few questions:
Is this serverless compute shared across all notebooks (and users), or does each notebook/user get a separate compute instance behind the scenes?
If it is shared, is there a max limit on CPU cores or concurrency for the serverless compute engine?
How does Databricks handle auto-scaling or resource isolation when multiple users are running queries using the same serverless compute?
Thanks in advance for the clarification!
โ06-03-2025 12:55 PM
โ06-03-2025 12:55 PM
โ06-03-2025 09:07 PM - edited โ06-03-2025 09:11 PM
Hi @BigRoux ,
I am using serverless compute for running a hash validation script across a large number of tables. While serverless is supposed to automatically adjust resources based on workload scaling up during peak and scaling down during idle periods I am noticing that the driver gets detached automatically, especially when processing large tables.
1. What is the reason the driver detaches automatically in serverless compute, especially when processing large tables?
2. What is the solution or best practice to prevent this issue and ensure stable processing?
โ06-04-2025 04:19 AM
Could you clarify what you mean by โThe driver detachesโ? If the driver detaches, the cluster would typically fail. Are you using Spark for processing, or is this a pure Python workload? If youโre using pure Python, only the driver node is utilized, since Python doesnโt support distributed execution in this context.
Please provide more details and any demonstrable evidence that the driver is being detached.
Thanks,
Louis
Passionate about hosting events and connecting people? Help us grow a vibrant local communityโsign up today to get started!
Sign Up Now