12-01-2025 11:43 PM - edited 12-01-2025 11:45 PM
I am encountering multiple issues in our Databricks environment and would appreciate guidance or best-practice recommendations for each. Details below:
Error:
Suspecting that Spark contexts are not being released properly.
Multiple scheduled notebooks may be causing accumulation.
Questions:
Common causes of hitting this 150 SparkContext limit?
How to inspect which jobs/notebooks are holding open contexts?
Any cleanup patterns or cluster settings recommended?
We trigger ~20 notebooks at the same time on the same cluster.
Questions:
Any Databricks concurrency limits at the cluster/job level?
How to throttle or queue notebook runs?
We’re hitting a request size restriction (~10,000 characters) when interacting with Databricks API.
Questions:
What is the official request/response size limit?
Is the 10k cap configurable?
Looking for:
Explanation of why these happen
How to diagnose root causes
Recommended best practices for preventing them
Any guidance or references to Databricks documentation would be very helpful.
12-02-2025 12:21 AM
Hello @adhi_databricks ,
Good Day! Below are the answers to your questions:
12-02-2025 12:21 AM
Hello @adhi_databricks ,
Good Day! Below are the answers to your questions:
12-02-2025 05:35 AM - edited 12-02-2025 05:35 AM
I would like to add my experience with 3. Databricks API 10k Character Limit
We had a similar issue, and this limit cannot be changed. Instead review concepts of sharing the input/output between Databricks and caller using cloud storage like ADLS. Provide ADLS URLs as input and output; this way we are not limited by size of payload.