How to get databricks performance metrics programmatically?
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
04-01-2024 09:52 PM
How to retrieve all Databricks performance metrics on an hourly basis. Is there a recommended method or API available for retrieving performance metrics ?
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
04-02-2024 07:31 PM
Hi @Retired_mod
I encountered the "Error: getaddrinfo ENOTFOUND http" while attempting to run the API.
Could you please provide guidance on resolving this issue? Are there any prerequisites that need to be followed before running the API? I'd appreciate any assistance you can offer.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
04-02-2024 05:38 AM
Thanks @Retired_mod for your response.
Could you please provide guidance on how we can achieve the step3 in AWS services?
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
04-03-2024 02:02 AM
Hi @Nandhini_Kumar, there's many performance metrics available - it depends on what you're looking to do with this data, and how you look to take action in real time. I would strongly recommend mapping out a user journey so you get only the metrics you need, and don't waste time trying to get these all hourly when they won't be used.
- Cluster metrics - you can export these with cluster logging. It's worth noting that ganglia is deprecated for newer runtimes
- Warehouse metrics - available through the API for query metrics
- System tables - if you're using unity catalog, this will update hourly
- Jobs performance - you can use the Jobs API
API docs can be found here: https://docs.databricks.com/api/workspace/introduction
System tables: https://docs.databricks.com/en/administration-guide/system-tables/index.html
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
11-06-2024 05:51 AM
Any response here? Are there any API that expose this? Specially the Job performance metrics?
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
11-18-2024 02:59 AM
Hi @rahuja can you be more specific on 'job performance'? Do you mean the spark metrics? Whether it was successful? How much it cost?
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
11-18-2024 04:34 AM
@holly yes we need spark metrics e.g: amount of Compute used, amount of memory used etc.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
11-21-2024 06:19 AM
The spark logs are available through cluster logging. This is enabled at the cluster level for you to choose the destination for the logs.
Just a heads up - interpreting them at scale is not trivial. I'd recommend having a read through the overwatch project to see how they're aggregated together.

