cancel
Showing results forย 
Search instead forย 
Did you mean:ย 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results forย 
Search instead forย 
Did you mean:ย 

Databricks driver logs

sagiatul
New Contributor II

I am running jobs on databricks clusters. When the cluster is running I am able to find the executor logs by going to Spark Cluster UI Master dropdown, selecting a worker and going through the stderr logs. However, once the job is finished and cluster terminates, I am unable to see those logs. I get below screen

image

2 REPLIES 2

Debayan
Esteemed Contributor III
Esteemed Contributor III

Hi, when you say the job is finished and the cluster terminates, can you still list the cluster in the cluster UI page? Also, could you please uncheck the auto-fetch logs?

Until and unless the cluster is deleted the logs should be there.

Please tag @Debayanโ€‹ with your next response which will notify me, Thank you!

Anonymous
Not applicable

Hi @Atul Aroraโ€‹ 

Thank you for your question! To assist you better, please take a moment to review the answer and let me know if it best fits your needs.

Please help us select the best solution by clicking on "Select As Best" if it does.

Your feedback will help us ensure that we are providing the best possible service to you.

Thank you!

Join 100K+ Data Experts: Register Now & Grow with Us!

Excited to expand your horizons with us? Click here to Register and begin your journey to success!

Already a member? Login and join your local regional user group! If there isn’t one near you, fill out this form and we’ll create one for you to join!