cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
cancel
Showing results for 
Search instead for 
Did you mean: 

Notebook fails in job but not in interactive mode

-werners-
Esteemed Contributor III

I have this notebook which is scheduled by Data Factory on a daily basis.

It works fine, up to today. All of a sudden I keep on getting NullpointerException when writing the data.

After some searching online, I disabled AQE. But this does not help.

The weird part is: when I run the notebook manually it does work.

So it seems something has changed in the scheduling of ADF or the job interface of Databricks?

Any idea in which direction I should look?

I also tried using fixed clusters (so no autoscaling), no spot instances, different dbrx-versions (9.0, 9,1, 10.0, 10.1).

3 REPLIES 3

-werners-
Esteemed Contributor III

After some tests it seems that if I run the notebook on an interactive cluster, I only get 80% of load (Ganglia metrics).

If I run the same notebook on a job cluster with the same VM types etc (so the only difference is interactive vs job), I get over 150% load => worker crash.

Weird.

Did you try on DBR 7.3 LTS? I think there was an issue with NPE that was affecting some versions but not 7.3.

-werners-
Esteemed Contributor III

Hm bizarre, it works fine now with any version.

Must have been some weird data or something.

Welcome to Databricks Community: Lets learn, network and celebrate together

Join our fast-growing data practitioner and expert community of 80K+ members, ready to discover, help and collaborate together while making meaningful connections. 

Click here to register and join today! 

Engage in exciting technical discussions, join a group with your peers and meet our Featured Members.