cancel
Showing results forย 
Search instead forย 
Did you mean:ย 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results forย 
Search instead forย 
Did you mean:ย 

Py4JJavaError: An error occurred while calling o465.coun

sinclair
New Contributor II

The following error occured when running .count() on a big sparkDF. 

Py4JJavaError: An error occurred while calling o465.count. : org.apache.spark.SparkException: Job aborted due to stage failure: Task 6 in stage 3.0 failed 4 times, most recent failure: Lost task 6.3 in stage 3.0 (TID 32) (10.139.64.6 executor 0): java.lang.NullPointerException

Data has close to 1.2mil records and have some nested json objects, so it's pretty big. I am using  cluster with 64GB memory and 16GB cores and 16gb worker with 4gb cores. 

Is there a way to solve this count? Is increasing cluster size the only solution?

6 REPLIES 6

iakshaykr
New Contributor III

 

DBR : ? 
Number of workers: ? 
number of driver node : assuming 1 driver

sinclair
New Contributor II

DBR: 13.3

Min 2 and max 8 workers

Driver node is 1 

p4pratikjain
Contributor

can you provide complete stacktrace ?

Pratik Jain

iakshaykr
New Contributor III

@sinclair - Can you please use num_workers=2 or 5 which will be constant instead of using auto scaling ? 

Meantime i am trying to reproduce the same issue in my end ! 

szymon_dybczak
Contributor III

Hi @sinclair,

Maybe try to increase number of partitions. There are no enough details, but assuming this error occur on worker due to insufficient memory, increasing partition number can help

Rishabh_Tiwari
Databricks Employee
Databricks Employee

Hi @sinclair ,

Thank you for reaching out to our community! We're here to help you. 

To ensure we provide you with the best support, could you please take a moment to review the response and choose the one that best answers your question? Your feedback not only helps us assist you better but also benefits other community members who may have similar questions in the future.

If you found the answer helpful, consider giving it a kudo. If the response fully addresses your question, please mark it as the accepted solution. This will help us close the thread and ensure your question is resolved.

We appreciate your participation and are here to assist you further if you need it!

Thanks,

Rishabh

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you wonโ€™t want to miss the chance to attend and share knowledge.

If there isnโ€™t a group near you, start one and help create a community that brings people together.

Request a New Group