01-15-2022 01:18 AM
HI,
I am running a Notebook job calling a JAR code (application code implmented in C#). in the Spark UI page for almost 2 hrs, it'w not showing any tasks and even the CPU usage is below 20%, memory usage is very small. Before this 2 hr window it shows taks and CPU usage.
from application code, it is writing to 2 tables. and before the 2 hr no task window, the data was already written to 1st table. after after this 2 hr window, it again started showing tasks for writing to 2 table.
job is running on 7.3 LTS with total 300 core and should be processing less than 40 million rows
Is this something UX issue Or actually the job was not doing any task. Attached is how UX shows.
I am trying to improve the performance, so any pointers on how to understand and optimize what job is doing this period will be helpful.
01-25-2022 09:03 PM
Thus could be the issue with spark UI not syncing up with spark job. May be you can take a look at this - https://docs.microsoft.com/en-us/azure/databricks/kb/scala/spark-ui-not-in-sync-with-job
01-15-2022 07:42 AM
in my opinion is problem with your code and architecture. It is not standard one so I think you need to keep debugging it.
01-15-2022 09:11 AM
Thanksfor reply. Any pointer where/what should be looking for to troubleshoot this.
01-25-2022 09:03 PM
Thus could be the issue with spark UI not syncing up with spark job. May be you can take a look at this - https://docs.microsoft.com/en-us/azure/databricks/kb/scala/spark-ui-not-in-sync-with-job
01-25-2022 09:03 PM
If I understood the issue correctly .
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.
If there isn’t a group near you, start one and help create a community that brings people together.
Request a New Group