01-07-2022 01:39 AM
I am currently working with a VNET injected databricks workspace. At the moment I have mounted a the databricks cluster on an ADLS G2 resource.
When running notebooks on a single node that read, transform, and write data we do not encounter any problems. However, when I run the same notebook on a multi-node cluster the spark job enters into an infinite waiting stage.
When looking deeper into the Spark cluster UI I find an active stage: "load at NativeMethodAccessorImpl.java:0 (Stage 2.0)". When going to the stage page we get a 403 error - Invalid or missing CSRF token.
Interestingly, for testing purposes we've deployed a databricks workspace without VNET injection. There we make use of a multi-node clusters without any issues.
01-10-2022 01:30 AM
Could it be your whitelisted firewall IP-ranges? wrong CIDR-interval?
Every note will have a separate IP-address, so if your firewall is not configured correctly, the nodes cannot communicate.
01-10-2022 01:30 AM
Could it be your whitelisted firewall IP-ranges? wrong CIDR-interval?
Every note will have a separate IP-address, so if your firewall is not configured correctly, the nodes cannot communicate.
01-10-2022 11:38 PM
@Werner Stinckens thank you for your reply. I will take a look into the netwerk configurations today.
05-07-2024 11:46 PM
@TheDataDexter Did you find a solution to your problem? I am facing the same issue
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.
If there isn’t a group near you, start one and help create a community that brings people together.
Request a New Group