2 weeks ago
Hi All,
I'm after some guidance on how to identify massive (100000%) spikes in bandwidth usage (and related costs) in the azure databricks provisioned/managed resource group storage account & stop them
These blips are adding 30-50% to our monthly costs. This is creating uncertainty in the organisation as we can't accurately predict our usage costs.
Our environment is still very much in the POC stage
I believe it's related to the geo-redundancy and databricks copying massive amounts of data to another region even when the clusters are turned off
I can't dig deeper because "system deny" security settings created by databtricks when provisioning the app lock me out even though I'm the subscription admin
(Azure support have not been helpful - just repeating that "yes, these costs were incurred by the databricks managed storage account")
This is weird because I use externally managed tables for everything linked to a storage account I control. I don't use DBFS for anything, especially not 1000s of gigs worth of data.
1.what is happening?
- the timestamps are on the weekend/overnight - our sql service and clusters are off outside business hours and are set to autostop & doesn't align to custers being turned on or any jobs - all of which are relatively simple and small
our 'raw data/datalake' is 2.9gig
the unity catalog + databricks tables are 54gig
2. why is this happening?
3. what do i do so it doesn't happen in the future?
a week ago
What exactly do you mean by the 'azure databricks provisioned/managed resource group storage account'?
The one that is linked to your UC metastore?
a week ago
Hi @jakubk,
Have you opened a case with MSFT? this looks to required a more in-depth analysis. Do you have an active support team with us?
a week ago
No, it's the 'databricks workspace' environment that's automatically created when you provision a new databricks workspace in azure. The resource group and storage account are locked down to the databricks service acc only. everyone else including the azure global admin are locked out
Yes, but the support i received was limited to 'it was caused by databricks - talk to your databricks admin (me) and turn on logging on the storage acc (not possible as because of the 'system deny' permissions which lock me out despite me being the global azure/subscription admin)
I've reached out to databricks support (Ticket Number: #00571565)
but haven't heard anything back
a week ago
That is strange, I think something went wrong in the deployment of the Databricks environment.
We have 3 databricks workspaces and do not see such behavior.
Any resource deployed on a subscription owned by you is accessible by some kind of admin. There should not be any locked down resource.
Thursday
@-werners- wrote:That is strange, I think something went wrong in the deployment of the Databricks environment.
We have 3 databricks workspaces and do not see such behavior.
Any resource deployed on a subscription owned by you is accessible by some kind of admin. There should not be any locked down resource.
You can't do this on the managed resource group created by Azure Databricks even if you're owner - it's a resource managed by Databricks
thanks
Thursday
OK I understand, the resource group with the workers etc.
Indeed the content of that storage account is not accessible unfortunately.
But this is indeed something to let databricks look into.
Here is the used capacity of our prod env (data lake of 10TB-ish capacity):
Yours seems way off.
Keep us updated.
a week ago
Thanks for opening a case with us, we will have a look at it.
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.
If there isn’t a group near you, start one and help create a community that brings people together.
Request a New Group