cancel
Showing results for 
Search instead for 
Did you mean: 
Administration & Architecture
Explore discussions on Databricks administration, deployment strategies, and architectural best practices. Connect with administrators and architects to optimize your Databricks environment for performance, scalability, and security.
cancel
Showing results for 
Search instead for 
Did you mean: 

Databricks All-purpose compute Pricing

karen_c
New Contributor

Hello, I am now struggling how to calculate the cost of my job cluster.

My configuration is as below:

karen_c_0-1733467288235.png

karen_c_1-1733467313502.png

If I have to run the above cluster 18 hours per day, in Standard Tier and East Asia Region, how much will be the pricing of the cluster?

Any help provided is greatly appreciated!! Please provide details if possible!!

Thanks 🙂

1 ACCEPTED SOLUTION

Accepted Solutions

TakuyaOmi
Contributor II

@karen_c 

The cost of Azure Databricks is primarily composed of the following three elements:

  1. Azure Virtual Machine (VM) instance cost: Charged based on the type and size of the VM used for data processing.
  2. Databricks Unit (DBU) cost: The software usage fee for Databricks, which varies depending on the workload type and cluster size.

In summary, the total cost can be calculated as:
Total Cost = (VM Instance Cost + DBU Cost) × Usage Hours

Strictly speaking, additional costs such as storage services are also required, but I’ve omitted them for simplicity.

This pricing can be easily calculated using Azure’s official documentation.

https://azure.microsoft.com/en-us/pricing/details/databricks/#pricing

Based on your specific conditions, I’ve done some research. The key section to look at is the "Pay As You Go Total Price."

It states $1.084/hour. However, since you’re using 1 driver node and 2 worker nodes (a total of 3 nodes), the cost becomes $3.252/hour. Running this for 18 hours would result in a total of $58.536.

One important note is that enabling Photon affects the DBU cost, which in turn impacts the overall cost. When calculating, be sure to select "with Photon," as shown in the attached image.

Please double-check my calculations, as there’s a chance I might have made a mistake. I hope this information is helpful to you!

TakuyaOmi_1-1733475597267.png

 

 

View solution in original post

2 REPLIES 2

TakuyaOmi
Contributor II

@karen_c 

The cost of Azure Databricks is primarily composed of the following three elements:

  1. Azure Virtual Machine (VM) instance cost: Charged based on the type and size of the VM used for data processing.
  2. Databricks Unit (DBU) cost: The software usage fee for Databricks, which varies depending on the workload type and cluster size.

In summary, the total cost can be calculated as:
Total Cost = (VM Instance Cost + DBU Cost) × Usage Hours

Strictly speaking, additional costs such as storage services are also required, but I’ve omitted them for simplicity.

This pricing can be easily calculated using Azure’s official documentation.

https://azure.microsoft.com/en-us/pricing/details/databricks/#pricing

Based on your specific conditions, I’ve done some research. The key section to look at is the "Pay As You Go Total Price."

It states $1.084/hour. However, since you’re using 1 driver node and 2 worker nodes (a total of 3 nodes), the cost becomes $3.252/hour. Running this for 18 hours would result in a total of $58.536.

One important note is that enabling Photon affects the DBU cost, which in turn impacts the overall cost. When calculating, be sure to select "with Photon," as shown in the attached image.

Please double-check my calculations, as there’s a chance I might have made a mistake. I hope this information is helpful to you!

TakuyaOmi_1-1733475597267.png

 

 

TakuyaOmi
Contributor II

@karen_c 

Let me make a small correction.

It seems that you have checked the option for Spot Instances, which should make the cost slightly lower. Please refer to the far-right column of the attached pricing table for more details.

Additionally, you have set your cluster to autoscale, allowing worker nodes to scale up to a maximum of 4 nodes. My calculation was based on the assumption that the cluster runs with a fixed 2 worker nodes, without scaling. Please keep this difference in mind as well.

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.

If there isn’t a group near you, start one and help create a community that brings people together.

Request a New Group