cancel
Showing results for 
Search instead for 
Did you mean: 
Administration & Architecture
cancel
Showing results for 
Search instead for 
Did you mean: 

How to determine if am using the same DBR minor version?

User16869510359
Esteemed Contributor

DBR minor version details are not exposed. However, in the documentation, it mentioned Databricks performs maintenance releases every 2 weeks. How can I determine if I am using the same minor version

1 ACCEPTED SOLUTION

Accepted Solutions

User16869510359
Esteemed Contributor

The below code snippet can help to determine the DBR Hash string for the DBR version. DBR hash string is unique for the DBR minor version.

val scalaVersion = scala.util.Properties.versionString
 
val hadoopVersion = org.apache.hadoop.util.VersionInfo.getVersion
 
val baseVersion = org.apache.spark.BuildInfo.sparkBranch
 
val baseSha = org.apache.spark.BuildInfo.apacheBase
 
val dbrSha = org.apache.spark.BuildInfo.gitHash
 
val universeSha = com.databricks.BuildInfo.gitHash
 
val universeTimestamp = com.databricks.BuildInfo.gitTimestamp
 
val imageId = spark.conf.get("spark.databricks.clusterUsageTags.sparkVersion")
 
 

View solution in original post

1 REPLY 1

User16869510359
Esteemed Contributor

The below code snippet can help to determine the DBR Hash string for the DBR version. DBR hash string is unique for the DBR minor version.

val scalaVersion = scala.util.Properties.versionString
 
val hadoopVersion = org.apache.hadoop.util.VersionInfo.getVersion
 
val baseVersion = org.apache.spark.BuildInfo.sparkBranch
 
val baseSha = org.apache.spark.BuildInfo.apacheBase
 
val dbrSha = org.apache.spark.BuildInfo.gitHash
 
val universeSha = com.databricks.BuildInfo.gitHash
 
val universeTimestamp = com.databricks.BuildInfo.gitTimestamp
 
val imageId = spark.conf.get("spark.databricks.clusterUsageTags.sparkVersion")