cancel
Showing results forย 
Search instead forย 
Did you mean:ย 
Community Platform Discussions
Connect with fellow community members to discuss general topics related to the Databricks platform, industry trends, and best practices. Share experiences, ask questions, and foster collaboration within the community.
cancel
Showing results forย 
Search instead forย 
Did you mean:ย 

databricks-jdbc lists `spark_catalog` among catalogs for Standard tier Azure workspace

HakanNordgren
New Contributor II

databricks-jdbc lists `spark_catalog` among catalogs for Standard tier Azure workspace. The UI lists `hive_metastore`. It would be better if these two were consistent.

4 REPLIES 4

Kaniz_Fatma
Community Manager
Community Manager

Hi @HakanNordgren, Letโ€™s address the inconsistency between spark_catalog and hive_metastore in Databricks.

Hereโ€™s what we know:

  1. spark_catalog:

    • This catalog is associated with Databricks and is used for managing metadata related to tables, views, and databases within Databricks.
    • It provides a unified view of data across various data sources and allows seamless querying using SQL.
    • The spark_catalog is an integral part of Databricksโ€™ functionality.
  2. hive_metastore:

    • The Hive Metastore is a component of Apache Hive, which is widely used for managing metadata in the Hadoop ecosystem.
    • It stores information about tables, partitions, and schemas.
    • Databricks leverages the Hive Metastore for compatibility and integration with existing Hive workloads.

Now, letโ€™s discuss consistency and recommendations:

  • Consistency:

    • Itโ€™s true that the UI and JDBC connection details should align to avoid confusion.
    • The UI displaying hive_metastore while JDBC lists spark_catalog can indeed be perplexing for users.

Thank you for flagging this.

@Kaniz_Fatma : Could I convince you to add the above as a bug to the backlog?

Also, @Kaniz_Fatma : Is it possible to have both a `hive_metastore` catalog and a `spark_catalog` catalog.

short answer is no

Databricks uses a single catalog -> Unity Catalog for managing tables created with Hive or Spark SQL, simplifying things. In spark sql you can do

spark-sql (default)> show catalogs;
spark_catalog

HTH

 

 

 

 

Mich Talebzadeh | Technologist | Data | Generative AI | Financial Fraud
London
United Kingdom

view my Linkedin profile



https://en.everybodywiki.com/Mich_Talebzadeh



Disclaimer: The information provided is correct to the best of my knowledge but of course cannot be guaranteed . It is essential to note that, as with any advice, quote "one test result is worth one-thousand expert opinions (Werner Von Braun)".

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you wonโ€™t want to miss the chance to attend and share knowledge.

If there isnโ€™t a group near you, start one and help create a community that brings people together.

Request a New Group