cancel
Showing results for 
Search instead for 
Did you mean: 
Data Governance
Join discussions on data governance practices, compliance, and security within the Databricks Community. Exchange strategies and insights to ensure data integrity and regulatory compliance.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

MadelynM
by Databricks Employee
  • 2089 Views
  • 0 replies
  • 0 kudos

[Recap] Data + AI Summit 2024 - Data Governance | Navigate the explosion of AI, data and tools

Here's your Data + AI Summit 2024 - Data Governance recap as you navigate the explosion of AI, data and tools in efforts to build a flexible and scalable governance framework that spans your entire data and AI estate. Keynote: Evolving Data Governan...

Screenshot 2024-07-03 at 9.27.29 AM.png
  • 2089 Views
  • 0 replies
  • 0 kudos
TSK
by New Contributor
  • 702 Views
  • 1 replies
  • 0 kudos

Set up GitLab server on Databricks Container Services (DCS)

I would like to set up GitLab and Grafana servers using Databricks Container Services (DCS). The reason is that our development team is small, and the management costs of using EKS are not justifiable. We want to make GitLab and Grafana accessible in...

  • 702 Views
  • 1 replies
  • 0 kudos
Latest Reply
ameame
New Contributor II
  • 0 kudos

 Yes, it’s possible to set up GitLab and Grafana servers using Databricks Container Services (DCS). Here’s a high-level overview of how you can achieve this:Prepare Your Docker Images:Create Docker images for GitLab and Grafana. You can use official ...

  • 0 kudos
jerickson
by New Contributor II
  • 1801 Views
  • 3 replies
  • 0 kudos

CREATE EXTERNAL LOCATION on a publicly available S3 bucket

I would like to create an external location on a publicly available S3 bucket, for which I don't have credentials. I get a syntax error unless I include credentials. Is there a way to do this? CREATE EXTERNAL LOCATION public_bucket URL 's3://public_b...

  • 1801 Views
  • 3 replies
  • 0 kudos
Latest Reply
filipniziol
Esteemed Contributor
  • 0 kudos

Hi @jerickson ,I have tested this on Databricks Runtime 14.3 LTS:Install on the cluster the below Maven packages:com.amazonaws:aws-java-sdk-bundle:1.12.262org.apache.hadoop:hadoop-aws:3.3.4Run the below code to read your csv file into dataframe: spar...

  • 0 kudos
2 More Replies
DG1
by New Contributor II
  • 1914 Views
  • 0 replies
  • 1 kudos

Data dictionary and glossar capabilities

Hi everyone!I would love to know whether Databricks has the following capabilities and what are the core functionalities of the tool:Data glossary: provides clear definitions of business terms, ensuring that everyone in the organization uses the same...

  • 1914 Views
  • 0 replies
  • 1 kudos
jv_v
by Contributor
  • 2296 Views
  • 5 replies
  • 6 kudos

How to Transfer Hive Metastore from One Workspace to Another

Hi everyone,I need to transfer tables present in the Hive metastore from one workspace to another. My ultimate goal is to migrate these tables to Unity Catalog. I already have a code that will handle the migration to Unity Catalog, but I need help wi...

  • 2296 Views
  • 5 replies
  • 6 kudos
Latest Reply
jose_gonzalez
Databricks Employee
  • 6 kudos

Are these tables managed tables or external tables? This is very important step for your migration. 

  • 6 kudos
4 More Replies
Gilg
by Contributor II
  • 2354 Views
  • 3 replies
  • 0 kudos

Delta Sharing within the same region

Hi Team,I have setup UC and have a Metastore for AUE. Assigned 2 workspaces to the AUE Metastore and both with different set of users and group . 1 workspace is the data provider and the other is a data consumerNow since 2 workspaces are in one Meta...

  • 2354 Views
  • 3 replies
  • 0 kudos
Latest Reply
NikunjKakadiya
New Contributor II
  • 0 kudos

Do we have any updates on this in terms of how we can share the data between different workspaces part of the same metastore?

  • 0 kudos
2 More Replies
Praveen
by New Contributor II
  • 1381 Views
  • 1 replies
  • 0 kudos

Enable external data access on the metastore

Hello Everyone,I have an AWS Databricks premium trial account and I’m trying to enable the configuration "Enable external data access on the metastore" for a metastore. I followed the steps in the official documentation (https://docs.databricks.com/e...

Praveen_0-1727338697909.png
  • 1381 Views
  • 1 replies
  • 0 kudos
Latest Reply
saikumar246
Databricks Employee
  • 0 kudos

Hi @Praveen Thank you for sharing your concern here. You are looking into the Catalog on the account console, but according to the document, you have to do this in the Workspace Catalog UI. Please follow the below steps, Log into any of the workspace...

  • 0 kudos
robbinbajaj
by New Contributor
  • 675 Views
  • 1 replies
  • 0 kudos

Restrict User to see schema using "all purpose compute"

Scenario Under Hive Metastore four diff schema exist S1,S2,S3 and S4 User1 want access to schema S1 but not allowed to see other 3 Schema which can be done using grant permission simple and easy case Same User1 has access to workspace. Now same user ...

  • 675 Views
  • 1 replies
  • 0 kudos
Latest Reply
gchandra
Databricks Employee
  • 0 kudos

The best option for Governance is to upgrade to Unity Catalog.  This is alternate, not the best. The cluster must be enabled for table access control. https://docs.databricks.com/en/data-governance/table-acls/object-privileges.html#requirements

  • 0 kudos
jv_v
by Contributor
  • 3033 Views
  • 1 replies
  • 0 kudos

Issues Creating External Tables in Hive Metastore vs. Unity Catalog

Hi Databricks Community,I’m encountering some challenges when working with external tables in Databricks. I’m using both Unity Catalog and Hive Meta store and have noticed differences in how external tables are handled.Unity Catalog: I can successful...

  • 3033 Views
  • 1 replies
  • 0 kudos
Latest Reply
Imalkov
New Contributor III
  • 0 kudos

hi @jv_v , that's possible to put a path to external storage i.e.:documentationCREATE OR REPLACE TABLE hive_metastore.schema.table LOCATION 's3://depts/finance/sec_filings'Can you share which issue do you have cause it's not clear from your questio...

  • 0 kudos
Karol1977
by New Contributor
  • 927 Views
  • 1 replies
  • 0 kudos

example how to search by both tag key and tag value

Hi,Could you please give me example how to search by both tag key and tag value?This article doesn't show such example: https://docs.databricks.com/en/search/index.html#tagsThanks,Karol

  • 927 Views
  • 1 replies
  • 0 kudos
Latest Reply
gchandra
Databricks Employee
  • 0 kudos

If you want to search by tag key alone, use the syntax: tag:<tag_key>. To search by both tag key and tag value, omit tag:. https://docs.databricks.com/en/search/index.html#use-tags-to-search-for-tables  

  • 0 kudos
MarcoRezende
by New Contributor II
  • 7292 Views
  • 1 replies
  • 0 kudos

Billing per user

I need to know how much money each user spend in my workspace, i already have the system billing table, but there is not much information in it. How can i achive that?

  • 7292 Views
  • 1 replies
  • 0 kudos
Latest Reply
gchandra
Databricks Employee
  • 0 kudos

You can tag the cluster and use this dashboard  https://learn.microsoft.com/en-us/azure/databricks/admin/account-settings/usage  

  • 0 kudos
dataguy555
by New Contributor
  • 1005 Views
  • 2 replies
  • 0 kudos
  • 1005 Views
  • 2 replies
  • 0 kudos
Latest Reply
filipniziol
Esteemed Contributor
  • 0 kudos

Hi @dataguy555 ,You will still need to have your external table in delta format. JSON or parquet are not supported:What is Delta Sharing? - Azure Databricks | Microsoft Learn 

  • 0 kudos
1 More Replies
pargit
by New Contributor
  • 1251 Views
  • 0 replies
  • 0 kudos

Tag dynamic allpurpose

hi..I want to use  1 all purpose cluster   and use dynamic taging in each notebook.for example tag  project  and department.I want to be able to change the tag from the notebook  so I will be able to understand the costs for each project and departme...

  • 1251 Views
  • 0 replies
  • 0 kudos
MarthinusB
by New Contributor
  • 1387 Views
  • 1 replies
  • 0 kudos

Pyspark SQL Write Error: SparkConnectGrpcException (com.microsoft.sqlserver.jdbc.SQLServerException)

When attempting to write a dataframe to SQL Server using the provided "sqlserver" connector or the old "jdbc" one, it seems to be breaking due to old values used to build the dataframe. The code I'm using to write is:```python(    df.write.format("sq...

  • 1387 Views
  • 1 replies
  • 0 kudos
Latest Reply
Brahmareddy
Honored Contributor III
  • 0 kudos

Hi @MarthinusB, How are you doing today?To overcome this, Give a try explicitly defining the schema for your DataFrame to ensure no unwanted metadata is retained. You can also recreate the DataFrame using rdd.toDF() or checkpoint it to drop unnecessa...

  • 0 kudos
prasadvaze
by Valued Contributor II
  • 1246 Views
  • 1 replies
  • 0 kudos

Delta sharing not working on tables with row filter or column mask

I have a delta table  with a column mask (azure databricks in eastus2 metastore) and another with no mask. I shared both tables with  westeurope unity catalog metastore  azure region via delta sharing (created share object, added the table as asset, ...

  • 1246 Views
  • 1 replies
  • 0 kudos
Latest Reply
daniel_sahal
Esteemed Contributor
  • 0 kudos

@prasadvaze That's correct, Delta Sharing does not work with RLS or column mask. It's well described in the documentation: https://docs.databricks.com/en/tables/row-and-column-filters.html#limitations

  • 0 kudos

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels