- 14365 Views
- 14 replies
- 4 kudos
In Azure Databricks Delta Lake not able to see unity catalog databases or tables in the drop down.
I have created an Azure data factory pipeline with a copy data function to copy data from adls path to a delta table .In the delta table drop downs i am able to see only the hive metastore database and tables only but the unity catalog tables are not...
- 14365 Views
- 14 replies
- 4 kudos
- 4 kudos
Hi all, Please try following this syntax in the database field <catalog_name>`.`<database_name> It is important that you use ` not ' This should force the connect to use a Unity Catalog Database.
- 4 kudos
- 5793 Views
- 5 replies
- 4 kudos
Resolved! Access table outside Databricks (hive metastore)
Hi guys, how are you ?How can I access tables outside the Databricks ? (hive metastore)I have a python script in my local structure but I need to access are tables stored in Databricks (hive metastore) , how ???? Have any idea ?Thank you guys
- 5793 Views
- 5 replies
- 4 kudos
- 4845 Views
- 6 replies
- 5 kudos
CREATE TABLE SHALLOW CLONE is not supported in Unity Catalog
Hi,I would like to clone the structure (schema only) of delta table without the content using SHALLOW CLONE command, as recommended in databricks documentation, but it doesn't work. I get an error message that this command is not supported in Unity C...
- 4845 Views
- 6 replies
- 5 kudos
- 5 kudos
You can shallow clone the table to an external location after sync the result of the shallow clone to unity catalog. Here is an example https://medium.com/@wahidatoui/sync-delta-tables-stored-in-dbfs-managed-or-external-to-unity-catalog-1fe29a3ae61b
- 5 kudos
- 1614 Views
- 2 replies
- 0 kudos
How to update change data UnityCatalog table after migrating hive metastore delta table
If I will migrate the Hive_metastore table into Unity Catalog table and subsequently hive metastore table is target of curation process. how I will migrate the delta record from Hive_metastore table to Unity Catalog table
- 1614 Views
- 2 replies
- 0 kudos
- 0 kudos
Hi @Vivek Singh Banafar​ Hope everything is going great.Just checking in. Does @Debayan Mukherjee​'s response answer your question? If yes, would you be happy to mark it as best so that other members can find the solution more quickly?We'd love to he...
- 0 kudos
- 9480 Views
- 12 replies
- 2 kudos
Create Schema with Unity enabled
i have unit setup, and am trying to create a delta schema and table, i am getting an error on the schema creation, i am able to list the files and folders in the ADLS Gen2 storage account, i am able to write a parquet file to ADLS, but i cannot crea...
- 9480 Views
- 12 replies
- 2 kudos
- 2 kudos
Hi @Shep Sheppard​ Hope all is well! Just wanted to check in if you were able to resolve your issue and would you be happy to share the solution or mark an answer as best? Else please let us know if you need more help. We'd love to hear from you.Than...
- 2 kudos
- 2527 Views
- 2 replies
- 2 kudos
Resolved! Sync delta tables across workspaces
What is the most cost-effective way to sync delta tables across workspaces? I'd like to sync a production table from the production workspace to a staging workspace for testing. It doesn't have to be in real-time, on a daily basis is acceptable. We a...
- 2527 Views
- 2 replies
- 2 kudos
- 2 kudos
Hi @Liz Zhang​ Hope everything is going great.Just wanted to check in if you were able to resolve your issue. If yes, would you be happy to mark an answer as best so that other members can find the solution more quickly? If not, please tell us so we ...
- 2 kudos
- 2016 Views
- 2 replies
- 0 kudos
Delta table error
Hi,we are creating an external table in the Azure databricks but facing this error:org.apache.spark.sql.AnalysisException: org.apache.hadoop.hive.ql.metadata.HiveException: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.meta...
- 2016 Views
- 2 replies
- 0 kudos
- 0 kudos
@Priyanka Biswas​ is there a place where we can have updates on metastore upgrades or control plane individual components. usually we can only see control plane region wide status
- 0 kudos
- 2262 Views
- 3 replies
- 20 kudos
Consume Delta table data using .net/python API's to present it on custom web application
How can we connect to unity catalog tables to create REST API's using Python/.NET?
- 2262 Views
- 3 replies
- 20 kudos
- 20 kudos
Did you mean create Delta Live Table Job using REST API ?
- 20 kudos
- 5623 Views
- 4 replies
- 6 kudos
Resolved! Cannot create a table having a column whose name contains commas in Hive metastore.
Hi Tried to create a delta table from spark data frame using below command:destination_path = "/dbfs/mnt/kidneycaredevstore/delta/df_corr_feats_spark_4"df_corr_feats_spark.write.format("delta").option("delta.columnMapping.mode", "name").option("path"...
- 5623 Views
- 4 replies
- 6 kudos
- 6 kudos
Hi @Shafiul Alam​ ,who gave those names to columns? you can rename you columns, replace spaces / special characters, for example:%python import re list_of_columns = df_corr_feats_spark.colums renamed_list_of_columns = [ re.sub(r'[^0-9a-zA-Z]+', "_", ...
- 6 kudos
-
Access
1 -
Access control
1 -
Access Mode
3 -
Account
4 -
Account Console
1 -
Account Level
2 -
Adf
1 -
ADLS
3 -
ADLS Gen2 Storage
1 -
ADLS Gen2 With ABFSS
1 -
AI
1 -
AmazonRDS
1 -
Amit
2 -
Analytics
1 -
API
4 -
AWS
8 -
Aws databricks
1 -
Aws glue
1 -
AWS Glue Catalog
1 -
Azure
21 -
Azure active directory
1 -
Azure Data Lake Storage
3 -
Azure databricks
16 -
Azure Databricks Delta Table
1 -
Azure DevOps
1 -
Azure Unity Catalog
1 -
Backup
1 -
Backup-restore
1 -
Bamboolib
1 -
Best practice
2 -
Best Practices
1 -
Best Way
1 -
Beta
1 -
Bug
1 -
Bug Report
1 -
Catalog
24 -
CatalogIntegration
1 -
CatalogShared Access
1 -
CatalogTable
1 -
CICD
1 -
Class Class
2 -
Cluster
3 -
Clusterlogs
2 -
Clusters
1 -
Code
2 -
Code Block
1 -
ColumnLevelSecurity
1 -
Columns
1 -
Company Email
1 -
ConfigurationInvalid Configuration Value
1 -
ContainerStorage
1 -
Create table
1 -
DAIS2023
1 -
Data
2 -
Data Engineering
2 -
Data Explorer
1 -
Data Governance
1 -
Data Ingestion
1 -
Data Science
2 -
Databrciks Runtime
1 -
databricks
4 -
Databricks Account
1 -
Databricks Audit Logs
1 -
Databricks Community
1 -
Databricks Data Engineer Associate
1 -
Databricks delta
1 -
Databricks Documentation
1 -
Databricks Feature Store
1 -
Databricks Lakehouse Platform
1 -
Databricks Migration Tool
1 -
Databricks notebook
3 -
Databricks Premium
2 -
Databricks Repos Api
1 -
Databricks Runtime
1 -
Databricks secrets
1 -
Databricks SQL
6 -
Databricks SQL Warehouse
1 -
Databricks Table Usage
3 -
Databricks Terraform
1 -
Databricks Unity Catalog
2 -
Databricks Workflows
1 -
Databricks workspace
4 -
Databricks-connect
2 -
DatabricksAWSAccount
1 -
DatabricksClusterLogs
1 -
Datagrip
1 -
DataMasking
1 -
DBFS
3 -
DBR
3 -
DBR10.4
2 -
Dbutils
1 -
Delete
1 -
Delt Lake
1 -
Delta
16 -
Delta Live
1 -
Delta Live Tables
3 -
Delta Schema
1 -
Delta Sharing
16 -
Delta table
9 -
Deploy Databricks
1 -
Dev
2 -
Different Account
1 -
DLT
7 -
DLT Pipelines
1 -
Documentation
2 -
Error
5 -
Error Message
3 -
Exception
1 -
Exists
1 -
External Hive
2 -
External Hive Metastore
2 -
External Metastore
1 -
External Metastores
1 -
External Table
2 -
Feature
1 -
Feature Store
1 -
File
2 -
File Trigger
1 -
Files
1 -
Function
1 -
GCP
2 -
GCP Databricks
2 -
Global init scripts
1 -
Governance
1 -
Grant
2 -
Groups
1 -
Hi
1 -
High Concurrency Cluster
1 -
Hive
7 -
Hive metastore
6 -
Hive Metastore Of Databricks
1 -
HiveCatalog
1 -
Iam
1 -
INFORMATION
1 -
Init Scripts
1 -
Instance Pool
1 -
Instance Profile
1 -
Integrations
1 -
INVALID STATE
1 -
IP Access List
1 -
JOBS
1 -
JSON
2 -
Limit
1 -
Lineage
1 -
List
1 -
Location
2 -
Locations
1 -
Machine Learning
1 -
ManagedIdentities
1 -
Metastore
15 -
Method
1 -
Method Public
1 -
Microsoft
1 -
Migration
1 -
ML
1 -
Mounts
1 -
Multi
1 -
Multiple Times
1 -
Multiple users
1 -
Network Security
1 -
New Features
2 -
New Workspace
1 -
Notebook
1 -
Notebook Display Widgets
1 -
Organize
1 -
Pandas
1 -
Party Libraries
1 -
Path
1 -
Permission
1 -
Permissions
3 -
Possible
1 -
PPC
1 -
Premium
1 -
Proven Practice
1 -
Public
1 -
Py4j.security.Py4JSecurityException
2 -
Pyspark
3 -
Python
5 -
Python Code
1 -
Python notebook
1 -
Python programming
1 -
Python Proxy
1 -
Python script
1 -
Python2
1 -
Query
2 -
Query History
1 -
Question
1 -
R
2 -
Rdd
1 -
Rest
1 -
Rest API
5 -
REST Endpoint
1 -
Row Level Security
1 -
S3 Params
1 -
S3 permission
1 -
Scala
1 -
Scala spark
1 -
Schema
1 -
Schema Browser
1 -
SCIM
1 -
Search
1 -
Secrets
2 -
Security
4 -
Security Controls
1 -
SEO
1 -
SEO Agency Sydney
1 -
Separate Workspace
1 -
Sequence Init Scripts
1 -
Series En LÃnea
1 -
Series Online
1 -
Service
1 -
Service endpoint
1 -
Service principal
1 -
SET MASK
1 -
SET ROW FILTER
1 -
Setup databricks
1 -
Shallow Clone
1 -
Shared Mode
4 -
SLDC
1 -
SLDC Environemtn
1 -
Snowflake
1 -
Snowflake Federation
1 -
Spark
5 -
Spark config
1 -
Spark Daatricks
1 -
Spark Path Format
1 -
Spark streaming
1 -
Spark view
1 -
SparkFiles
1 -
Sparklyr
1 -
SQL
10 -
SQL Editor
2 -
SQL View
1 -
Sql Warehouse
1 -
Status Code
1 -
Storage
1 -
Storage account
1 -
Stored Procedures
1 -
Strange Behavior
1 -
Streaming
1 -
Summit22
1 -
SVC
1 -
SVC Principal
1 -
Sync Users
1 -
Table
5 -
Table access control
2 -
Table ACLs
1 -
Table Integration
1 -
TABLE SHALLOW CLONE
1 -
Tableau Desktop
1 -
Tables
4 -
TBL
1 -
Terraform
5 -
Terraform Provider
1 -
Trying
1 -
TV En Streaming
1 -
UC
10 -
UC - Service Principal
1 -
UC NOT ENABLED
1 -
Udf
3 -
Unclear Error Message
1 -
Unclear Validation Error Message
1 -
Understanding Delta Lake
1 -
UNDROP
1 -
Unify Catalog
1 -
Unity
3 -
Unity Catalog
185 -
Unity Catalog +
1 -
Unity Catalog API
1 -
Unity Catalog Data Lineage
1 -
Unity Catalog Error
1 -
Unity Catalog Metastore
1 -
Unity Catalog Objects
1 -
Unity Catalog Request
1 -
Unity Catalog Support
1 -
Unity Catalog Sync
1 -
Unity Catalog Table
1 -
Unity Catalogue
1 -
Unity Setup
1 -
UnityCalatog
1 -
UnityCatalog
1 -
UnityCatalog Tables
1 -
Unstructured
1 -
Usa
1 -
User
2 -
User Access Tokens
1 -
Ver Series
1 -
Ver Series En Español
1 -
Version
1 -
Views
1 -
Vpc
1 -
Wildrose
1 -
Wildrose Flixlatino
1 -
Workers
1 -
Workflows
2 -
Workspace
4 -
Write
1
- « Previous
- Next »