- 11660 Views
- 7 replies
- 4 kudos
no information on character_maximum_length and is_identity in unity catalog table system.information_schema.columns
no information on character_maximum_length and is_identity in unity catalog table system.information_schema.columnstable definition exampleCREATE TABLE tyable_name ( UniqueID bigint not null GENERATED ALWAYS AS IDENTITY , ID varchar(256))looks like ...
- 11660 Views
- 7 replies
- 4 kudos
- 4 kudos
Hello Team,Still i am facing same issue of value zero in character_maximum_length. Is there any other way we can get character_maximum_length of columns?When we can expect to resolve this bug?
- 4 kudos
- 4898 Views
- 3 replies
- 1 kudos
- 4898 Views
- 3 replies
- 1 kudos
- 1 kudos
Starting DBR 12.1 you can run the UNDROP command to undrop a managed or external table. The table must be in Unity Catalog for this feature to work. See https://docs.databricks.com/sql/language-manual/sql-ref-syntax-ddl-undrop-table.html for more det...
- 1 kudos
- 2109 Views
- 2 replies
- 0 kudos
How to update change data UnityCatalog table after migrating hive metastore delta table
If I will migrate the Hive_metastore table into Unity Catalog table and subsequently hive metastore table is target of curation process. how I will migrate the delta record from Hive_metastore table to Unity Catalog table
- 2109 Views
- 2 replies
- 0 kudos
- 0 kudos
Hi @Vivek Singh Banafar​ Hope everything is going great.Just checking in. Does @Debayan Mukherjee​'s response answer your question? If yes, would you be happy to mark it as best so that other members can find the solution more quickly?We'd love to he...
- 0 kudos
- 7149 Views
- 4 replies
- 6 kudos
Resolved! Cannot create a table having a column whose name contains commas in Hive metastore.
Hi Tried to create a delta table from spark data frame using below command:destination_path = "/dbfs/mnt/kidneycaredevstore/delta/df_corr_feats_spark_4"df_corr_feats_spark.write.format("delta").option("delta.columnMapping.mode", "name").option("path"...
- 7149 Views
- 4 replies
- 6 kudos
- 6 kudos
Hi @Shafiul Alam​ ,who gave those names to columns? you can rename you columns, replace spaces / special characters, for example:%python import re list_of_columns = df_corr_feats_spark.colums renamed_list_of_columns = [ re.sub(r'[^0-9a-zA-Z]+', "_", ...
- 6 kudos
- 6049 Views
- 3 replies
- 1 kudos
Resolved! Table ACLs, secrets, and compute clusters
I am adding a credential to access MongoDB through the Python MongoDB Spark connector. It is stored as a secret and the `users` group has `MANAGE` ACL privileges. I turned on table ACLs and created a `High Concurrency` compute cluster. I then run a s...
- 6049 Views
- 3 replies
- 1 kudos
- 1 kudos
@Erik Louie​ looks the issue is resolved. Please let me know if you need further help on this. Thanks!
- 1 kudos
-
Access control
1 -
Access Mode
3 -
Account Console
1 -
Account Level
2 -
ADLS Gen2 Storage
1 -
ADLS Gen2 With ABFSS
1 -
AmazonRDS
1 -
Amit
2 -
Analytics
1 -
AWS
8 -
Aws databricks
1 -
AWS Glue Catalog
1 -
Azure
21 -
Azure active directory
1 -
Azure Data Lake Storage
3 -
Azure databricks
16 -
Azure Databricks Delta Table
1 -
Azure Unity Catalog
1 -
Backup
1 -
Backup-restore
1 -
Bamboolib
1 -
Beta
1 -
CatalogShared Access
1 -
Class Class
2 -
Cluster
3 -
Clusterlogs
2 -
Code Block
1 -
ColumnLevelSecurity
1 -
Company Email
1 -
ContainerStorage
1 -
Create table
1 -
Data
2 -
Data Engineering
2 -
Data Explorer
1 -
Data Governance
1 -
Data Ingestion
1 -
Databrciks Runtime
1 -
databricks
4 -
Databricks Account
1 -
Databricks Community
1 -
Databricks Data Engineer Associate
1 -
Databricks delta
1 -
Databricks Documentation
1 -
Databricks Feature Store
1 -
Databricks Lakehouse Platform
1 -
Databricks notebook
3 -
Databricks Premium
2 -
Databricks Runtime
1 -
Databricks secrets
1 -
Databricks SQL
6 -
Databricks SQL Warehouse
1 -
Databricks Table Usage
3 -
Databricks Terraform
1 -
Databricks Unity Catalog
2 -
Databricks Workflows
1 -
Databricks workspace
4 -
Databricks-connect
2 -
DatabricksAWSAccount
1 -
Datagrip
1 -
DataMasking
1 -
DBFS
3 -
Delete
1 -
Delt Lake
1 -
Delta
16 -
Delta Live
1 -
Delta Live Tables
3 -
Delta Sharing
17 -
Deploy Databricks
1 -
Dev
2 -
Different Account
1 -
DLT
7 -
DLT Pipelines
1 -
Exists
1 -
External Hive
2 -
External Hive Metastore
2 -
External Metastore
1 -
External Metastores
1 -
External Table
2 -
Feature Store
1 -
File Trigger
1 -
GCP Databricks
2 -
Global init scripts
1 -
Governance
1 -
Grant
2 -
Groups
1 -
Hi
1 -
High Concurrency Cluster
1 -
Hive metastore
6 -
Hive Metastore Of Databricks
1 -
HiveCatalog
1 -
Iam
1 -
INFORMATION
1 -
Instance Pool
1 -
Instance Profile
1 -
Integrations
1 -
IP Access List
1 -
Limit
1 -
Lineage
1 -
Location
2 -
Machine Learning
1 -
Method
1 -
Method Public
1 -
Migration
1 -
Mounts
1 -
Multi
1 -
Multiple users
1 -
Network Security
1 -
New Features
2 -
New Workspace
1 -
Notebook
1 -
Notebook Display Widgets
1 -
Organize
1 -
Permissions
3 -
Possible
1 -
Premium
1 -
Public
1 -
Py4j.security.Py4JSecurityException
2 -
Pyspark
3 -
Python
5 -
Python Code
1 -
Python programming
1 -
Python Proxy
1 -
Query History
1 -
Row Level Security
1 -
Scala
1 -
Scala spark
1 -
Search
1 -
Secrets
2 -
Security
4 -
Security Controls
1 -
Separate Workspace
1 -
Service
1 -
Shallow Clone
1 -
Shared Mode
4 -
Spark
5 -
Spark config
1 -
Spark view
1 -
SparkFiles
1 -
Sparklyr
1 -
SQL
10 -
SQL Editor
2 -
Sql Warehouse
1 -
Storage
1 -
Strange Behavior
1 -
Table
5 -
Table access control
2 -
TBL
1 -
Trying
1 -
Understanding Delta Lake
1 -
UNDROP
1 -
Unity
3 -
Unity Catalog
198 -
Unity Catalog Metastore
1 -
Unity Catalogue
1 -
UnityCalatog
1 -
Views
1 -
Workers
1 -
Workflows
2 -
Workspace
4 -
Write
1
- « Previous
- Next »