cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

MBV3
by Contributor
  • 15611 Views
  • 5 replies
  • 7 kudos

Resolved! External table from parquet partition

Hi,I have data in parquet format in GCS buckets partitioned by name eg. gs://mybucket/name=ABCD/I am trying to create a table in Databaricks as followsDROP TABLE IF EXISTS name_test; CREATE TABLE name_testUSING parquetLOCATION "gs://mybucket/name=*/...

  • 15611 Views
  • 5 replies
  • 7 kudos
Latest Reply
Pat
Esteemed Contributor
  • 7 kudos

Hi @M Baig​ ,the error doesn't tell me much, but you could try:CREATE TABLE name_test USING parquet PARTITIONED BY ( name STRING) LOCATION "gs://mybucket/";

  • 7 kudos
4 More Replies
ac0
by Contributor
  • 2414 Views
  • 0 replies
  • 0 kudos

Get size of metastore specifically

Currently my Databricks Metastore is in the the same location as the data for my production catalog. We are moving the data to a separate storage account. In advance of this, I'm curious if there is a way to determine the size of the metastore itself...

  • 2414 Views
  • 0 replies
  • 0 kudos
DylanS
by New Contributor II
  • 5923 Views
  • 7 replies
  • 6 kudos

FileNotFoundError: [Errno 2] No such file or directory: 'pylsp'

We are intermittently experiencing the below issue when running mundane code in our databricks notebook environment using 13.3 LTS runtime, with a compute pool with r6id.large on-demand instances, using local storage.We first noticed this late last w...

DylanS_0-1707756410914.png
  • 5923 Views
  • 7 replies
  • 6 kudos
Latest Reply
engixcmt
New Contributor II
  • 6 kudos

Hello @Navya_R ,We are facing a similar issue when using 14.3LTS with DCSFor us, certain Global Inits are not getting applied. Is there a patch we can use for 14.3 LTS as well?

  • 6 kudos
6 More Replies
SandeepG
by New Contributor
  • 3746 Views
  • 1 replies
  • 0 kudos

not able to create temporary tables in unity catalog

We are using a unity catalog environment and when trying to create a temporary table the statement errored out. 

SandeepG_0-1710776912989.png
  • 3746 Views
  • 1 replies
  • 0 kudos
Latest Reply
Sampath_Kumar
New Contributor II
  • 0 kudos

Hi @SandeepG Could you please share the need of a temporary table?Here are the possible waysTables:External Tables: Tables can be created on top of files which are externally located.Managed Tables: The usual tables which will be stored and managed a...

  • 0 kudos
Gilg
by Contributor II
  • 2546 Views
  • 1 replies
  • 0 kudos

Autoloader - File Notification mode

Hi All,I have set up a DLT pipeline that is using Autoloader in a file notification mode.Everything runs smoothly for the first time. However, it seems like the next micro-batch did not trigger as I can see some events coming in the queue.But if I lo...

Gilg_0-1710827649089.png Gilg_1-1710827662118.png
  • 2546 Views
  • 1 replies
  • 0 kudos
cpd
by New Contributor II
  • 4080 Views
  • 1 replies
  • 0 kudos

Ingesting geospatial data into a table

I'm just getting started with Databricks and wondering if it is possible to ingest a GeoJSON or GeoParquet file into a new table without writing code? My goal here is to load vector data into a table and perform H3 polyfill operations on all the vect...

  • 4080 Views
  • 1 replies
  • 0 kudos
Latest Reply
cpd
New Contributor II
  • 0 kudos

Thank you @Retired_mod - much appreciated!

  • 0 kudos
ashish577
by New Contributor III
  • 2519 Views
  • 0 replies
  • 0 kudos

How do we pass parameters which have a "," with bundle run ?

So I have a query "select col1, col2 from table" that I need to pass as a parameter to a databricks job that I am triggering through the bundle run command. Issue is, when I pass this via --params="query=select col1, col2 from table" it splits it bas...

  • 2519 Views
  • 0 replies
  • 0 kudos
Brichj
by New Contributor II
  • 2712 Views
  • 2 replies
  • 0 kudos

%run ../Includes/Classroom-Setup-02.1

I ran the code in the cell as it was given in the presentation. But it failed. Can someone please help?The presentation is the second lesson in the second model of Data Engineering Associate exam prep.

  • 2712 Views
  • 2 replies
  • 0 kudos
Latest Reply
Brichj
New Contributor II
  • 0 kudos

Thanks Ajay-Pandey!This is error that I keep getting when I run the following: %run ./Includes/Classroom-Setup-02.3LI have run dbutils.library.restartPython(), but it did not help.Note: you may need to restart the kernel using dbutils.library.restart...

  • 0 kudos
1 More Replies
Brad
by Contributor II
  • 2963 Views
  • 3 replies
  • 0 kudos

Inconsistent behavior when displaying chart in notebook

Hi, I'm trying to create some 3D charts. With the same code and same cluster, sometimes it can show, sometimes it cannot. Previously it cannot display, but last week I opened a notebook with failed run and found the result can be shown by itself (as ...

  • 2963 Views
  • 3 replies
  • 0 kudos
Latest Reply
Brad
Contributor II
  • 0 kudos

Also, with same code, same browser, different workspaces, one works, other one not. In the notebook with "script error", if I "Export cell" and get its iframe html and use displayHTML to display it, it works, so this means the JS and HTML inside is o...

  • 0 kudos
2 More Replies
VGS777
by New Contributor III
  • 1207 Views
  • 0 replies
  • 0 kudos

Regarding Cloning dbc file from git

Hey FolksI have dbc file in a git repo and i cloned in the databricks when tried to open the .dbc file it is saying ```Failed to load file. The file encoding is not supported```can anyone please advice me on this #help #beginner

  • 1207 Views
  • 0 replies
  • 0 kudos
LoiNguyen
by New Contributor II
  • 19633 Views
  • 5 replies
  • 2 kudos

The authentication type 10 is not supported

I use below code to connect to postgresql. df = spark.read \ .jdbc("jdbc:postgresql://hostname:5432/dbname", "schema.table", properties={"user": "user", "password": "password"})\ .load() df.printSchema() However, I got the ...

  • 19633 Views
  • 5 replies
  • 2 kudos
Latest Reply
simboss
New Contributor II
  • 2 kudos

But how are we going to do this for those who use Windows?

  • 2 kudos
4 More Replies
PassionateDBD
by New Contributor II
  • 1521 Views
  • 1 replies
  • 1 kudos

Is it possible to create/update non dlt table in init phase of dlt task?

We have a dlt task that is written in python. Is it possible to create or update a delta table programatically from inside a dlt task? The delta table would not be managed from inside the dlt task because we never want to fully refresh that table. Th...

  • 1521 Views
  • 1 replies
  • 1 kudos
Latest Reply
PassionateDBD
New Contributor II
  • 1 kudos

Thanks for you reply @Retired_mod ! I'm aware of the possibility to create or not create a table based on some parameter.What I'm trying to figure out is basically how to achieve following:-DLT pipeline starts and logs some information to a delta tab...

  • 1 kudos

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels