cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

SamAdams
by New Contributor III
  • 35 Views
  • 2 replies
  • 1 kudos

Migrating source directory in an existing DLT Pipeline with Autoloader

I have a DLT pipeline that reads data in S3 into an append-only bronze layer using Autoloader. The data sink needs to be changed to a new s3 bucket in a new account, and data in the existing s3 bucket migrated to the new one.Will Autoloader still be ...

  • 35 Views
  • 2 replies
  • 1 kudos
Latest Reply
Brahmareddy
Honored Contributor II
  • 1 kudos

Hi SamAdams,How are you doing today?, As per my understanding, You're on the right track here! When you change the S3 path for Autoloader, even if the files are exactly the same and just copied from the old bucket, Autoloader will treat them as new f...

  • 1 kudos
1 More Replies
Mulder81
by New Contributor II
  • 91 Views
  • 3 replies
  • 2 kudos

PDF Generation via databricks Job

WE have a databricks Job that will aggregate some data and create some data tables. This needs to be exported out in a PDF format.I have seen a few python libraries that can generate PDF, but was wondering if the PDF can be generated and dropped in a...

  • 91 Views
  • 3 replies
  • 2 kudos
Latest Reply
Mulder81
New Contributor II
  • 2 kudos

Are there any specific ways to generate the PDF file from a dataframe? and libraries that work

  • 2 kudos
2 More Replies
PraveenReddy21
by New Contributor III
  • 803 Views
  • 3 replies
  • 0 kudos

how to create catalog

Hi ,I am trying to create  catalog  and database  its not  allowing  databricks  , please  suggest  .Here my code .base_dir = "/mnt/files"spark.sql(f"CREATE CATALOG IF NOT EXISTS dev")spark.sql(f"CREATE DATABASE IF NOT EXISTS dev.demo_db") first i ne...

  • 803 Views
  • 3 replies
  • 0 kudos
Latest Reply
JairoCollante
  • 0 kudos

I got a similar error trying create a catalog with "databricks.sdk" library I resolved it add the parameter "storage_root": w.catalogs.create(name=c.name, storage_root='s3://databricks-workspace-bucket/unity-catalog/426335709') In my case all catalog...

  • 0 kudos
2 More Replies
TomHauf
by Visitor
  • 38 Views
  • 0 replies
  • 0 kudos

Sending my weather data to a clients cloud storage

Hi, One of our clients is asking to switch from our API feed to have weather data delivered automatically to their Cloud Storage.  What steps do I need to take from my end?  Do I need to join Databricks to do so? Thanks. Tom

  • 38 Views
  • 0 replies
  • 0 kudos
MadhuB
by Contributor III
  • 75 Views
  • 1 replies
  • 0 kudos

Installing Maven (3rd party) libraries on Job Cluster

I'm trying to install Maven Libraries on the job cluster (non interactive cluster) as part of databricks workflow. I've added the context in the cluster configuration as part of deployment which I cant find the same in the post deployment configurati...

MadhuB_0-1742919949369.png
  • 75 Views
  • 1 replies
  • 0 kudos
Latest Reply
MadhuB
Contributor III
  • 0 kudos

I found the workaround. Below are the steps:1. Add the required library to the Allowed list at the workspace level (require workspace/metastore admin access); you might need coordinates groupdd:artifactId:version2. At the task level, include under De...

  • 0 kudos
usabuysmm5
by Visitor
  • 48 Views
  • 0 replies
  • 0 kudos

How do I verify my PayPal account?

Get Reliable PayPal Account Assistance Are you looking for a seamless way to manage international transactions? Our team provides secure solutions to help you verify your PayPal account and ensure hassle-free payment acceptance across the globe.

  • 48 Views
  • 0 replies
  • 0 kudos
Pu_123
by Visitor
  • 91 Views
  • 1 replies
  • 0 kudos

Cluster configuration

Please help me configure/choose the cluster configuration. I need to process and merge 6 million records into Azure SQL DB. At the end of the week, 9 billion records need to be processed and merged into Azure SQL DB, and a few transformations need to...

  • 91 Views
  • 1 replies
  • 0 kudos
Latest Reply
Shua42
Databricks Employee
  • 0 kudos

It will depend on the transformations and how you're loading them. Assuming it's mostly in spark, I recommend starting small using a job compute cluster with autoscaling enabled for cost efficiency. For daily loads (6 million records), a driver and 2...

  • 0 kudos
walgt
by Visitor
  • 79 Views
  • 1 replies
  • 0 kudos

Permission Issue in Delta Lake Course

Hi everyone,I'm new to Databricks and working on the "Data Ingestion with Delta Lake" course. I encountered a permission error with the following query:Can anyone help with this?Thanks! 

walgt_0-1742915294633.png
  • 79 Views
  • 1 replies
  • 0 kudos
Latest Reply
Advika
Databricks Employee
  • 0 kudos

Hello @walgt! Apologies for the inconvenience. This was a known issue, but it has now been fixed! You should now be able to run your query without any problems. Thanks for your patience!

  • 0 kudos
subhas
by New Contributor
  • 53 Views
  • 0 replies
  • 0 kudos

Auto Loader bringing NULL Records

Hi        I am using auto loader to fetch some records stored in two files. Please see below my code. It fetches records from two files correctly and then it starts fetching NULL records. I attach option("cleanSource",    ) to  readStream. But it is ...

  • 53 Views
  • 0 replies
  • 0 kudos
chexa_Wee
by Visitor
  • 211 Views
  • 5 replies
  • 0 kudos

How to manage two separate projects ?

Hi all, I am managing one project in Databricks, with one more coming soon. Can anyone guide me on how to use Unity Catalog or any other method for this?"

  • 211 Views
  • 5 replies
  • 0 kudos
Latest Reply
mnorland
New Contributor III
  • 0 kudos

There are a wide variety of needs that need to be considered such as governance, compute and storage.  This depends on the size of your projects.

  • 0 kudos
4 More Replies
ayushmangal72
by Visitor
  • 113 Views
  • 1 replies
  • 0 kudos

Revert cluster DBR version to last DBR

Hi Team,We have updated our clusters DBR version, later we got to know that some of our jobs started failing, now we wanted to revert to DBR version to the previos one only but we forgot the DBR version on which job was running fine.Is there any way ...

  • 113 Views
  • 1 replies
  • 0 kudos
Latest Reply
adhi_databricks
New Contributor III
  • 0 kudos

 Hey @ayushmangal72 , try using the Databricks Job Run API (/api/2.2/jobs/runs/list) to fetch older run IDs for the job.Once you have the run_id, make a request to the API at /api/2.2/jobs/runs/get. You'll be able to find the DBR version in the API r...

  • 0 kudos
cmathieu
by New Contributor II
  • 71 Views
  • 0 replies
  • 0 kudos

DAB - All projects files deployed

I have an issue with DAB where all the project files, starting from root ., get deployed to the /files folder in the bundle. I would prefer being able to deploy certain util notebooks, but not all the files of the project. I'm able to not deploy any ...

  • 71 Views
  • 0 replies
  • 0 kudos

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels