cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

nguyenthuymo
by New Contributor III
  • 2945 Views
  • 3 replies
  • 1 kudos

Resolved! Unable to import a library before restarting the cluster

Hi everyone,Anyone has problem when import a library in Databricks notebook? I found it failed to import. Then I restarted the cluster, run it again and it successfully imported the library.My concern here is that: I scheduled to run the notebook at ...

  • 2945 Views
  • 3 replies
  • 1 kudos
Latest Reply
nguyenthuymo
New Contributor III
  • 1 kudos

Thanks LR. That looks like a great response!

  • 1 kudos
2 More Replies
skosaraju
by New Contributor III
  • 2626 Views
  • 5 replies
  • 1 kudos

Resolved! Getting an error "You cannot use dbutils within a spark job"

I am using mlflow to register my custom model using a simple code below. The DatabricksParams extracts all the params from dbutils and sets the params dictionary and dbutils is not used anywhere else within the rest of my code base. The code fails wh...

  • 2626 Views
  • 5 replies
  • 1 kudos
Latest Reply
lingareddy_Alva
Esteemed Contributor
  • 1 kudos

Thanks for the update @skosaraju .

  • 1 kudos
4 More Replies
daan_dw
by New Contributor III
  • 2322 Views
  • 3 replies
  • 0 kudos

Resolved! Auto Loader for copying files on s3

Hey community,I have a folder on s3 with around 5 million small files. On a daily basis new files are added. I would like to simply copy those new files to another folder on s3. My approach is to use an Auto Loader of which I attached the code below....

  • 2322 Views
  • 3 replies
  • 0 kudos
Latest Reply
daan_dw
New Contributor III
  • 0 kudos

Hey LRALVAThe first time running your code I got the error:  PicklingError: Could not serialize object: Exception: You cannot use dbutils within a spark job You cannot use dbutils within a spark job or otherwise pickle it.So I changed the copy_single...

  • 0 kudos
2 More Replies
asrivas
by New Contributor II
  • 1702 Views
  • 1 replies
  • 0 kudos

Resolved! Is SSL cert needed for Azure Databricks API calls?

Hi, I'm using Azure Databricks and making API calls to the endpoints with verify=False Eg.response = requests.get('https://%s/api/2.0/clusters/list' % (databricks_domain),headers=request_headers,verify=False)Security scanners are flagging the use of ...

  • 1702 Views
  • 1 replies
  • 0 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 0 kudos

In API communication with Azure Databricks, it is not necessary to manually install or configure an SSL certificate on the client side because Azure Databricks endpoints inherently use TLS (Transport Layer Security) for secure communication. To addre...

  • 0 kudos
azimbhuiyan
by New Contributor II
  • 5024 Views
  • 2 replies
  • 1 kudos

Docker image fails to initalize - Databricks Container Service

I am trying to use a public docker image hosted in Dockerhub for my job. I have not included public_auth field as per this guide since the image type is public and when the cluster is trying to initialize it fails with the error:  Invalid argument: C...

  • 5024 Views
  • 2 replies
  • 1 kudos
Latest Reply
AskeM
New Contributor II
  • 1 kudos

Did you figure this issue out? I am having the exact same problem - I am also having trouble finding thorough documentation/examples on the use of DCS, so if you have stumbled upon any links in that regard, i'd also appreciate it.  

  • 1 kudos
1 More Replies
v_n66
by New Contributor III
  • 3311 Views
  • 8 replies
  • 2 kudos

Resolved! dbutils.notebooks.exit() is not returning results to the variable only on some notebooks

dbutils.notebooks.exit() is not returning results to the variable only on some notebooks issue is Parent child notebook communication and need some solution

  • 3311 Views
  • 8 replies
  • 2 kudos
Latest Reply
v_n66
New Contributor III
  • 2 kudos

@SteveW @lorenzoscandola Its working without any issue from Yesterday. Thank you guys for the support.

  • 2 kudos
7 More Replies
Rishabh-Pandey
by Databricks MVP
  • 17256 Views
  • 8 replies
  • 8 kudos

Resolved! connect databricks to teradata

hey i want to know can we connect databricks to the teradata database and if yes what will be the procedure ??? help would be appreciated

  • 17256 Views
  • 8 replies
  • 8 kudos
Latest Reply
BroData
New Contributor II
  • 8 kudos

There are two main ways to connect to Teradata from Databricks using Python.Way 1: Using Python Libraries (e.g., sqlalchemy, pyjdbc, pyodbc, jaydebeapi, and so on)Pros: Provides a comprehensive solution, allowing us to: Query data, Trigger stored pro...

  • 8 kudos
7 More Replies
Nick_Pacey
by New Contributor III
  • 4860 Views
  • 3 replies
  • 1 kudos

Issue when trying to create a Foreign Catalog to a On Prem SQL Server Instance

Hi,We are creating a lakehouse federated connection to our 2016 On Prem SQL Server.  This has an instance in place, so we only want and need to connect to this instance.  From this connection, we want to create a foreign catalog of a database on the ...

  • 4860 Views
  • 3 replies
  • 1 kudos
Latest Reply
trueray_3150
New Contributor II
  • 1 kudos

Hi @Nick_Pacey  Thank you I already did that using in the codejdbc_url = "jdbc:sqlserver://999.99.999.99\\instance:7777;encrypt=true;trustServerCertificate=true;database=mydatabase"jdbc_username = "myusername"jdbc_password = "mypassword"jdbc_driver =...

  • 1 kudos
2 More Replies
Direo
by Contributor II
  • 40297 Views
  • 6 replies
  • 1 kudos

Resolved! Importing CA certificate into a Databricks cluster

Hi!I was following guide outlined here:https://kb.databricks.com/en_US/python/import-custom-ca-cert(also tried this: https://stackoverflow.com/questions/73043589/configuring-tls-ca-on-databricks)to add ca root certificate into Databricks cluster, but...

  • 40297 Views
  • 6 replies
  • 1 kudos
Latest Reply
jash281098
New Contributor II
  • 1 kudos

@Debayan One question - Will same approach work for JKS file containing private key certificate for X.509 authentication to Mongo Atlas database.Usual way of adding below spark config's is not working. spark.driver.extraJavaOptions -Djavax.net.ssl.ke...

  • 1 kudos
5 More Replies
Stentone
by New Contributor III
  • 1665 Views
  • 3 replies
  • 0 kudos

DLT Direct Publish Mode does not Handle Constraint Dependencies

I'm having some issues with the direct publish mode when defining a DLT workflow that includes tables where their schema defines foreign key constraints. When the foreign constraints reference tables that are not directly defined in any joins of the ...

  • 1665 Views
  • 3 replies
  • 0 kudos
Latest Reply
lingareddy_Alva
Esteemed Contributor
  • 0 kudos

@Stentone This is a tricky situation where you want to leverage the metadata benefits (like the ERD visualization) without running into execution dependencies. Let me help you solve this issue.The error suggests that DLT is trying to validate the for...

  • 0 kudos
2 More Replies
HQJaTu
by New Contributor III
  • 3527 Views
  • 3 replies
  • 2 kudos

Custom container doesn't launch systemd

Quite soon after moving from VMs to containers, I started crafting my own images. That way notebooks have all the necessary libraries already there and no need to do any Pipping/installing in the notebook.As requirements get more complex, now I'm at ...

  • 3527 Views
  • 3 replies
  • 2 kudos
Latest Reply
futurewasfree
New Contributor II
  • 2 kudos

Are there any updates on this? I'm also very interested in having full-fledged Databricks system services integrated into DCS.

  • 2 kudos
2 More Replies
Enrique1987
by New Contributor III
  • 4822 Views
  • 2 replies
  • 3 kudos

Resolved! when to activate photon and when not to ?

Photon appears as an option to check and uncheck as appropriate.The use of Photon leads to higher consumption of DBUs and higher costs.At what point does it pay off and when not to enable it.More costs for the use of photon, but at the same time less...

  • 4822 Views
  • 2 replies
  • 3 kudos
Latest Reply
sunlight
New Contributor II
  • 3 kudos

Hi All,Based on the discussion, can we load a huge flat file (csv) 10 GB using photon accelerated Runtime .Just dump that file into delta from cloud storage like S3 or BlobStorage.Is this one of those ideal usecases for using photon where it will be ...

  • 3 kudos
1 More Replies
Optum
by Databricks Partner
  • 14769 Views
  • 9 replies
  • 5 kudos

Databricks JDBC & Remote Write

Hello,I'm trying to write to a Delta Table in my Databricks instance from a remote Spark session on a different cluster with the Simba Spark driver. I can do reads, but when I attempt to do a write, I get the following error:{  df.write.format("jdbc...

  • 14769 Views
  • 9 replies
  • 5 kudos
Latest Reply
RoK1
New Contributor II
  • 5 kudos

Any update on the issue?

  • 5 kudos
8 More Replies
Fraip
by New Contributor
  • 4190 Views
  • 1 replies
  • 0 kudos

Unable to read files or write to from external location S3 (DataBricks Free Trial)

Hi! I'm trying DataBricks free trial and I tried to link it to an S3 Bucket I set up but I get errors related to serverless policies and unauthorized access whether I tried to read or write to S3, but I have no problem just listing the files that exi...

  • 4190 Views
  • 1 replies
  • 0 kudos
Latest Reply
MuthuLakshmi
Databricks Employee
  • 0 kudos

Your error may be caused by serverless network policy restrictions and/or missing S3 permissions. In the free trial, you cannot use your own S3 buckets with serverless compute. For full access, use a paid workspace and configure both network policy a...

  • 0 kudos
BalaRamesh
by New Contributor II
  • 1233 Views
  • 3 replies
  • 0 kudos

Delta Live tables - If there is no target schema defined , where live tables will create.

Currently i am working Delta live tables.  one of  my ex - team member designed the job and they did not defined in target schema as empty  in destinations (settings -->destination ---> Target Shema). where delta live tables will create if it is empt...

BalaRamesh_0-1746779647873.png
  • 1233 Views
  • 3 replies
  • 0 kudos
Latest Reply
MuthuLakshmi
Databricks Employee
  • 0 kudos

@BalaRamesh If you have catalog specified, there will be storage location for that and you will see this MV created there.Refer to this doc to understand about storage location: https://docs.databricks.com/aws/en/connect/unity-catalog/cloud-storage/m...

  • 0 kudos
2 More Replies
Labels