cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

skosaraju
by New Contributor III
  • 2455 Views
  • 5 replies
  • 1 kudos

Resolved! Getting an error "You cannot use dbutils within a spark job"

I am using mlflow to register my custom model using a simple code below. The DatabricksParams extracts all the params from dbutils and sets the params dictionary and dbutils is not used anywhere else within the rest of my code base. The code fails wh...

  • 2455 Views
  • 5 replies
  • 1 kudos
Latest Reply
lingareddy_Alva
Esteemed Contributor
  • 1 kudos

Thanks for the update @skosaraju .

  • 1 kudos
4 More Replies
daan_dw
by New Contributor III
  • 2133 Views
  • 3 replies
  • 0 kudos

Resolved! Auto Loader for copying files on s3

Hey community,I have a folder on s3 with around 5 million small files. On a daily basis new files are added. I would like to simply copy those new files to another folder on s3. My approach is to use an Auto Loader of which I attached the code below....

  • 2133 Views
  • 3 replies
  • 0 kudos
Latest Reply
daan_dw
New Contributor III
  • 0 kudos

Hey LRALVAThe first time running your code I got the error:  PicklingError: Could not serialize object: Exception: You cannot use dbutils within a spark job You cannot use dbutils within a spark job or otherwise pickle it.So I changed the copy_single...

  • 0 kudos
2 More Replies
asrivas
by New Contributor II
  • 1609 Views
  • 1 replies
  • 0 kudos

Resolved! Is SSL cert needed for Azure Databricks API calls?

Hi, I'm using Azure Databricks and making API calls to the endpoints with verify=False Eg.response = requests.get('https://%s/api/2.0/clusters/list' % (databricks_domain),headers=request_headers,verify=False)Security scanners are flagging the use of ...

  • 1609 Views
  • 1 replies
  • 0 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 0 kudos

In API communication with Azure Databricks, it is not necessary to manually install or configure an SSL certificate on the client side because Azure Databricks endpoints inherently use TLS (Transport Layer Security) for secure communication. To addre...

  • 0 kudos
azimbhuiyan
by New Contributor II
  • 4967 Views
  • 2 replies
  • 1 kudos

Docker image fails to initalize - Databricks Container Service

I am trying to use a public docker image hosted in Dockerhub for my job. I have not included public_auth field as per this guide since the image type is public and when the cluster is trying to initialize it fails with the error:  Invalid argument: C...

  • 4967 Views
  • 2 replies
  • 1 kudos
Latest Reply
AskeM
New Contributor II
  • 1 kudos

Did you figure this issue out? I am having the exact same problem - I am also having trouble finding thorough documentation/examples on the use of DCS, so if you have stumbled upon any links in that regard, i'd also appreciate it.  

  • 1 kudos
1 More Replies
v_n66
by New Contributor III
  • 3130 Views
  • 8 replies
  • 2 kudos

Resolved! dbutils.notebooks.exit() is not returning results to the variable only on some notebooks

dbutils.notebooks.exit() is not returning results to the variable only on some notebooks issue is Parent child notebook communication and need some solution

  • 3130 Views
  • 8 replies
  • 2 kudos
Latest Reply
v_n66
New Contributor III
  • 2 kudos

@SteveW @lorenzoscandola Its working without any issue from Yesterday. Thank you guys for the support.

  • 2 kudos
7 More Replies
Rishabh-Pandey
by Databricks MVP
  • 16951 Views
  • 8 replies
  • 8 kudos

Resolved! connect databricks to teradata

hey i want to know can we connect databricks to the teradata database and if yes what will be the procedure ??? help would be appreciated

  • 16951 Views
  • 8 replies
  • 8 kudos
Latest Reply
BroData
New Contributor II
  • 8 kudos

There are two main ways to connect to Teradata from Databricks using Python.Way 1: Using Python Libraries (e.g., sqlalchemy, pyjdbc, pyodbc, jaydebeapi, and so on)Pros: Provides a comprehensive solution, allowing us to: Query data, Trigger stored pro...

  • 8 kudos
7 More Replies
Nick_Pacey
by New Contributor III
  • 4733 Views
  • 3 replies
  • 1 kudos

Issue when trying to create a Foreign Catalog to a On Prem SQL Server Instance

Hi,We are creating a lakehouse federated connection to our 2016 On Prem SQL Server.  This has an instance in place, so we only want and need to connect to this instance.  From this connection, we want to create a foreign catalog of a database on the ...

  • 4733 Views
  • 3 replies
  • 1 kudos
Latest Reply
trueray_3150
New Contributor II
  • 1 kudos

Hi @Nick_Pacey  Thank you I already did that using in the codejdbc_url = "jdbc:sqlserver://999.99.999.99\\instance:7777;encrypt=true;trustServerCertificate=true;database=mydatabase"jdbc_username = "myusername"jdbc_password = "mypassword"jdbc_driver =...

  • 1 kudos
2 More Replies
Direo
by Contributor II
  • 39865 Views
  • 6 replies
  • 1 kudos

Resolved! Importing CA certificate into a Databricks cluster

Hi!I was following guide outlined here:https://kb.databricks.com/en_US/python/import-custom-ca-cert(also tried this: https://stackoverflow.com/questions/73043589/configuring-tls-ca-on-databricks)to add ca root certificate into Databricks cluster, but...

  • 39865 Views
  • 6 replies
  • 1 kudos
Latest Reply
jash281098
New Contributor II
  • 1 kudos

@Debayan One question - Will same approach work for JKS file containing private key certificate for X.509 authentication to Mongo Atlas database.Usual way of adding below spark config's is not working. spark.driver.extraJavaOptions -Djavax.net.ssl.ke...

  • 1 kudos
5 More Replies
Stentone
by New Contributor III
  • 1487 Views
  • 3 replies
  • 0 kudos

DLT Direct Publish Mode does not Handle Constraint Dependencies

I'm having some issues with the direct publish mode when defining a DLT workflow that includes tables where their schema defines foreign key constraints. When the foreign constraints reference tables that are not directly defined in any joins of the ...

  • 1487 Views
  • 3 replies
  • 0 kudos
Latest Reply
lingareddy_Alva
Esteemed Contributor
  • 0 kudos

@Stentone This is a tricky situation where you want to leverage the metadata benefits (like the ERD visualization) without running into execution dependencies. Let me help you solve this issue.The error suggests that DLT is trying to validate the for...

  • 0 kudos
2 More Replies
HQJaTu
by New Contributor III
  • 3470 Views
  • 3 replies
  • 2 kudos

Custom container doesn't launch systemd

Quite soon after moving from VMs to containers, I started crafting my own images. That way notebooks have all the necessary libraries already there and no need to do any Pipping/installing in the notebook.As requirements get more complex, now I'm at ...

  • 3470 Views
  • 3 replies
  • 2 kudos
Latest Reply
futurewasfree
New Contributor II
  • 2 kudos

Are there any updates on this? I'm also very interested in having full-fledged Databricks system services integrated into DCS.

  • 2 kudos
2 More Replies
Enrique1987
by New Contributor III
  • 4677 Views
  • 2 replies
  • 3 kudos

Resolved! when to activate photon and when not to ?

Photon appears as an option to check and uncheck as appropriate.The use of Photon leads to higher consumption of DBUs and higher costs.At what point does it pay off and when not to enable it.More costs for the use of photon, but at the same time less...

  • 4677 Views
  • 2 replies
  • 3 kudos
Latest Reply
sunlight
New Contributor II
  • 3 kudos

Hi All,Based on the discussion, can we load a huge flat file (csv) 10 GB using photon accelerated Runtime .Just dump that file into delta from cloud storage like S3 or BlobStorage.Is this one of those ideal usecases for using photon where it will be ...

  • 3 kudos
1 More Replies
Optum
by Databricks Partner
  • 14479 Views
  • 9 replies
  • 5 kudos

Databricks JDBC & Remote Write

Hello,I'm trying to write to a Delta Table in my Databricks instance from a remote Spark session on a different cluster with the Simba Spark driver. I can do reads, but when I attempt to do a write, I get the following error:{  df.write.format("jdbc...

  • 14479 Views
  • 9 replies
  • 5 kudos
Latest Reply
RoK1
New Contributor II
  • 5 kudos

Any update on the issue?

  • 5 kudos
8 More Replies
Fraip
by New Contributor
  • 4044 Views
  • 1 replies
  • 0 kudos

Unable to read files or write to from external location S3 (DataBricks Free Trial)

Hi! I'm trying DataBricks free trial and I tried to link it to an S3 Bucket I set up but I get errors related to serverless policies and unauthorized access whether I tried to read or write to S3, but I have no problem just listing the files that exi...

  • 4044 Views
  • 1 replies
  • 0 kudos
Latest Reply
MuthuLakshmi
Databricks Employee
  • 0 kudos

Your error may be caused by serverless network policy restrictions and/or missing S3 permissions. In the free trial, you cannot use your own S3 buckets with serverless compute. For full access, use a paid workspace and configure both network policy a...

  • 0 kudos
BalaRamesh
by New Contributor II
  • 1158 Views
  • 3 replies
  • 0 kudos

Delta Live tables - If there is no target schema defined , where live tables will create.

Currently i am working Delta live tables.  one of  my ex - team member designed the job and they did not defined in target schema as empty  in destinations (settings -->destination ---> Target Shema). where delta live tables will create if it is empt...

BalaRamesh_0-1746779647873.png
  • 1158 Views
  • 3 replies
  • 0 kudos
Latest Reply
MuthuLakshmi
Databricks Employee
  • 0 kudos

@BalaRamesh If you have catalog specified, there will be storage location for that and you will see this MV created there.Refer to this doc to understand about storage location: https://docs.databricks.com/aws/en/connect/unity-catalog/cloud-storage/m...

  • 0 kudos
2 More Replies
eyalo
by New Contributor II
  • 2419 Views
  • 1 replies
  • 0 kudos

Ingest from FTP server doesn't work

Hi,I am trying to connect my FTP server and store the files to a dataframe with the following code:%pip install ftputilfrom ftputil import FTPHostHost = "92.118.67.49"Login = "StrideNBM-DF_BO"Passwd = "Sdf123456"ftp_dir = "/dwh-reports/"with FTPHost(...

  • 2419 Views
  • 1 replies
  • 0 kudos
Latest Reply
Kayla
Valued Contributor II
  • 0 kudos

I'm afraid I don't have an answer, and I know this is an old post, but if you haven't already, I'd recommend changing the password if that is/was a genuine password.

  • 0 kudos
Labels