cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Taha_Hussain
by Databricks Employee
  • 1298 Views
  • 0 replies
  • 4 kudos

Databricks Office Hours Register for Office Hours to participate in a live Q&A session with Databricks experts! Our next event is scheduled for Ju...

Databricks Office HoursRegister for Office Hours to participate in a live Q&A session with Databricks experts! Our next event is scheduled for June 22nd from 8:00 am - 9:00am PT.This is your opportunity to connect directly with our experts to ask any...

  • 1298 Views
  • 0 replies
  • 4 kudos
wyzer
by Contributor II
  • 5017 Views
  • 4 replies
  • 2 kudos

Resolved! Unable to delete a DBFS folder

Hello everyone,I've created by error a DBFS folder named : ${env]But when I run this command :dbutils.fs.rm("/mnt/${env]")It returns me this error : java.net.URISyntaxException: Illegal character in path at index 12: /mnt/$%7Benv]How can I do please ...

  • 5017 Views
  • 4 replies
  • 2 kudos
Latest Reply
User16764241763
Honored Contributor
  • 2 kudos

Hello @Salah K.​ Can you try below?%sh rm -r /dbfs/mnt/$\{env\]

  • 2 kudos
3 More Replies
shubhamb
by New Contributor III
  • 4630 Views
  • 2 replies
  • 3 kudos

Why does my Notebook fails if I try to load a function from another Notebook in Repos in Databricks

My function in func.pydef lower_events(df): return df.withColumn("event",f.lower(f.col("event")))My main notebook import pyspark.sql.functions as f from pyspark.sql.functions import udf, col, lower import sys   sys.path.append("..") from folder.func...

  • 4630 Views
  • 2 replies
  • 3 kudos
Latest Reply
shubhamb
New Contributor III
  • 3 kudos

@Kaniz Fatma​  https://community.databricks.com/s/question/0D58Y00008ouo6xSAA/how-to-fetch-environmental-variables-saved-in-one-notebook-into-another-notebook-in-databricks-repos-and-notebooksCan you please look into this

  • 3 kudos
1 More Replies
mortenhaga
by Contributor
  • 3752 Views
  • 2 replies
  • 2 kudos

Resolved! Importing python function with spark.read.jdbc in to Repos

Hi all!Before we used Databricks Repos we used the run magic to run various utility python functions from one notebook inside other notebooks, fex like reading from a jdbc connections. We now plan to switch to repos to utilize the fantastic CI/CD pos...

image.png image
  • 3752 Views
  • 2 replies
  • 2 kudos
Latest Reply
mortenhaga
Contributor
  • 2 kudos

Thats...odd. I was sure I had tried that, but now it works somehow. I guess it has to be that now I did it with double quotation marks. Thanks anyway! Works like a charm.

  • 2 kudos
1 More Replies
Marra
by New Contributor III
  • 8858 Views
  • 7 replies
  • 2 kudos

Read temporary views in SQL Analytics

I'm having issues trying to read temporary views in SQL Analytics module. Ive managed to create temporary views based on a query but I don't know how to read from them? Just using the name of the view returns "Table or view not found".

  • 8858 Views
  • 7 replies
  • 2 kudos
Latest Reply
Marra
New Contributor III
  • 2 kudos

No, I'm actually having issues reading from the view in the same session that created it. Using the same view name I get a table or view not found.

  • 2 kudos
6 More Replies
AndriusVitkausk
by New Contributor III
  • 2076 Views
  • 1 replies
  • 1 kudos

Autoloader event vs directory ingestion

For a production work load containing around 15k gzip compressed json files per hour all in a YYYY/MM/DD/HH/id/timestamp.json.gz directoryWhat would be the better approach on ingesting this into a delta table in terms of not only the incremental load...

  • 2076 Views
  • 1 replies
  • 1 kudos
Latest Reply
AndriusVitkausk
New Contributor III
  • 1 kudos

@Kaniz Fatma​ So i've not found a fix for the small file problem using autoloader, seems to struggle really badly against large directories, had a cluster running for 8h stuck on "listing directory" part with no end, cluster seemed completely idle to...

  • 1 kudos
karthikeyanr
by New Contributor II
  • 6784 Views
  • 4 replies
  • 6 kudos

Unable to import .dbc files in Databricks for "Databricks Developer Foundation Capstone"

Hi,I am not able to import .dbc file into Databricks workspace for "Databricks Developer Foundation Capstone". When I click import the error message is displayed. Secondly when I click the github link in the 'Download the capstone' error 404 is displ...

errorgit import error
  • 6784 Views
  • 4 replies
  • 6 kudos
Latest Reply
Atanu
Databricks Employee
  • 6 kudos

Hello @Karthikeyan.r3@cognizant.com R​ I agreed with Hubert. please write to  https://help.databricks.com

  • 6 kudos
3 More Replies
Constantine
by Contributor III
  • 5070 Views
  • 1 replies
  • 1 kudos

Resolved! Can we reuse checkpoints in Spark Streaming?

I am reading data from a Kafka topic, say topic_a. I have an application, app_one which uses Spark Streaming to read data from topic_a. I have a checkpoint location, loc_a to store the checkpoint. Now, app_one has read data till offset 90.Can I creat...

  • 5070 Views
  • 1 replies
  • 1 kudos
Latest Reply
jose_gonzalez
Databricks Employee
  • 1 kudos

Hi @John Constantine​,Is not recommended to share the checkpoint with your queries. Every streaming query should have their own checkpoint. If you can to start at the offset 90 in another query, then you can define it when starting your job. You can ...

  • 1 kudos
bhargavi1
by New Contributor II
  • 2027 Views
  • 1 replies
  • 1 kudos
  • 2027 Views
  • 1 replies
  • 1 kudos
Latest Reply
jose_gonzalez
Databricks Employee
  • 1 kudos

Hi,Could you share more details on what you have tried? please provide more details.

  • 1 kudos
pbezz
by New Contributor III
  • 9782 Views
  • 5 replies
  • 9 kudos

Visualisation libraries does not work on Databricks

Why is it that certain Python visualisation libraries do not work on Databricks? I am trying to install (via pip) and work with some data visualisation libraries - they work perfectly in a normal Jupyter Notebook but not on a Databricks notebook envi...

  • 9782 Views
  • 5 replies
  • 9 kudos
Latest Reply
pbezz
New Contributor III
  • 9 kudos

No switched to using html widgets.

  • 9 kudos
4 More Replies
Manoj
by Contributor II
  • 20581 Views
  • 13 replies
  • 6 kudos

Resolved! Did any one in the community create Permanent Functions using using python script ? I tried but i am getting the below error, Please advise

Hi Team, When i am trying to register a permanant function i am getting the below error.%sqlCREATE FUNCTION simple_udf AS 'SimpleUdf'  USING JAR '/tmp/SimpleUdf.jar';%sqlselect simple_udf(2)Error Details : com.databricks.backend.common.rpc.Databricks...

  • 20581 Views
  • 13 replies
  • 6 kudos
Latest Reply
Manoj
Contributor II
  • 6 kudos

hi @Werner Stinckens​  @Jose Gonzalez​  @Hubert Dudek​ @Kaniz Fatma​ â€‹Thanks for all the help, Appreciate it. I was able to create permanent functions and use eclipse to create the runnable jar. However, Does anyone have any idea on how to deploy t...

  • 6 kudos
12 More Replies
rbarrero
by New Contributor III
  • 7369 Views
  • 9 replies
  • 7 kudos

Resolved! Error saving changes on Job Cluster

Hello all and thanks.After apply to serving a model, I go to edit corresponding Job Cluster to configure its init_script but when I try to save changes (Confirm and restart) it thrown the following error:Error: Cannot edit cluster 0503-141315-hu3wd4i...

  • 7369 Views
  • 9 replies
  • 7 kudos
Latest Reply
rbarrero
New Contributor III
  • 7 kudos

Sorry for the delay in responding. Finally a partner could fix the problem, he can edit without problems the cluster and add the init_script.Thank you!

  • 7 kudos
8 More Replies
mick042
by New Contributor III
  • 2597 Views
  • 1 replies
  • 3 kudos

Resolved! Spark Executor Rdd.Pipe call not finding file location that exists in Sparkfiles.get()

In a Databricks notebook, I need to run text files (stdin, stdout) through a function from an external library. I have used sparkContext.AddFiles({external_library_name}) to add the external library so that it is available to all executors.when I run...

  • 2597 Views
  • 1 replies
  • 3 kudos
Latest Reply
mick042
New Contributor III
  • 3 kudos

Thanks Kaniz, yes I tried that. Did not work. Falling back on init scripts now and that works.

  • 3 kudos
alefred
by New Contributor II
  • 5583 Views
  • 3 replies
  • 2 kudos

Resolved! Automation creation of Databricks Top Level Repositories

I'm looking to automate the creation of Top Level repositories in Databricks however isn't possible using cli or API if this repo is private repository(Azure DevOps Repository) because require setup the token in user setting.databricks repos create \...

  • 5583 Views
  • 3 replies
  • 2 kudos
Latest Reply
Atanu
Databricks Employee
  • 2 kudos

We have API availabe for repos https://docs.databricks.com/dev-tools/api/latest/repos.html#operation/get-repos also we are currently supporting SP. Step 1: As an admin Create a Service PrincipalUse this API SCIM API 2.0 (ServicePrincipals) | Databric...

  • 2 kudos
2 More Replies
RajibRajib_Mand
by New Contributor III
  • 5640 Views
  • 7 replies
  • 6 kudos

Unicode field separator to create unamanged table in databricks for csv file

We are getting \u318a (ㆊ) separated csv file. We want to create unmanaged table in databricks, Here is the table creation script.create table IF NOT EXISTS db_test_raw.t_data_otc_poc (`caseidt` String,`worktype` String,`doctyp` String,`brand` String,...

  • 5640 Views
  • 7 replies
  • 6 kudos
Latest Reply
Hubert-Dudek
Esteemed Contributor III
  • 6 kudos

sep "\u318a"delimeter " \x318a"sep " \x318a"Try to use sep instead or/and x instead.

  • 6 kudos
6 More Replies

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels