cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

aline_alvarez
by New Contributor III
  • 7251 Views
  • 6 replies
  • 7 kudos

Resolved! How can I delete a file in DBFS with Illegal character?

How can I delete a file in DBFS with Illegal character?Someone put the file named "planejamento_[4098.]___SHORT_SAIA_JEANS__.xlsx" inside the folder /FileStore and I can delete it, because of this error: java.net.URISyntaxException: Illegal character...

  • 7251 Views
  • 6 replies
  • 7 kudos
Latest Reply
-werners-
Esteemed Contributor III
  • 7 kudos

try this %sh ls -li /dbfsif the file is located in a subdirectory you can change the path mentioned above.the %sh magic command gives you access to linux shell commands.

  • 7 kudos
5 More Replies
EDDatabricks
by Databricks Partner
  • 8577 Views
  • 2 replies
  • 0 kudos

Resolved! Pool Max Capacity vs Cluster Max Workers

Hi all, we have a databricks instance on Azure with a Compute Cluster version 7.3 LTS. Currently the cluster has 4 max workers (min workers: 1) of type: Standard_D13_v2 and 1 driver of the same type. There are several jobs that are running on this cl...

  • 8577 Views
  • 2 replies
  • 0 kudos
Latest Reply
Anonymous
Not applicable
  • 0 kudos

Hi @EDDatabricks EDDatabricks​ Hope all is well! Just wanted to check in if you were able to resolve your issue and would you be happy to share the solution or mark an answer as best? Else please let us know if you need more help. We'd love to hear f...

  • 0 kudos
1 More Replies
tinendra
by New Contributor III
  • 5952 Views
  • 5 replies
  • 5 kudos

How to reduce time while loading data into the azure synapse table?

Hi All,I just wanted to know if is there any option to reduce time while loading Pyspark Dataframe into the Azure synapse table using Databricks.like..I have a pyspark dataframe that has around 40k records and I am trying to load data into the azure ...

  • 5952 Views
  • 5 replies
  • 5 kudos
Latest Reply
Anonymous
Not applicable
  • 5 kudos

Hi @Tinendra Kumar​ Hope all is well! Just wanted to check in if you were able to resolve your issue and would you be happy to share the solution or mark an answer as best? Else please let us know if you need more help. We'd love to hear from you.Tha...

  • 5 kudos
4 More Replies
Murthy1
by Contributor II
  • 9774 Views
  • 3 replies
  • 3 kudos

Resolved! Impacts of running multiple jobs in parallel that refers the same notebook

Can I run multiple jobs(for example: 100+) in parallel that refers the same notebook? I supply each job with a different parameter. If we can do this, what would be the impact? (for example: reliability, performance, troubleshooting etc. )Example: N...

  • 9774 Views
  • 3 replies
  • 3 kudos
Latest Reply
Anonymous
Not applicable
  • 3 kudos

Hi @Murthy Ramalingam​ Hope all is well! Just wanted to check in if you were able to resolve your issue and would you be happy to share the solution or mark an answer as best? Else please let us know if you need more help. We'd love to hear from you....

  • 3 kudos
2 More Replies
venkat-bodempud
by New Contributor III
  • 4063 Views
  • 4 replies
  • 3 kudos

Resolved! Databricks Design Help

Hello Community,I am currently working on populating gold layer tables. Source for these gold layer tables are silver layer tables. A query is going to run on silver layer tables, spark sql query contains joins between multiple tables.ex:select colum...

  • 4063 Views
  • 4 replies
  • 3 kudos
Latest Reply
Anonymous
Not applicable
  • 3 kudos

Hi @bodempudi venkat​ Hope all is well! Just wanted to check in if you were able to resolve your issue and would you be happy to share the solution or mark an answer as best? Else please let us know if you need more help. We'd love to hear from you.T...

  • 3 kudos
3 More Replies
MelMazz
by Databricks Partner
  • 4243 Views
  • 2 replies
  • 0 kudos

Data in transit security when using jdbc connections.

When using spark.read.jdbc() for a direct connection, how is data in transit secured?

  • 4243 Views
  • 2 replies
  • 0 kudos
Latest Reply
Anonymous
Not applicable
  • 0 kudos

Hi @Melody Mazaiwana​ Hope all is well! Just wanted to check in if you were able to resolve your issue and would you be happy to share the solution or mark an answer as best? Else please let us know if you need more help. We'd love to hear from you.T...

  • 0 kudos
1 More Replies
167559
by Databricks Partner
  • 4058 Views
  • 3 replies
  • 4 kudos

Internal error message: Failed to launch spark container on instance i-xxxxxxxxxxxxx. Exception: Unexpected internal error, please contact Databricks support

I have a docker image for debezium in my ECR repo and added IAM roles from Databricks to pull this image for my cluster and seeing this error when cluster is created

  • 4058 Views
  • 3 replies
  • 4 kudos
Latest Reply
167559
Databricks Partner
  • 4 kudos

I have been trying for more than a week now..... where is the logs for these pulls,....error message is not helpful

  • 4 kudos
2 More Replies
vinaykumar
by Databricks Partner
  • 4418 Views
  • 1 replies
  • 0 kudos

Resolved! Pros and cons - running SQL query in databricks notebook and serverless warehouse sql editor

Can you please help technical pros and cons for running SQL query in databricks notebook (data engineering workspace) and serverless warehouse SQL editor

  • 4418 Views
  • 1 replies
  • 0 kudos
Latest Reply
Anonymous
Not applicable
  • 0 kudos

NotebookPROS: More traditional cluster, git integration, choose DBR versionCONS: Cluster startup time, photon not automatically part of the clusterServerlessPROS: Faster, almost immediate startup time, less expensive for a single query, photon enable...

  • 0 kudos
ABVectr
by New Contributor III
  • 5643 Views
  • 6 replies
  • 1 kudos

Resolved! Maven Package install failing on DBR 11.3 LTS

Hi Databricks Community,I ran into the following issue when setting up a new cluster with the latest LTS Databricks runtime (11.3). When trying to install the package with the coordinates com.microsoft.azure.kusto:kusto-spark_3.0_2.12:3.1.4 from Mave...

  • 5643 Views
  • 6 replies
  • 1 kudos
Latest Reply
Anonymous
Not applicable
  • 1 kudos

Hi @Andrei Bondarenko​ Hope all is well! Just wanted to check in if you were able to resolve your issue and would you be happy to share the solution or mark an answer as best? Else please let us know if you need more help. We'd love to hear from you....

  • 1 kudos
5 More Replies
User16270906190
by Databricks Employee
  • 7723 Views
  • 8 replies
  • 5 kudos

customer is trying to generate a Databricks token for a service principal (SP). They’ve created the SP in Azure AD and have used the Databricks rest ...

customer is trying to generate a Databricks token for a service principal (SP). They’ve created the SP in Azure AD and have used the Databricks rest api to add it as an admin. When using the Databricks rest api /api/2.0/token-management/on-behalf-of...

  • 7723 Views
  • 8 replies
  • 5 kudos
Latest Reply
Buxert
New Contributor II
  • 5 kudos

Having the same problem here. @Nitisha Nigam​ did you solve it?

  • 5 kudos
7 More Replies
cmilligan
by Contributor II
  • 3203 Views
  • 2 replies
  • 0 kudos

Delay in retrieving task variables from a notebook utilizing %run

I have a databricks workflow where the first task is to set up task parameters for other notebooks to use later in the process. Since these are variables that are used in all of my notebooks, I have opted to assign them in a shared notebook and call ...

  • 3203 Views
  • 2 replies
  • 0 kudos
Latest Reply
Debayan
Databricks Employee
  • 0 kudos

Hi, Please refer: https://docs.databricks.com/notebooks/notebook-workflows.htmlCould you please provide the error received?

  • 0 kudos
1 More Replies
Jayesh
by New Contributor III
  • 21209 Views
  • 13 replies
  • 13 kudos

How to deploy Azure Databricks Sql (Table, Views, etc) into higher environments?

We have build multiple tables and Views under Databricks Sql and unable to figure out how can we take this code and deploy at our higher environments? Need some guidance as we're unable to get any information from searched documentations.

  • 21209 Views
  • 13 replies
  • 13 kudos
Latest Reply
mmlime
New Contributor III
  • 13 kudos

Hi, I agree with @Josef Prakljacic​ . If Databricks would like to compete SQL DWH/Synapse or Snowflake and target DWH users, It should prepare some guidelines how to manage "database" objects. Yea, @Werner Stinckens​ with Engineering workspace and py...

  • 13 kudos
12 More Replies
theSoyf
by New Contributor II
  • 5864 Views
  • 2 replies
  • 1 kudos

How to write to Salesforce object using Spark Salesforce Library

Hi I'm facing an issue when writing to a salesforce object. I'm using the springml/spark-salesforce library. I have the above libraries installed as recommended based on my research.I try to write like this:(_sqldf .write .format("com.springml.spar...

Screen Shot 2022-12-14 at 8.18.07 AM
  • 5864 Views
  • 2 replies
  • 1 kudos
Latest Reply
Gauthy
Databricks Partner
  • 1 kudos

Im facing the same issue while trying to write to Salesforce, if you have found a resolution could you please share it ?

  • 1 kudos
1 More Replies
knawara
by Contributor
  • 5776 Views
  • 4 replies
  • 1 kudos

Delta Live Tables: reading from output

I'm trying to implement an incremental ingestion logic in the following way:database tables have DbUpdatedDate columnDuring initial load I perform a full copy of the database tableDuring incremental load I:scan the data already in the DLT to see what...

  • 5776 Views
  • 4 replies
  • 1 kudos
Latest Reply
fecavalc08
New Contributor III
  • 1 kudos

Hi @Chris Nawara​, I had the same issue you had. I was trying to avoid the apply_changes but we in the end I implemented it and I'm happier that I expected heheand if you have any additional standardization columns that you need to implement, you can...

  • 1 kudos
3 More Replies
Bie1234
by New Contributor III
  • 3684 Views
  • 2 replies
  • 3 kudos

Resolved! accidently delete paquet file in dbfs

I accidently  delete manual paquet file in dbfs how can I recovery this recovery this file

  • 3684 Views
  • 2 replies
  • 3 kudos
Latest Reply
Ajay-Pandey
Databricks MVP
  • 3 kudos

Hi @pansiri panaudom​ ,There is no option restore deleted files in databricks .

  • 3 kudos
1 More Replies
Labels