cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

NextIT
by New Contributor
  • 715 Views
  • 0 replies
  • 0 kudos

www.nextitvision.com

Online IT Training: ERP/SAP Online Training | JAVA Online Training | C++Online Training | ORACLE Online Training | Online Python Training | Machine Learning Training. If you Need more Details and Information Regarding IT Online Training. Please Visi...

  • 715 Views
  • 0 replies
  • 0 kudos
sh_abrishami_ie
by New Contributor II
  • 4654 Views
  • 1 replies
  • 3 kudos

Resolved! Driver is up but is not responsive, likely due to GC.

Hi,I have a problem with writing an excel file into the mounted file.after 10 mins I see the Driver is up but is not responsive, likely due to GC on the log events.I'm using the following script:df.repartition(1).write .format("com.crealytics.spark....

  • 4654 Views
  • 1 replies
  • 3 kudos
Latest Reply
Hubert-Dudek
Esteemed Contributor III
  • 3 kudos

It is not solution to that problem but I recommend to handle excel reads and writes with Spark Koalas https://koalas.readthedocs.io/en/latest/reference/api/databricks.koalas.DataFrame.to_excel.html just give it a try maybe it will solve your issue

  • 3 kudos
Robbie
by New Contributor III
  • 2439 Views
  • 1 replies
  • 2 kudos

How can I avoid this 'java.sql.SQLException: Too many connections' error?

I'm having difficulty with a job (parent) that triggers multiple parallel runs of another job (child) in batches (e.g. 10 parallel runs per batch).Occasionally some of the parallel "child" jobs will crash a few minutes in-- either during or immediate...

  • 2439 Views
  • 1 replies
  • 2 kudos
Latest Reply
Hubert-Dudek
Esteemed Contributor III
  • 2 kudos

It is MariaDB JDBC error so probably database which you are trying to connect can not handle this amount of concurrent connections (alternatively if you are not connecting to MariaDB databse, MariaDB is used also for hive metastore in your case maria...

  • 2 kudos
bchaubey
by Contributor II
  • 1099 Views
  • 1 replies
  • 1 kudos

Azure Databricks Certification

@Hubert Dudek​  what is the Certification name of Azure Databricks?

  • 1099 Views
  • 1 replies
  • 1 kudos
Latest Reply
Hubert-Dudek
Esteemed Contributor III
  • 1 kudos

Hi @Bhagwan Chaubey​ ,There is Spark developer certification from Databricks - https://databricks.com/learn/training/home (and some higher levels as well)In Azure databricks is included in DP-100 and DP-203 certification (together with around 10 diff...

  • 1 kudos
Ashish
by New Contributor II
  • 7440 Views
  • 4 replies
  • 3 kudos

Resolved! Cost of individual jobs running on a shared Databricks cluster

Hi All,I am working on a requirement where I need to calculate the cost of each spark job individually on a shared Azure/AWS Databricks cluster. There can be multiple jobs running on the cluster parallelly.Cost needs to be calculated after job comple...

  • 7440 Views
  • 4 replies
  • 3 kudos
Latest Reply
alexott
Databricks Employee
  • 3 kudos

There is a built-in functionality for getting the costs:AWS - https://docs.databricks.com/administration-guide/account-settings-e2/usage.htmlAzure - via built-in Cost Management + BillingThe main problem with that functionality is that the smallest g...

  • 3 kudos
3 More Replies
Autel
by New Contributor II
  • 3968 Views
  • 3 replies
  • 0 kudos

Resolved! concurrent update to same hive or deltalake table

HI, I'm interested to know if multiple executors to append the same hive table using saveAsTable or insertInto sparksql. will that cause any data corruption? What configuration do I need to enable concurrent write to same hive table? what about the s...

  • 3968 Views
  • 3 replies
  • 0 kudos
Latest Reply
-werners-
Esteemed Contributor III
  • 0 kudos

The Hive table will not like this, as the underlying data is parquet format which is not ACID compliant.Delta lake however is:https://docs.delta.io/0.5.0/concurrency-control.htmlYou can see that inserts do not give conflicts.

  • 0 kudos
2 More Replies
Maverick1
by Valued Contributor II
  • 26307 Views
  • 9 replies
  • 7 kudos

Resolved! How to deploy a databricks managed workspace model to sagemaker from databricks notebook

I wanted to deploy a registered model present in databricks managed MLFlow to a sagemaker via databricks notebook?As of now, it is not able to run mlflow sagemaker build-and-push container command directly. What all configurations or steps needed to ...

  • 26307 Views
  • 9 replies
  • 7 kudos
Latest Reply
User16871418122
Contributor III
  • 7 kudos

@Saurabh Verma​ Please try!import mlflow.sagemaker as mfs sys.stdout.fileno = lambda: 0 mfs.run_local(model_uri=model_uri,port=8000,image="test")

  • 7 kudos
8 More Replies
cbynum
by New Contributor III
  • 3332 Views
  • 4 replies
  • 1 kudos

Resolved! Terraform authentication with SSO enabled

After enabling SSO on my account I now don't have any way to change my terraform for provisioning AWS workspaces because username/password is disabled. Is there a workaround for this?

  • 3332 Views
  • 4 replies
  • 1 kudos
Latest Reply
cbynum
New Contributor III
  • 1 kudos

Never mind, the account owner creds do work, but I had to add the account owner to all of the workspaces. The terraform didn't give me an informative error, it just hung forever when applying.

  • 1 kudos
3 More Replies
Ketna
by New Contributor
  • 1622 Views
  • 1 replies
  • 0 kudos

I have included SparkJDBC42.jar in my war file. but when i start my application using tomcat, i get EOFExceptions from log4j classes. I need help with what is causing this and How to resolve this issue? Please help.

Below is part of the exceptions I am getting:org.apache.catalina.startup.ContextConfig processAnnotationsJarSEVERE: Unable to process Jar entry [com/simba/spark/jdbc42/internal/apache/logging/log4j/core/pattern/ThreadIdPatternConverter.class] from Ja...

  • 1622 Views
  • 1 replies
  • 0 kudos
Latest Reply
Anonymous
Not applicable
  • 0 kudos

Hello, @Ketna Khalasi​ ! My name is Piper, and I'm a moderator here at Databricks. Thank you for posting your question and I'm sorry to hear you're having this problem. We generally give the community a chance to respond before jumping in. Thanks in ...

  • 0 kudos
Anonymous
by Not applicable
  • 6553 Views
  • 2 replies
  • 4 kudos

Cluster does not have proper permissions to view DBFS mount point to Azure ADLS Gen 2.

I've created other mount points and am now trying to use the OAUTH method. I'm able to define the mount point using the OAUTH Mount to ADLS Gen 2 Storage.I've created an App Registration with Secret, added the App Registration as Contributor to the ...

  • 6553 Views
  • 2 replies
  • 4 kudos
Latest Reply
Gerbastanovic
New Contributor II
  • 4 kudos

Also check if you set the right permissions for the app on the containers ACLhttps://docs.microsoft.com/en-us/azure/storage/blobs/data-lake-storage-access-control

  • 4 kudos
1 More Replies
Erik
by Valued Contributor III
  • 3574 Views
  • 5 replies
  • 2 kudos

Resolved! Powerbi databricks connector should import column description

I posted this idea in ideas.powerbi.com as well, but it is quite unclear to me whether the powerbi databricks connector is in fact made by MS or Databricks, so I post it here as well!It is possible to add comments/descriptions to databricks database ...

  • 3574 Views
  • 5 replies
  • 2 kudos
Latest Reply
Atanu
Databricks Employee
  • 2 kudos

@Erik Parmann​  the connector is the colab product of MS and Databricks. But I feel this is a nice feature to add. I request to raise a feature request with us from here - https://ideas.databricks.com/ . Our product team definitely will take a look a...

  • 2 kudos
4 More Replies
Matt_Johnston
by New Contributor III
  • 4896 Views
  • 4 replies
  • 4 kudos

Resolved! Disk Type in Azure Databricks

Hi There,How are the disks tiers determined in Azure Databricks? We are currently using a pool which is using Standard DS3 v2 Virtual Machines, all with Premium SSD disks. Is there a way to change the tier of the disks?Thanks

  • 4896 Views
  • 4 replies
  • 4 kudos
Latest Reply
Atanu
Databricks Employee
  • 4 kudos

I think we do not have option to change the disk type at this moment. but I would like to request you to raise a feature request through azure support if you are azure databricks user. if aws you can do the same from - https://docs.databricks.com/res...

  • 4 kudos
3 More Replies
Shridhar
by New Contributor
  • 16063 Views
  • 2 replies
  • 2 kudos

Resolved! Load multiple csv files into a dataframe in order

I can load multiple csv files by doing something like: paths = ["file_1", "file_2", "file_3"] df = sqlContext.read .format("com.databricks.spark.csv") .option("header", "true") .load(paths) But this doesn't seem to preserve the...

  • 16063 Views
  • 2 replies
  • 2 kudos
Latest Reply
Jaswanth_Saniko
New Contributor III
  • 2 kudos

val diamonds = spark.read.format("csv") .option("header", "true") .option("inferSchema", "true") .load("/FileStore/tables/11.csv","/FileStore/tables/12.csv","/FileStore/tables/13.csv")   display(diamonds)This is working for me @Shridhar​ 

  • 2 kudos
1 More Replies
Itachi_Naruto
by New Contributor II
  • 8754 Views
  • 3 replies
  • 0 kudos

hdbscan package error

I try to import **hdbscan** but it throws this following error/databricks/python_shell/dbruntime/PythonPackageImportsInstrumentation/__init__.py in import_patch(name, globals, locals, fromlist, level) 156 # Import the desired module. ...

  • 8754 Views
  • 3 replies
  • 0 kudos
Latest Reply
Atanu
Databricks Employee
  • 0 kudos

does this help @Rajamannar Aanjaram​ ?

  • 0 kudos
2 More Replies

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.

If there isn’t a group near you, start one and help create a community that brings people together.

Request a New Group
Labels