cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

NWIEFInance
by New Contributor
  • 2346 Views
  • 1 replies
  • 0 kudos

Connect to EXCEL

I have hardtime connecting my existing EXCEL file to source data from DataBricks and need help

  • 2346 Views
  • 1 replies
  • 0 kudos
Latest Reply
User16539034020
Databricks Employee
  • 0 kudos

Hi, Thanks for contacting Databricks Support. We doesn't support direct Excel-Databricks connectivity. However, Databricks can be accessed through ODBC and JDBC interfaces, and we can leverage these with Excel's Power Query functionality for indirect...

  • 0 kudos
matanper
by New Contributor III
  • 3973 Views
  • 5 replies
  • 1 kudos

Custom docker image fails to initalize

I'm trying to use a custom docker image for my job. This is my docker file:FROM databricksruntime/standard:12.2-LTS COPY . . RUN /databricks/python3/bin/pip install -U pip RUN /databricks/python3/bin/pip install -r requirements.txt USER rootMy job ...

  • 3973 Views
  • 5 replies
  • 1 kudos
Latest Reply
Debayan
Databricks Employee
  • 1 kudos

Hi, I think, disabling iptables will be better in this case, could you please try the below command and confirm? $ sudo iptables -S

  • 1 kudos
4 More Replies
Łukasz
by New Contributor III
  • 4718 Views
  • 6 replies
  • 5 kudos

Resolved! Dense rank possible bug

I have the case of deduplicating data source over specific business key using dense_rank function. Currently the data source does not have any duplicates, so the function should return 1 in all cases. The issue is that dense rank does not return prop...

  • 4718 Views
  • 6 replies
  • 5 kudos
Latest Reply
saipujari_spark
Databricks Employee
  • 5 kudos

Hey @Łukasz Thanks for reporting.As I see Spark 3.4.0 introduced an improvement that looks to be the cause for this issue.Improvement: https://issues.apache.org/jira/browse/SPARK-37099Similar Bug: https://issues.apache.org/jira/browse/SPARK-44448This...

  • 5 kudos
5 More Replies
415963
by New Contributor II
  • 3105 Views
  • 3 replies
  • 2 kudos

Not able to catch structured streaming exception

I would like to catch and handle an exception in a structured streaming job.The databricks notebook still displays the exception, regardless of added exception handling (see attached screenshot)I guess that the exception is displayed by the cell outp...

  • 3105 Views
  • 3 replies
  • 2 kudos
Latest Reply
Debayan
Databricks Employee
  • 2 kudos

Hi, I understand, could you please also provide the last line of the error after scrolling down in the notebook cell? 

  • 2 kudos
2 More Replies
Retko
by Contributor
  • 7812 Views
  • 4 replies
  • 2 kudos

Running Command is often stuck on "Running Command..."

Hi,when running command, it often gets stuck and message below it says: "Running Command..."What can I do with it besides of restarting cluster?Also tried reattaching and clearing state, but no help here.Thanks

  • 7812 Views
  • 4 replies
  • 2 kudos
Latest Reply
Debayan
Databricks Employee
  • 2 kudos

Hi, do you see this while running a command in the notebook? Please tag @Debayan  with your next comment which will notify me. Thanks!

  • 2 kudos
3 More Replies
DennisB
by New Contributor III
  • 4133 Views
  • 4 replies
  • 2 kudos

Resolved! Better Worker Node Core Utilisation

Hi everyone,Hoping someone can help me with this problem. I have an embarrassingly parallel workload, which I'm parallelising over 4 worker nodes (of type Standard_F4, so 4 cores each). Each workload is single-threaded, so I believe that only one cor...

  • 4133 Views
  • 4 replies
  • 2 kudos
Latest Reply
DennisB
New Contributor III
  • 2 kudos

So I managed to get the 1-core-per-executor working successfully. The bit that wasn't working was spark.executor.memory -- this was too high, but lowering it so that the sum of the executors memory was ~90% of the worker node's memory allowed it to w...

  • 2 kudos
3 More Replies
Priyag1
by Honored Contributor II
  • 2228 Views
  • 2 replies
  • 11 kudos

Query parameters in dashboardsQueries can optionally leverage parameters or static values. When a visualization based on a parameterized query is adde...

Query parameters in dashboardsQueries can optionally leverage parameters or static values. When a visualization based on a parameterized query is added to a dashboard, the visualization can either be configured to use a:Widget parameterWidget paramet...

  • 2228 Views
  • 2 replies
  • 11 kudos
Latest Reply
Natalie_NL
New Contributor II
  • 11 kudos

Hi, I build a dashboard with dashboard parameters, it works pretty easy!The advantage of dashboard parameters is that you do not have to set a default (it can be: all). This is convenient when you need to filter on values that change every time the q...

  • 11 kudos
1 More Replies
The_raj
by New Contributor
  • 4567 Views
  • 0 replies
  • 0 kudos

Error while reading file <file path>. [DEFAULT_FILE_NOT_FOUND]

Hi,I have a workflow created where there are 5 notebooks in it. One of the notebooks is failing with below error. I have tried refreshing the table. Still facing the same issue. When I try to run the notebook manually, it works fine. Can someone plea...

  • 4567 Views
  • 0 replies
  • 0 kudos
dvmentalmadess
by Valued Contributor
  • 1864 Views
  • 3 replies
  • 0 kudos

Ingestion Time Clustering on initial load

We are migrating our data into Databricks and I was looking at the recommendations for partitioning here: https://docs.databricks.com/tables/partitions.html. This recommends not specifying partitioning and allowing "Ingestion Time Partitioning" (ITP)...

  • 1864 Views
  • 3 replies
  • 0 kudos
Latest Reply
Anonymous
Not applicable
  • 0 kudos

Hi @dvmentalmadess  Hope all is well! Just wanted to check in if you were able to resolve your issue and would you be happy to share the solution or mark an answer as best? Else please let us know if you need more help.  We'd love to hear from you. T...

  • 0 kudos
2 More Replies
NWIEFInance
by New Contributor
  • 1020 Views
  • 0 replies
  • 0 kudos

Connect to EXCEL

> I have hard time connecting to Excel, any help connecting Data Bricks to EXCEL

  • 1020 Views
  • 0 replies
  • 0 kudos
RamozanbekS
by New Contributor III
  • 1929 Views
  • 1 replies
  • 0 kudos

Resolved! Databricks SQL Statement Execution API

I'm trying to follow the example provided here https://github.com/databricks-demos/dbsql-rest-api/blob/main/python/external_links.pyIt fails when it comes to downloading the data chunks. The statement status turns from SUCCEEDED to CLOSED right away ...

  • 1929 Views
  • 1 replies
  • 0 kudos
Latest Reply
RamozanbekS
New Contributor III
  • 0 kudos

It turns out that if the response is small and can fit 16mb limit, then status check will also provide single external link to download the data.So I need a condition here. Maybe even something like thisif len(chunks) == 1: external_url = respons...

  • 0 kudos
MadrasSenpai
by New Contributor II
  • 1786 Views
  • 2 replies
  • 2 kudos

How to install cmdstanpy in dbx cluster

I have built an HMC model using cmdstand. In my local machine, I have install cmdstan for the following approach. import cmdstanpy cmdstanpy.install_cmdstan()But in Databricks I need to reinstall it every time when I train a new model, from the noteb...

  • 1786 Views
  • 2 replies
  • 2 kudos
Latest Reply
Anonymous
Not applicable
  • 2 kudos

Hi @Rajamannar Aanjaram Krishnamoorthy​ Great to meet you, and thanks for your question! Let's see if your peers in the community have an answer to your question. Thanks.

  • 2 kudos
1 More Replies
Asterol
by New Contributor III
  • 1295 Views
  • 1 replies
  • 1 kudos

Creating a test schema - what is the best practice?

Hey, I've created a schema with few tables with historical data (prod), now I would like to have a Dev/testing environment with exactly the same data.What do you recommend? CTAS? Shallow clone? Deep clone? I wonder if shallow clone would be sufficien...

Data Engineering
Clone ctas
  • 1295 Views
  • 1 replies
  • 1 kudos
Latest Reply
Tharun-Kumar
Databricks Employee
  • 1 kudos

@Asterol If you would like to have the same data for your Dev/testing environment, I would recommend using Deep Clone. Deep clone copies the metadata and creates an independent copy of the table data. Shallow clone only copies the metadata and will h...

  • 1 kudos
NathanSundarara
by Contributor
  • 1600 Views
  • 0 replies
  • 0 kudos

Sample code to read json from service bus queue in Azure

Hi,I'm looking for sample notebook or code snippet to read messages from Azure Service bus queues. I looked for documentation couldn't find anything. Any help would be appreciated. First we are thinking of batch mode before we move on to Streaming. P...

Data Engineering
azure
deltalivetable
messagequeue
servicebus
Servicebus azure deltalivetables message queue
  • 1600 Views
  • 0 replies
  • 0 kudos
Navashakthi
by New Contributor
  • 2037 Views
  • 4 replies
  • 2 kudos

Resolved! Community Edition Sign-up Issue

Hi, I'm trying to signup community edition for learning purpose. The sign up page has issue in selecting country. The select dropdown doesn't work and continue option redirects to same page. Couldn't complete signup. Kindly help!

  • 2037 Views
  • 4 replies
  • 2 kudos
Latest Reply
amitdas2k6
New Contributor II
  • 2 kudos

for me it is alwas displaying below error but entered correct user name and passowrd,my user name : amit.das2k16@gmail.com Invalid email address or passwordNote: Emails/usernames are case-sensitive 

  • 2 kudos
3 More Replies

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.

If there isn’t a group near you, start one and help create a community that brings people together.

Request a New Group
Labels