cancel
Showing results for 
Search instead for 
Did you mean: 
Get Started Discussions
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and introductory topics. Connect with beginners and experts alike to kickstart your Databricks experience.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Miasu
by New Contributor II
  • 2832 Views
  • 1 replies
  • 0 kudos

FileAlreadyExistsException error while analyzing table in Notebook

Databricks experts, I'm new to Databricks, and encounter an issue with the ANALYZE TABLE command in the Notebook. I created two tables nyc_taxi and nyc_taxi2, from one csv file.When executing the following command in Notebook, analyze table nyc_taxi2...

  • 2832 Views
  • 1 replies
  • 0 kudos
Dlt
by New Contributor III
  • 10699 Views
  • 4 replies
  • 0 kudos

Running an exe file in databricks

Hello I have an executable file which i want to host and run from databricks.  is this possible in databricks using DBFS ?If NOT what are the other ways to it in databricks ? 

  • 10699 Views
  • 4 replies
  • 0 kudos
Latest Reply
BR_DatabricksAI
Contributor III
  • 0 kudos

Hello, I don't have much information on what kind of executables you would like to run in databricks however, I can think of two solutions : Solution 1: Deploy your code in azure container registry as an image and use the endpoint in data bricks. Sol...

  • 0 kudos
3 More Replies
Wycliff
by New Contributor II
  • 3288 Views
  • 1 replies
  • 0 kudos

JWT Encoding error while using Azure secret key

My secret value in Azure key vault is like below.private_key="""-----BEGIN RSA PRIVATE KEY-----********-----END RSA PRIVATE KEY-----"""Running this command in Databricks notebook - jwt.encode(claim_set,private_key,algorithm='RS256')While using the ab...

  • 3288 Views
  • 1 replies
  • 0 kudos
Latest Reply
Wycliff
New Contributor II
  • 0 kudos

Thanks much for your troubleshooting methods.Validated the secret scopes, accessing secrets. These looks fine.Key format - I feel problem is with the key format only. As of now I'm awaiting on Azure subscription access. But I printed the secret value...

  • 0 kudos
sanjay
by Valued Contributor II
  • 9851 Views
  • 3 replies
  • 1 kudos

stop autoloader with continuous trigger programatically

Hi,I am running autoloader with continuous trigger. How can I stop this trigger during some specific time, only if no data pending and current batch process is complete. How to check how many records pending in queue and current state.Regards,Sanjay

  • 9851 Views
  • 3 replies
  • 1 kudos
Latest Reply
RamonaMraz
New Contributor II
  • 1 kudos

Hello, I am new here, Can I ask a question?

  • 1 kudos
2 More Replies
Lucidity
by New Contributor II
  • 24447 Views
  • 2 replies
  • 1 kudos

Resolved! Deny assignment modification to allow attach/detach of disks in azure databricks

Our application does storage autoscaling on Azure. We would like to deploy our solution with Azure databricks. But even though the service principal associated with our application has the necessary roles and permissions to attach/detach a disk from ...

  • 24447 Views
  • 2 replies
  • 1 kudos
Latest Reply
Lucidity
New Contributor II
  • 1 kudos

Thank you for your replyIs there any way databricks provides to bypass the deny assignment for specific apps? I noticed in the deny assignment unity-catalog-access-connector has been provided exlusion under the excludePrincipals section. is there a w...

  • 1 kudos
1 More Replies
amama
by New Contributor II
  • 1447 Views
  • 1 replies
  • 0 kudos

Create persistent Scala UDF

I've created a UDF in Scala (using Databricks notebook) which checks if the value of a field in a table is greater than 0, then sum those values. UDF is working as expected. The problem is, that this UDF will be used by multiple SQLs across different...

  • 1447 Views
  • 1 replies
  • 0 kudos
Latest Reply
feiyun0112
Honored Contributor
  • 0 kudos

In the child notebook:def my_udf(x): return x + 1   spark.udf.register("my_udf", my_udf)   dbutils.notebook.exit(my_udf)In the master notebook:child_udf = dbutils.notebook.run("PathToChildnotebook", timeout_seconds=600) spark.udf.register("my_udf",...

  • 0 kudos
AadityaBhatt
by New Contributor II
  • 3637 Views
  • 3 replies
  • 1 kudos

[UNBOUND_SQL_PARAMETER] When running Placeholder query

I am using the databricks-sql-go library version 1.5.2. I am trying to run a query with placeholders of type '?'The query looks like params, args := databricksParams(values)sql := fmt.Sprintf(`SELECT COUNT(*) FROM %s.%s WHERE %s IN (%s)`, schema, tab...

  • 3637 Views
  • 3 replies
  • 1 kudos
Latest Reply
SergeRielau
Databricks Employee
  • 1 kudos

Can you print out an example after the Sprintf substitutions?It seems you generated a query with a named parameter: ":_58" But args (which should be a Map) does not have a key named "_58".

  • 1 kudos
2 More Replies
coltonflowers
by New Contributor III
  • 2212 Views
  • 1 replies
  • 0 kudos

DLT: Only STREAMING tables can have multiple queries.

I am trying to to do a one-time back-fill on a DLT table following the example here: dlt.table() def test(): # providing a starting version return (spark.readStream.format("delta") .option("readChangeFeed", "true") .option("...

  • 2212 Views
  • 1 replies
  • 0 kudos
Latest Reply
coltonflowers
New Contributor III
  • 0 kudos

I should also add that when I drop the `backfill` function, validation happens successfully and we get the following pipeline DAG:

  • 0 kudos
Sujitha
by Databricks Employee
  • 13165 Views
  • 1 replies
  • 1 kudos

Introducing AI Model Sharing with Databricks!

Today, we're excited to announce that AI model sharing is available in both Databricks Delta Sharing and on the Databricks Marketplace. With Delta Sharing you can now easily share and serve AI models securely within your organization or externally ac...

Screenshot 2024-02-06 at 7.01.48 PM.png
  • 13165 Views
  • 1 replies
  • 1 kudos
Latest Reply
johnsonit
New Contributor II
  • 1 kudos

I'm eager to dive in and leverage these new features to elevate my AI game with Databricks.This is Johnson from KBS Technologies.Thanks for your update.

  • 1 kudos
Sweetness
by New Contributor II
  • 3582 Views
  • 3 replies
  • 0 kudos

Cannot create a repo because the parent path does not exist

I tried following this docWork With Large Monorepos With Sparse Checkout Support in Databricks Repos | Databricks BlogWhen I hook it up to my repos using Azure DevOps Services and check mark Sparse checkout mode, I pass in a subdirectory in my Cone p...

  • 3582 Views
  • 3 replies
  • 0 kudos
Latest Reply
Sweetness
New Contributor II
  • 0 kudos

This is Azure Databricks

  • 0 kudos
2 More Replies
Frantz
by New Contributor III
  • 10845 Views
  • 2 replies
  • 0 kudos

Resolved! Show Existing Header From CSV I External Table

Hello, is there a way to load csv data into an external table without the _c0, _c1 columns showing?

  • 10845 Views
  • 2 replies
  • 0 kudos
Latest Reply
Frantz
New Contributor III
  • 0 kudos

My question was answered in a separate thread here.

  • 0 kudos
1 More Replies
Frantz
by New Contributor III
  • 3795 Views
  • 3 replies
  • 0 kudos

Resolved! Unable to load csv data with correct header values in External tables

Hello, is there a way to load "CSV" data into an external table without the _c0, _c1 columns showing?I've tried using the options within the sql statement that does not appear to work.Which results in this table 

Frantz_0-1707258246022.png Frantz_1-1707258264972.png
Get Started Discussions
External Tables
Unity Catalog
  • 3795 Views
  • 3 replies
  • 0 kudos
Latest Reply
feiyun0112
Honored Contributor
  • 0 kudos

you need set "USING data_source"https://community.databricks.com/t5/data-engineering/create-external-table-using-multiple-paths-locations/td-p/44042 

  • 0 kudos
2 More Replies
Lambda
by New Contributor
  • 1637 Views
  • 1 replies
  • 0 kudos

What is the REST API payload for "not sending alert when alert is back to normal"?

Greeting,I am using REST API to create & schedule sql alert by default these alerts will send notifications when they are back to normal which I don't want, In the UI I have the option to uncheck the box (shown in the picture) but I can't find any do...

Lambda_0-1702306187563.png
  • 1637 Views
  • 1 replies
  • 0 kudos
Latest Reply
Yeshwanth
Databricks Employee
  • 0 kudos

@Lambda Did you refer this documentation: https://docs.databricks.com/api/workspace/alerts/create

  • 0 kudos
Phani1
by Valued Contributor II
  • 1833 Views
  • 1 replies
  • 0 kudos

Spark VCore for Databrick vCPU

Hi Team,What is the equivalent Spark VCore for Databrick vCPU? , like for below example for DS3 v2 , vCPU=4 and RAM = 14.00 GiB, would like to know equivalent Spark VCore for DS3 v2 as in  Azure Databricks Pricing | Microsoft AzureRegards,Phanindra

  • 1833 Views
  • 1 replies
  • 0 kudos
Latest Reply
Yeshwanth
Databricks Employee
  • 0 kudos

@Phani1 In Databricks, the vCPU count is equivalent to the number of Spark vCores. This means that if the DS3 v2 instance has 4 vCPU, it would also have 4 Spark vCores. Please note that the Spark VCore count is based on the vCPU count of the underlyi...

  • 0 kudos
AtomicBoy99
by New Contributor
  • 1403 Views
  • 1 replies
  • 0 kudos

Can't enter 'edit mode' using shortcut

Hi,With the recently added AI Assistant to databricks notebooks, I'm having issues entering the 'edit mode' of a notebook cell. Previously, I could simply press the 'Enter' key in order to do this but this no longer works.Is anyone else having the sa...

  • 1403 Views
  • 1 replies
  • 0 kudos
Latest Reply
Yeshwanth
Databricks Employee
  • 0 kudos

@AtomicBoy99, it seems like the fix is in place. Can you confirm if it is working well now?

  • 0 kudos

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels