cancel
Showing results for 
Search instead for 
Did you mean: 
Community Platform Discussions
Connect with fellow community members to discuss general topics related to the Databricks platform, industry trends, and best practices. Share experiences, ask questions, and foster collaboration within the community.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Miasu
by New Contributor II
  • 2187 Views
  • 1 replies
  • 0 kudos

FileAlreadyExistsException error while analyzing table in Notebook

Databricks experts, I'm new to Databricks, and encounter an issue with the ANALYZE TABLE command in the Notebook. I created two tables nyc_taxi and nyc_taxi2, from one csv file.When executing the following command in Notebook, analyze table nyc_taxi2...

  • 2187 Views
  • 1 replies
  • 0 kudos
Akira
by New Contributor II
  • 2054 Views
  • 3 replies
  • 0 kudos

"PutWithBucketOwnerFullControl" privilege missing for storage configuration

Hi. I've been unable to create workspaces manually for a while now. The error I get is "MALFORMED_REQUEST: Failed storage configuration validation checks: List,Put,PutWithBucketOwnerFullControl,Delete".  The storage configuration is on a bucket that ...

putwithownercontrols_error.trimmed.png
  • 2054 Views
  • 3 replies
  • 0 kudos
Latest Reply
Akira
New Contributor II
  • 0 kudos

> Yes, it does look like the bucket permissions are not properly set up, but ...To avoid potential misunderstanding: I mean yes the error message does make it sound like the bucket permissions are wrong. I don't meant I found a problem with the ones ...

  • 0 kudos
2 More Replies
sanjay
by Valued Contributor II
  • 8751 Views
  • 3 replies
  • 1 kudos

stop autoloader with continuous trigger programatically

Hi,I am running autoloader with continuous trigger. How can I stop this trigger during some specific time, only if no data pending and current batch process is complete. How to check how many records pending in queue and current state.Regards,Sanjay

  • 8751 Views
  • 3 replies
  • 1 kudos
Latest Reply
RamonaMraz
New Contributor II
  • 1 kudos

Hello, I am new here, Can I ask a question?

  • 1 kudos
2 More Replies
amama
by New Contributor II
  • 1002 Views
  • 1 replies
  • 0 kudos

Create persistent Scala UDF

I've created a UDF in Scala (using Databricks notebook) which checks if the value of a field in a table is greater than 0, then sum those values. UDF is working as expected. The problem is, that this UDF will be used by multiple SQLs across different...

  • 1002 Views
  • 1 replies
  • 0 kudos
Latest Reply
feiyun0112
Honored Contributor
  • 0 kudos

In the child notebook:def my_udf(x): return x + 1   spark.udf.register("my_udf", my_udf)   dbutils.notebook.exit(my_udf)In the master notebook:child_udf = dbutils.notebook.run("PathToChildnotebook", timeout_seconds=600) spark.udf.register("my_udf",...

  • 0 kudos
coltonflowers
by New Contributor III
  • 1388 Views
  • 1 replies
  • 0 kudos

DLT: Only STREAMING tables can have multiple queries.

I am trying to to do a one-time back-fill on a DLT table following the example here: dlt.table() def test(): # providing a starting version return (spark.readStream.format("delta") .option("readChangeFeed", "true") .option("...

  • 1388 Views
  • 1 replies
  • 0 kudos
Latest Reply
coltonflowers
New Contributor III
  • 0 kudos

I should also add that when I drop the `backfill` function, validation happens successfully and we get the following pipeline DAG:

  • 0 kudos
Sujitha
by Databricks Employee
  • 12500 Views
  • 1 replies
  • 1 kudos

Introducing AI Model Sharing with Databricks!

Today, we're excited to announce that AI model sharing is available in both Databricks Delta Sharing and on the Databricks Marketplace. With Delta Sharing you can now easily share and serve AI models securely within your organization or externally ac...

Screenshot 2024-02-06 at 7.01.48 PM.png
  • 12500 Views
  • 1 replies
  • 1 kudos
Latest Reply
johnsonit
New Contributor II
  • 1 kudos

I'm eager to dive in and leverage these new features to elevate my AI game with Databricks.This is Johnson from KBS Technologies.Thanks for your update.

  • 1 kudos
Frantz
by New Contributor III
  • 9639 Views
  • 2 replies
  • 0 kudos

Resolved! Show Existing Header From CSV I External Table

Hello, is there a way to load csv data into an external table without the _c0, _c1 columns showing?

  • 9639 Views
  • 2 replies
  • 0 kudos
Latest Reply
Frantz
New Contributor III
  • 0 kudos

My question was answered in a separate thread here.

  • 0 kudos
1 More Replies
Frantz
by New Contributor III
  • 2453 Views
  • 3 replies
  • 0 kudos

Resolved! Unable to load csv data with correct header values in External tables

Hello, is there a way to load "CSV" data into an external table without the _c0, _c1 columns showing?I've tried using the options within the sql statement that does not appear to work.Which results in this table 

Frantz_0-1707258246022.png Frantz_1-1707258264972.png
Community Platform Discussions
External Tables
Unity Catalog
  • 2453 Views
  • 3 replies
  • 0 kudos
Latest Reply
feiyun0112
Honored Contributor
  • 0 kudos

you need set "USING data_source"https://community.databricks.com/t5/data-engineering/create-external-table-using-multiple-paths-locations/td-p/44042 

  • 0 kudos
2 More Replies
Kaizen
by Valued Contributor
  • 2755 Views
  • 2 replies
  • 0 kudos

Python Logging cant save log in DBFS

Hi! I am trying to integrate logging into my project. Got the library and logs to work but cant log the file into DBFS directly.Have any of you been able to save and append the log file directly to dbfs? From what i came across online the best way to...

Kaizen_0-1707174350136.png
  • 2755 Views
  • 2 replies
  • 0 kudos
Latest Reply
feiyun0112
Honored Contributor
  • 0 kudos

you can use  azure_storage_loggingSet Python Logging to Azure Blob, but Can not Find Log File there - Stack Overflow

  • 0 kudos
1 More Replies
Jasonh202222
by New Contributor II
  • 4166 Views
  • 2 replies
  • 1 kudos

Databricks notebook how to stop truncating numbers when export the query result to csv

I use Databricks notebook to query databases and export / download result to csv. I just accidentally close a pop-up window asking if need to truncate the numbers, I accidentally chose yes and don't ask again. Now all my long digit numbers are trunca...

  • 4166 Views
  • 2 replies
  • 1 kudos
Latest Reply
shan_chandra
Databricks Employee
  • 1 kudos

@Jasonh202222  - Kindly check the below navigation path user settings -> Account settings -> Display -> Download and Export.  Under Download and Export, Enable the checkbox -  "Prompt for formatting large numbers when downloading or exporting" and cl...

  • 1 kudos
1 More Replies
Kaizen
by Valued Contributor
  • 2839 Views
  • 2 replies
  • 0 kudos

Resolved! Using Python RPA Library on Databricks

Hi I didn't see any conversations regarding using python RPA package on Data bricks clusters. Is anyone doing this or have gotten it to successfully work on the clusters? I ran into the following errors:1) Initially I was getting the error below rega...

Kaizen_0-1706743633855.png Kaizen_1-1706743734836.png
  • 2839 Views
  • 2 replies
  • 0 kudos
Latest Reply
feiyun0112
Honored Contributor
  • 0 kudos

If you want to capture browser screenshot, you can use playwright%sh pip install playwright playwright install sudo apt-get update playwright install-deps  from playwright.async_api import async_playwright async with async_playwright() as p: ...

  • 0 kudos
1 More Replies
pshuk
by New Contributor III
  • 1664 Views
  • 1 replies
  • 0 kudos

how to create volume using databricks cli commands

I am new to using volumes on databricks. Is there a way to create volume using CLI commands.On the similar note, is there a way to create DBFS directories and subdirectories using single command.for example: I want to copy file here dbfs:/FileStore/T...

  • 1664 Views
  • 1 replies
  • 0 kudos
Latest Reply
Walter_C
Databricks Employee
  • 0 kudos

Creates a new volume. The user could create either an external volume or a managed volume. An external volume will be created in the specified external location, while a managed volume will be located in the default location which is specified bythe...

  • 0 kudos
al2co33
by New Contributor
  • 1444 Views
  • 1 replies
  • 0 kudos

Can I update a table comment using REST API?

https://docs.databricks.com/api/workspace/tablesIt seems I could only list/delete tables, is there a way to update a table's metadata like comment or detail fields by REST API?

  • 1444 Views
  • 1 replies
  • 0 kudos
Latest Reply
Ayushi_Suthar
Databricks Employee
  • 0 kudos

Hi @al2co33 , We don't currently provide any APIs for updating table comments, however you can utilize the SQL Statement Execution API to do it. You can use the following tutorial to ALTER TABLE/COLUMN COMMENT. https://learn.microsoft.com/en-us/azure...

  • 0 kudos

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.

If there isn’t a group near you, start one and help create a community that brings people together.

Request a New Group
Top Kudoed Authors