cancel
Showing results for 
Search instead for 
Did you mean: 
Community Platform Discussions
Connect with fellow community members to discuss general topics related to the Databricks platform, industry trends, and best practices. Share experiences, ask questions, and foster collaboration within the community.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

JasonAckman
by New Contributor
  • 7592 Views
  • 0 replies
  • 0 kudos

Data Engineer – Databricks - Remote

Data Engineer – Databricks - RemoteApply Here: Job Application for Data Engineer – Databricks at Jenzabar (greenhouse.io)Jenzabar Website: Higher Education Software Solutions - JenzabarFor over four decades, the higher education experts at Jenzabar h...

  • 7592 Views
  • 0 replies
  • 0 kudos
Sujitha
by Databricks Employee
  • 10308 Views
  • 0 replies
  • 1 kudos

Calling all innovators and visionaries! The 2024 Data Team Awards are open for nominations

Each year, we celebrate the amazing customers that rely on Databricks to innovate and transform their organizations — and the world — with the power of data and AI. The nomination form is now open to submit nominations. Nominations will close on Marc...

Screenshot 2024-02-12 at 10.52.07 AM.png
  • 10308 Views
  • 0 replies
  • 1 kudos
RobsonNLPT
by Contributor III
  • 3023 Views
  • 4 replies
  • 0 kudos

Databricks XML - Bypassing rootTag and rowTag

I see the current conversion of dataframe to xml need to be improved.My dataframe schema is a perfect nested schema based on structs but when I create a xml I have the follow issues:1) I can't add elements to root2) rootTag and rowTag are requiredIn ...

  • 3023 Views
  • 4 replies
  • 0 kudos
Latest Reply
sandip_a
Databricks Employee
  • 0 kudos

Here is one of the ways to use the struct field name as rowTag:     import org.apache.spark.sql.types._ val schema = new StructType().add("Record", new StructType().add("age", IntegerType).add("name", StringType)) val data = Seq(Row(Row(18, "John ...

  • 0 kudos
3 More Replies
Israel_H
by New Contributor III
  • 2403 Views
  • 3 replies
  • 1 kudos

The risks of code execution by default on widget change

Taking from my experience, the default action of widgets triggering code execution upon value change poses risks that outweigh the convenience in certain scenarios. While this feature may seem advantageous in some cases, it can lead to unintended con...

  • 2403 Views
  • 3 replies
  • 1 kudos
Latest Reply
Kayla
Valued Contributor II
  • 1 kudos

I definitely have to agree with the original point- if you have a notebook that you import, and you touch any widget value you're running code, most likely accidentally. I'd love to see a workspace or user type option where you can change the default...

  • 1 kudos
2 More Replies
RobsonNLPT
by Contributor III
  • 1438 Views
  • 2 replies
  • 1 kudos

databricks spark XML Writer

Hi.I'm trying to generate XML as output base on my nested dataframe. Everything is ok except by I don't know how to add elements to rootTag.I can add elements from rowtag but not in rootTag. Same problems to add attributes to root <books  version = "...

  • 1438 Views
  • 2 replies
  • 1 kudos
Latest Reply
Ayushi_Suthar
Databricks Employee
  • 1 kudos

Hi @RobsonNLPT ,Thanks for bringing up your concerns, always happy to help  Can you please refer to the below document to read and write the XML files? https://docs.databricks.com/en/query/formats/xml.html Please let me know if this helps and leave a...

  • 1 kudos
1 More Replies
Miasu
by New Contributor II
  • 2287 Views
  • 1 replies
  • 0 kudos

FileAlreadyExistsException error while analyzing table in Notebook

Databricks experts, I'm new to Databricks, and encounter an issue with the ANALYZE TABLE command in the Notebook. I created two tables nyc_taxi and nyc_taxi2, from one csv file.When executing the following command in Notebook, analyze table nyc_taxi2...

  • 2287 Views
  • 1 replies
  • 0 kudos
Akira
by New Contributor II
  • 2198 Views
  • 3 replies
  • 0 kudos

"PutWithBucketOwnerFullControl" privilege missing for storage configuration

Hi. I've been unable to create workspaces manually for a while now. The error I get is "MALFORMED_REQUEST: Failed storage configuration validation checks: List,Put,PutWithBucketOwnerFullControl,Delete".  The storage configuration is on a bucket that ...

putwithownercontrols_error.trimmed.png
  • 2198 Views
  • 3 replies
  • 0 kudos
Latest Reply
Akira
New Contributor II
  • 0 kudos

> Yes, it does look like the bucket permissions are not properly set up, but ...To avoid potential misunderstanding: I mean yes the error message does make it sound like the bucket permissions are wrong. I don't meant I found a problem with the ones ...

  • 0 kudos
2 More Replies
sanjay
by Valued Contributor II
  • 8885 Views
  • 3 replies
  • 1 kudos

stop autoloader with continuous trigger programatically

Hi,I am running autoloader with continuous trigger. How can I stop this trigger during some specific time, only if no data pending and current batch process is complete. How to check how many records pending in queue and current state.Regards,Sanjay

  • 8885 Views
  • 3 replies
  • 1 kudos
Latest Reply
RamonaMraz
New Contributor II
  • 1 kudos

Hello, I am new here, Can I ask a question?

  • 1 kudos
2 More Replies
amama
by New Contributor II
  • 1044 Views
  • 1 replies
  • 0 kudos

Create persistent Scala UDF

I've created a UDF in Scala (using Databricks notebook) which checks if the value of a field in a table is greater than 0, then sum those values. UDF is working as expected. The problem is, that this UDF will be used by multiple SQLs across different...

  • 1044 Views
  • 1 replies
  • 0 kudos
Latest Reply
feiyun0112
Honored Contributor
  • 0 kudos

In the child notebook:def my_udf(x): return x + 1   spark.udf.register("my_udf", my_udf)   dbutils.notebook.exit(my_udf)In the master notebook:child_udf = dbutils.notebook.run("PathToChildnotebook", timeout_seconds=600) spark.udf.register("my_udf",...

  • 0 kudos
coltonflowers
by New Contributor III
  • 1483 Views
  • 1 replies
  • 0 kudos

DLT: Only STREAMING tables can have multiple queries.

I am trying to to do a one-time back-fill on a DLT table following the example here: dlt.table() def test(): # providing a starting version return (spark.readStream.format("delta") .option("readChangeFeed", "true") .option("...

  • 1483 Views
  • 1 replies
  • 0 kudos
Latest Reply
coltonflowers
New Contributor III
  • 0 kudos

I should also add that when I drop the `backfill` function, validation happens successfully and we get the following pipeline DAG:

  • 0 kudos
Sujitha
by Databricks Employee
  • 12570 Views
  • 1 replies
  • 1 kudos

Introducing AI Model Sharing with Databricks!

Today, we're excited to announce that AI model sharing is available in both Databricks Delta Sharing and on the Databricks Marketplace. With Delta Sharing you can now easily share and serve AI models securely within your organization or externally ac...

Screenshot 2024-02-06 at 7.01.48 PM.png
  • 12570 Views
  • 1 replies
  • 1 kudos
Latest Reply
johnsonit
New Contributor II
  • 1 kudos

I'm eager to dive in and leverage these new features to elevate my AI game with Databricks.This is Johnson from KBS Technologies.Thanks for your update.

  • 1 kudos
Frantz
by New Contributor III
  • 9777 Views
  • 2 replies
  • 0 kudos

Resolved! Show Existing Header From CSV I External Table

Hello, is there a way to load csv data into an external table without the _c0, _c1 columns showing?

  • 9777 Views
  • 2 replies
  • 0 kudos
Latest Reply
Frantz
New Contributor III
  • 0 kudos

My question was answered in a separate thread here.

  • 0 kudos
1 More Replies
Frantz
by New Contributor III
  • 2631 Views
  • 3 replies
  • 0 kudos

Resolved! Unable to load csv data with correct header values in External tables

Hello, is there a way to load "CSV" data into an external table without the _c0, _c1 columns showing?I've tried using the options within the sql statement that does not appear to work.Which results in this table 

Frantz_0-1707258246022.png Frantz_1-1707258264972.png
Community Platform Discussions
External Tables
Unity Catalog
  • 2631 Views
  • 3 replies
  • 0 kudos
Latest Reply
feiyun0112
Honored Contributor
  • 0 kudos

you need set "USING data_source"https://community.databricks.com/t5/data-engineering/create-external-table-using-multiple-paths-locations/td-p/44042 

  • 0 kudos
2 More Replies
Kaizen
by Valued Contributor
  • 3034 Views
  • 2 replies
  • 0 kudos

Python Logging cant save log in DBFS

Hi! I am trying to integrate logging into my project. Got the library and logs to work but cant log the file into DBFS directly.Have any of you been able to save and append the log file directly to dbfs? From what i came across online the best way to...

Kaizen_0-1707174350136.png
  • 3034 Views
  • 2 replies
  • 0 kudos
Latest Reply
feiyun0112
Honored Contributor
  • 0 kudos

you can use  azure_storage_loggingSet Python Logging to Azure Blob, but Can not Find Log File there - Stack Overflow

  • 0 kudos
1 More Replies