cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Hubert-Dudek
by Esteemed Contributor III
  • 1045 Views
  • 1 replies
  • 10 kudos

databricks SQL now supports getting the SQLSTATE code of a query to identify errors. You can use the e.getSqlState() method in a try/catch block to ge...

databricks SQL now supports getting the SQLSTATE code of a query to identify errors. You can use the e.getSqlState() method in a try/catch block to get the five-character code that indicates the success or failure of an SQL command. I am still thinki...

Untitled
  • 1045 Views
  • 1 replies
  • 10 kudos
Latest Reply
jose_gonzalez
Moderator
  • 10 kudos

Thank you for sharing @Hubert Dudek​ 

  • 10 kudos
Ajay-Pandey
by Esteemed Contributor III
  • 1875 Views
  • 3 replies
  • 5 kudos

Support of running multiple cells at a time in databricks notebook Hi all,Now databricks notebook supports parallel run of commands in a single notebo...

Support of running multiple cells at a time in databricks notebookHi all,Now databricks notebook supports parallel run of commands in a single notebook that will help run ad hoc queries simultaneously without creating a separate notebook.Once you run...

image.png image
  • 1875 Views
  • 3 replies
  • 5 kudos
Latest Reply
Anonymous
Not applicable
  • 5 kudos

Hi @Ajay Pandey​ Hope everything is going great.Just wanted to check in if you were able to resolve your issue. If yes, would you be happy to mark an answer as best so that other members can find the solution more quickly? If not, please tell us so w...

  • 5 kudos
2 More Replies
Data_Engineer3
by Contributor II
  • 1569 Views
  • 2 replies
  • 0 kudos

Default maximum spark streaming chunk size in delta files in each batch?

working with delta files spark structure streaming , what is the maximum default chunk size in each batch?How do identify this type of spark configuration in databricks?#[Databricks SQL]​ #[Spark streaming]​ #[Spark structured streaming]​ #Spark​ 

  • 1569 Views
  • 2 replies
  • 0 kudos
Latest Reply
NandiniN
Valued Contributor II
  • 0 kudos

Hello @KARTHICK N​ ,The default value for spark.sql.files.maxPartitionBytes is 128 MB. These defaults are in the Apache Spark documentation https://spark.apache.org/docs/latest/sql-performance-tuning.html (unless there might be some overrides).To che...

  • 0 kudos
1 More Replies
Kaniz
by Community Manager
  • 1082 Views
  • 3 replies
  • 4 kudos

Databricks SQL is curious !!!!

Hello Databricks Community!We're conducting a quick poll to gather insights on your experience with SQL in Databricks.Your input will help us tailor our content and discussions to serve the community's needs better. Please take a moment to answer the...

  • 1082 Views
  • 3 replies
  • 4 kudos
Latest Reply
rendorHaevyn
New Contributor III
  • 4 kudos

If you're asking about the Databricks SQL service: (8) (8.1) Repos integration of SQL Queries (requested feature)(8.2) Drag-and-drop GUI organisation of SQL queries and dashboards by features / tags / user, and generally a significantly improved UI (...

  • 4 kudos
2 More Replies
Anonymous
by Not applicable
  • 464 Views
  • 1 replies
  • 2 kudos

Join our Community Social Group and Never Miss a Beat! Are you looking to connect with like-minded individuals and stay on top of the latest news and ...

Join our Community Social Group and Never Miss a Beat!Are you looking to connect with like-minded individuals and stay on top of the latest news and events in your community? Look no further than our special group on Community called the "Community S...

  • 464 Views
  • 1 replies
  • 2 kudos
Latest Reply
Ajay-Pandey
Esteemed Contributor III
  • 2 kudos

@Rishabh Pandey​ 

  • 2 kudos
Anonymous
by Not applicable
  • 938 Views
  • 4 replies
  • 4 kudos

How to create a new group in the Databricks community? Dear esteemed community users, It is with great pleasure that we inform you of an important upd...

How to create a new group in the Databricks community?Dear esteemed community users,It is with great pleasure that we inform you of an important update regarding the creation of Groups on Community. As part of our continuous efforts to enhance your e...

  • 938 Views
  • 4 replies
  • 4 kudos
Latest Reply
Anonymous
Not applicable
  • 4 kudos

HI @Hubert Dudek​  and @Ratna Chaitanya Raju Bandaru​ : Thanks for pointing this out. This is going to be a design decision which we will take after looking into the ask carefully. Thanks for getting the conversation going. This really helps us.

  • 4 kudos
3 More Replies
oleole
by Contributor
  • 4912 Views
  • 3 replies
  • 2 kudos

Resolved! Using "FOR XML PATH" in Spark SQL in sql syntax

I'm using spark version 3.2.1 on databricks (DBR 10.4 LTS), and I'm trying to convert sql server sql query to a new sql query that runs on a spark cluster using spark sql in sql syntax. However, spark sql does not seem to support XML PATH as a functi...

input output
  • 4912 Views
  • 3 replies
  • 2 kudos
Latest Reply
oleole
Contributor
  • 2 kudos

Posting the solution that I ended up using:%sql DROP TABLE if exists UserCountry; CREATE TABLE if not exists UserCountry ( UserID INT, Country VARCHAR(5000) ); INSERT INTO UserCountry SELECT L.UserID AS UserID, CONCAT_WS(',', co...

  • 2 kudos
2 More Replies
JamieN
by New Contributor III
  • 11779 Views
  • 5 replies
  • 16 kudos

Resolved! Querying columns with special characters

I'm trying to do a select on a column with special characters in Databricks sql.I've tried the following%sql select ex$col from database.table limit 10;   %sql select `ex$col ` from database.table limit 10;They both don't work and will return "ex do...

  • 11779 Views
  • 5 replies
  • 16 kudos
Latest Reply
Anonymous
Not applicable
  • 16 kudos

Hi @Jamie Nathan​ Hope all is well! Just wanted to check in if you were able to resolve your issue and would you be happy to share the solution or mark an answer as best? Else please let us know if you need more help. We'd love to hear from you.Thank...

  • 16 kudos
4 More Replies
wyzer
by Contributor II
  • 28946 Views
  • 15 replies
  • 7 kudos

Resolved! What's the equivalent of "DECLARE..." in Databricks SQL ?

Hello everyone,I'm new in Databricks SQL, and I'm comming from SQL Server.I would like to know what's the equivalent of :DECLARE @P_Name varchar(50) = 'BackOffice'It's for use it like this : CREATE DATABASE @P_NameThanks.

  • 28946 Views
  • 15 replies
  • 7 kudos
Latest Reply
Amit_352107
New Contributor III
  • 7 kudos

Hi @Salah K.​ you can go through this code block%python P_Name = 'BackOffice'spark.sql(f""" create database {P_name} """)

  • 7 kudos
14 More Replies
sanq
by New Contributor II
  • 1975 Views
  • 3 replies
  • 7 kudos

what formatter is used to format SQL cell in databricks

Databricks launched formatter Black which formats python cells, I can also see SQL cell getting formatted, but not sure which formatter is being used for SQL cell formatting. No clarity given on docs.

  • 1975 Views
  • 3 replies
  • 7 kudos
Latest Reply
Anonymous
Not applicable
  • 7 kudos

Hi @SANJAY qq​ Hope all is well! Just wanted to check in if you were able to resolve your issue and would you be happy to share the solution or mark an answer as best? Else please let us know if you need more help. We'd love to hear from you.Thanks!

  • 7 kudos
2 More Replies
JustinDM
by New Contributor II
  • 705 Views
  • 2 replies
  • 2 kudos

I have been getting a False Alert on Data bricks SQL.

When ever a legitimate alert is triggered, I get a False Alert with 0.00 triggered at 12.00AM the next day. I tried Altering the Query but its still the same. Not posting examples as data is not shareable but I can give an example. if the alert is se...

  • 705 Views
  • 2 replies
  • 2 kudos
Latest Reply
Anonymous
Not applicable
  • 2 kudos

Hi @justin moorthy​ Hope all is well! Just wanted to check in if you were able to resolve your issue and would you be happy to share the solution or mark an answer as best? Else please let us know if you need more help. We'd love to hear from you.Tha...

  • 2 kudos
1 More Replies
ramankr48
by Contributor II
  • 11721 Views
  • 5 replies
  • 8 kudos

Resolved! How to get all the tables name with a specific column or columns in a database?

let's say there is a database db in which 700 tables are there, and we need to find all the tables name in which column "project_id" is present.just an example for ubderstanding the questions.

  • 11721 Views
  • 5 replies
  • 8 kudos
Latest Reply
Anonymous
Not applicable
  • 8 kudos

databaseName = "db" desiredColumn = "project_id" database = spark.sql(f"show tables in {databaseName} ").collect() tablenames = [] for row in database: cols = spark.table(row.tableName).columns if desiredColumn in cols: tablenames.append(row....

  • 8 kudos
4 More Replies
KVNARK
by Honored Contributor II
  • 14846 Views
  • 12 replies
  • 10 kudos

Resolved! Fetching top 1 L rows in Databricks SQL

how to fetch the top 1L rows from a long SQL query as the query is returning 5gb data, I want to check only the first 1L ROWS.

  • 14846 Views
  • 12 replies
  • 10 kudos
Latest Reply
Sangram
New Contributor III
  • 10 kudos

use the below commandsdf.show(100000)

  • 10 kudos
11 More Replies
Tewks
by New Contributor
  • 1143 Views
  • 2 replies
  • 5 kudos

Resolved! Databricks SQL External Connections

Lakehouse architectures seem enticing, especially from the standpoint of querying the data lake directly as it sits (as opposed to first migrating the data to an external data warehouse). While documentation and support seems pretty clear regarding ...

  • 1143 Views
  • 2 replies
  • 5 kudos
Latest Reply
Aviral-Bhardwaj
Esteemed Contributor III
  • 5 kudos

these are really awesome details

  • 5 kudos
1 More Replies
Sujitha
by Community Manager
  • 412 Views
  • 1 replies
  • 1 kudos

Weekly Release Notes RecapHere’s a quick recap of the latest release notes updates from the past one week. Databricks platform release notesFebruary 2...

Weekly Release Notes RecapHere’s a quick recap of the latest release notes updates from the past one week.Databricks platform release notesFebruary 21 - 28, 2023Ray on Databricks (Public Preview)With Databricks Runtime 12.0 and above, you can create ...

  • 412 Views
  • 1 replies
  • 1 kudos
Latest Reply
jose_gonzalez
Moderator
  • 1 kudos

Thank you for sharing!!!

  • 1 kudos
Labels