cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

dkxxx-rc
by Contributor
  • 10594 Views
  • 3 replies
  • 2 kudos

Resolved! CREATE TEMP TABLE

The Databricks assistant tells me (sometimes) that `CREATE TEMP TABLE` is a valid SQL operation.  And other sources (e.g., https://www.freecodecamp.org/news/sql-temp-table-how-to-create-a-temporary-sql-table/) say the same.But in actual practice, thi...

  • 10594 Views
  • 3 replies
  • 2 kudos
Latest Reply
dkxxx-rc
Contributor
  • 2 kudos

In addition to accepting KaranamS's answer, I will note a longer and useful discussion, with caveats, at https://community.databricks.com/t5/data-engineering/how-do-temp-views-actually-work/m-p/20137/highlight/true#M13584.

  • 2 kudos
2 More Replies
Upendra_Dwivedi
by Contributor
  • 2093 Views
  • 5 replies
  • 0 kudos

Remote SQL Server Instance Connection using JDBC

Hi All,I am connecting Remote SQL Server Instance using JDBC Driver, I have enabled TCP/IP and Setup the Firewall Rule. When i am querying the instance i am getting this error:(com.microsoft.sqlserver.jdbc.SQLServerException) The TCP/IP connection to...

  • 2093 Views
  • 5 replies
  • 0 kudos
Latest Reply
turagittech
Contributor
  • 0 kudos

If you want to access a local SQL server, you'll need a Private Link to access the server. If it's on your own local machine, that's likely not possible. Creating a VPN to your machine is a unique problem, and you would be better off using a VM or a ...

  • 0 kudos
4 More Replies
sanq
by New Contributor II
  • 7518 Views
  • 3 replies
  • 7 kudos

what formatter is used to format SQL cell in databricks

Databricks launched formatter Black which formats python cells, I can also see SQL cell getting formatted, but not sure which formatter is being used for SQL cell formatting. No clarity given on docs.

  • 7518 Views
  • 3 replies
  • 7 kudos
Latest Reply
mitch_DE
New Contributor II
  • 7 kudos

The formatter is mentioned here: Develop code in Databricks notebooks - Azure Databricks | Microsoft LearnIt is this npm package: @gethue/sql-formatter - npm

  • 7 kudos
2 More Replies
BobCat62
by New Contributor III
  • 4043 Views
  • 8 replies
  • 3 kudos

Resolved! How to copy notebooks from local to the tarrget folder via asset bundles

Hi all,I am able to deploy Databricks assets to the target workspace. Jobs and workflows can also be created successfully.But I have aspecial requirement, that I copy the note books to the target folder on databricks workspace.Example:on Local I have...

  • 4043 Views
  • 8 replies
  • 3 kudos
Latest Reply
kmodelew
New Contributor III
  • 3 kudos

What are the permissions to this databricks directory? Can someone delete this directory or any file? On Shared workspace everyone can delete bundle files or bundle directory, even if in databricks.yml I provided permissions only to admins ('CAN MANA...

  • 3 kudos
7 More Replies
TomHauf
by New Contributor II
  • 862 Views
  • 1 replies
  • 1 kudos

Sending my weather data to a clients cloud storage

Hi, One of our clients is asking to switch from our API feed to have weather data delivered automatically to their Cloud Storage.  What steps do I need to take from my end?  Do I need to join Databricks to do so? Thanks. Tom

  • 862 Views
  • 1 replies
  • 1 kudos
Latest Reply
XP
Databricks Employee
  • 1 kudos

Hey @TomHauf, while it may not be essential in your case, you should at least consider using Databricks to facilitate loading data into your customers cloud storage. Databricks gives you a few options to make sharing with third parties simple as per ...

  • 1 kudos
Long_Tran
by New Contributor
  • 4387 Views
  • 2 replies
  • 0 kudos

Can job 'run_as' be assigned to users/principals who actually run it?

Can job 'run_as' be assigned to users/principals who actually run it? instead of always a fixed creator/user/pricipal?When a job is run, I would like to see in the job setting "run_as" the name of the actual user/principal who runs it.Currently, "run...

  • 4387 Views
  • 2 replies
  • 0 kudos
Latest Reply
701153
New Contributor II
  • 0 kudos

Yeah, the functionality is odd. You can't change the Run As user to anyone but yourself. But you can run it using the Run As setting previously used. This sort of makes sense if the workflow is created to be run as a service principal with specific p...

  • 0 kudos
1 More Replies
JJ_LVS1
by New Contributor III
  • 2347 Views
  • 1 replies
  • 0 kudos

CLOUD_PROVIDER_RESOURCE_STOCKOUT (Azure)

Hey All,Anyone ran into this 'out of stock' error on certain types of clusters?  We've spent months building on  Standard_D8ads_v5 (delta cache) and this morning a see of red because there are none available.  I can't even spin up a small interactive...

  • 2347 Views
  • 1 replies
  • 0 kudos
Latest Reply
Prabakar
Databricks Employee
  • 0 kudos

Hi JJ,  The CLOUD_PROVIDER_RESOURCE_STOCKOUT error code indicates that the cloud provider is out of physical capacity underneath virtual machines. The failure was caused by the cloud provider and I would recommend you reaching out to the respective c...

  • 0 kudos
valjas
by New Contributor III
  • 4197 Views
  • 3 replies
  • 0 kudos

Warehouse Name in System Tables

Hello.I am creating a table to monitor the usage of All-purpose Compute and SQL Warehouses. From the tables in 'system' catalog, I can get cluster_name and cluster_id. However only warehouse_id is available and not warehouse name. Is there a way to g...

  • 4197 Views
  • 3 replies
  • 0 kudos
Latest Reply
aranjan99
Contributor
  • 0 kudos

I have enabled the compute system schema, but i dont see this tabke. What am i missing?

  • 0 kudos
2 More Replies
cmathieu
by New Contributor III
  • 1761 Views
  • 4 replies
  • 0 kudos

DAB - All projects files deployed

I have an issue with DAB where all the project files, starting from root ., get deployed to the /files folder in the bundle. I would prefer being able to deploy certain util notebooks, but not all the files of the project. I'm able to not deploy any ...

  • 1761 Views
  • 4 replies
  • 0 kudos
Latest Reply
ashraf1395
Honored Contributor
  • 0 kudos

@cmathieu , It will support  deployment of whole directory and not others as well.

  • 0 kudos
3 More Replies
DylanStout
by Contributor
  • 1329 Views
  • 2 replies
  • 0 kudos

Resolved! Error while reading file from Cloud Storage

The code we are executing: df = spark.read.format("parquet").load("/mnt/g/drb/HN/") df.write.mode('overwrite').saveAsTable("bronze.HN")the error it throws:org.apache.spark.SparkException: Job aborted due to stage failure: Task 44 in stage 642.0 faile...

  • 1329 Views
  • 2 replies
  • 0 kudos
Latest Reply
DylanStout
Contributor
  • 0 kudos

spark.conf.set("spark.sql.parquet.enableVectorizedReader", "false")

  • 0 kudos
1 More Replies
Rajt1
by New Contributor
  • 858 Views
  • 1 replies
  • 0 kudos

Job , Task, Stage Creation

I am running below code -df = spark.read.json('xyz.json')df.countI want to understand the actual working of the spark. How many jobs & stages will be created. I want to understand the detailed & easier concept of how it works?

  • 858 Views
  • 1 replies
  • 0 kudos
Latest Reply
Advika
Community Manager
  • 0 kudos

Hello @Rajt1! When you execute df = spark.read.json('xyz.json’), Spark does not read the file immediately. Data is only read when an action like count() is triggered. Job: df.count() triggers one job because it's an action.Stage: Reading JSON and cou...

  • 0 kudos
khangnguyen164
by New Contributor II
  • 2068 Views
  • 3 replies
  • 0 kudos

Error "insert concurrent to Delta Lake" when 2 streaming merge data to same table at the same time

Hello everyone ,We currently have 2 streaming (Bronze job) created on 2 tasks in the same job, running the same compute job and both merge data into the same table (Silver table). If I create it like above, sometimes I get an error related to "insert...

  • 2068 Views
  • 3 replies
  • 0 kudos
Latest Reply
khangnguyen164
New Contributor II
  • 0 kudos

 Anyone else can help me this case

  • 0 kudos
2 More Replies
YOUKE
by New Contributor III
  • 2457 Views
  • 2 replies
  • 0 kudos

Resolved! Connecting to SQL on Databricks Using SQLAlchemy or pyodbc

On Databricks, when I try to connect to SQL using SQLAlchemy or pyodbc to run delete queries on a specific table, I get this error: (pyodbc.Error) ('01000', "[01000] [unixODBC][Driver Manager]Can't open lib 'ODBC Driver 17 for SQL Server' : file not ...

  • 2457 Views
  • 2 replies
  • 0 kudos
Latest Reply
YOUKE
New Contributor III
  • 0 kudos

I was able to solve the problem! the problem was because the driver was missing and so pyodbc or sqlAlchemy can't find it. So I used the native Java API and it is working.This is the example code:jdbcUsername = "username"jdbcPassword = "password"driv...

  • 0 kudos
1 More Replies
IGRACH
by New Contributor III
  • 855 Views
  • 1 replies
  • 1 kudos

Resolved! Unable to delete a table

When I try to delete a table, I'm getting this error:[ErrorClass=INVALID_STATE] TABLE catalog.schema.table_name cannot be deleted because it is being shared via Delta Sharing.I have checked on the internet about it, but could not find any info about ...

  • 855 Views
  • 1 replies
  • 1 kudos
Latest Reply
ashraf1395
Honored Contributor
  • 1 kudos

Hi @IGRACH ,You are facing this issue bcz I guess the table you want to delete is being shared by delta sharing. You can go to the shared object by following this dochttps://docs.databricks.com/aws/en/delta-sharing/create-share#update-sharesandThen, ...

  • 1 kudos
HoussemBL
by New Contributor III
  • 2300 Views
  • 3 replies
  • 1 kudos

External tables in DLT pipelines

Hello community,I have implemented a DLT pipeline.In the "Destination" setting of the pipeline I have specified a unity catalog with target schema of type external referring to an S3 destination.My DLT pipeline works well. Yet, I noticed that all str...

  • 2300 Views
  • 3 replies
  • 1 kudos
Latest Reply
Sushil_saini
New Contributor II
  • 1 kudos

This won't work.best approach is create dlt sink to write to delta external table. This pipeline should only be 1 step. Read table and append flow using data sink. It works fine. 

  • 1 kudos
2 More Replies
Labels