cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

techg
by New Contributor II
  • 303 Views
  • 4 replies
  • 1 kudos

Missing selection for Parameter error

Hi All,I have created three parameters in an SQL query in Databricks. If no value is entered for a parameter, I would like the query to retrieve all values for that particular column. Currently, I'm getting an error message: "Missing selection for Pa...

  • 303 Views
  • 4 replies
  • 1 kudos
Latest Reply
techg
New Contributor II
  • 1 kudos

I'm creating this query with parameters in SQL Editor in Databricks and added it to the SQL Dashboard.Do we need to create Widget while creating parameters in SQL Editor? When i tried creating widget in SQL editor, Im getting syntax error near Widget...

  • 1 kudos
3 More Replies
Gusman
by New Contributor II
  • 246 Views
  • 2 replies
  • 1 kudos

Resolved! Natural language queries through REST API?

Natural language queries provided by Genie are really powerful and a compeling tool.Is there any way to execute these natural language queries through the REST API to integrate them into in-house applications?

  • 246 Views
  • 2 replies
  • 1 kudos
Latest Reply
stacey45
New Contributor II
  • 1 kudos

@Gusman wrote:Natural language queries provided by Genie are really powerful and a compeling tool.Is there any way to execute these natural language queries through the REST API to integrate them into in-house applications?While there's no direct RES...

  • 1 kudos
1 More Replies
Clara
by New Contributor
  • 135 Views
  • 1 replies
  • 1 kudos

Retrieve data older than the one year window : system.access.table_lineage

Hello,I am currently using table_lineage from system.access.table_lineage. It is a great feature but I am experiencing missing data. After some search I have seen that "Because lineage is computed on a one-year rolling window, lineage collected more ...

  • 135 Views
  • 1 replies
  • 1 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 1 kudos

Hi @Clara ,I don't think so. But you can build history such history tables by yourself. Desing ETL process that will extract data from system tables and store them in your own data tables. 

  • 1 kudos
sboxi
by New Contributor II
  • 214 Views
  • 2 replies
  • 1 kudos

Can we create Materialized view n exsting view and table?

Dear All,Is it possible to create Materialized view on view and table (Joining view and table)?I suspect it is not possible. Please suggest.Also please provide best way to schedule the refresh of Materialized view. Regards,Surya 

  • 214 Views
  • 2 replies
  • 1 kudos
Latest Reply
sboxi
New Contributor II
  • 1 kudos

Thanks @Alberto_Umana . I will try that.

  • 1 kudos
1 More Replies
TimW
by New Contributor
  • 3898 Views
  • 4 replies
  • 1 kudos

Resolved! Help - Can't create table from tutorial. Is my setup wrong?

Trying out databricks for the first time and followed the Get Started steps. I managed to successfully create a cluster and ran the simple sql tutorial to query data from a notebook. However, got the following error:Query:DROP TABLE IF EXISTS diamond...

  • 3898 Views
  • 4 replies
  • 1 kudos
Latest Reply
patwilliams
New Contributor III
  • 1 kudos

It seems as though you're doing great with your Databricks arrangement, however this sort of mistake could be connected with a couple of expected issues. In light of the subtleties you've shared, here are a few things you should check:Group Setup: Gu...

  • 1 kudos
3 More Replies
ashraf1395
by Valued Contributor
  • 341 Views
  • 5 replies
  • 1 kudos

Empty Streaming tables in dlt

I want to create empty streaming tables in dlt with only schema specified. Is it possible ?I want to do it in dlt python.

  • 341 Views
  • 5 replies
  • 1 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 1 kudos

Hi @ashraf1395, The term "rate" refers to a special source in Apache Spark's Structured Streaming that generates data at a specified rate. This source is primarily used for testing and benchmarking purposes. When you use spark.readStream.format("rate...

  • 1 kudos
4 More Replies
ashraf1395
by Valued Contributor
  • 319 Views
  • 2 replies
  • 2 kudos

Resolved! applying column tags

Can anyone tell me the correct syntax for applying a column tag to a specific tableThese are what I tried   ALTER TABLE accounts_and_customer.bronze.BB1123_loans ALTER/CHANGE COLUMN loan_number SET TAGS ('classification' = 'confidential')<p>I got thi...

  • 319 Views
  • 2 replies
  • 2 kudos
Latest Reply
ashraf1395
Valued Contributor
  • 2 kudos

Hi there @TakuyaOmi ,I agree. The syntax was correct. I was facing some completely different problems with schemas and I solved it. Thanks though. Or I would have spent hours banging my head to find the reason for error.

  • 2 kudos
1 More Replies
JothyGanesan
by New Contributor II
  • 231 Views
  • 1 replies
  • 1 kudos

CDF table partition - Real time Data

Hi team,We are currently working in loading CDF table using data events from Kafka. The table is going to hold data across geographies. When we tried partitioning it is slowing down the ingestion time. But without partition the downstream application...

  • 231 Views
  • 1 replies
  • 1 kudos
Latest Reply
ozaaditya
Contributor
  • 1 kudos

 1. Instead of using many small partitions (e.g., country or region), opt for larger partitions, such as continent or time-based partitions (e.g., weekly or monthly). This will reduce the number of partitions and improve performance.2. Write data to ...

  • 1 kudos
binsel
by New Contributor III
  • 361 Views
  • 3 replies
  • 2 kudos

Resolved! UNPIVOT VARIANT data in SQL

Hi All,Have a VARIANT column with the following data;CREATE TABLE unpivot_valtype AS SELECT parse_json( '{ "Id": 1234567, "Result": { "BodyType": "NG", "ProdType": "Auto", "ResultSets": [ { "R1": { "AIn...

result.png
  • 361 Views
  • 3 replies
  • 2 kudos
Latest Reply
filipniziol
Contributor III
  • 2 kudos

Hi @binsel ,You need to use variant_explode function.Here is the working code:WITH first_explode AS ( SELECT uv.rowData:Id AS Id, uv.rowData:Result:BodyType AS BodyType, uv.rowData:Result:ProdType AS ProdType, v.value AS result_se...

  • 2 kudos
2 More Replies
erigaud
by Honored Contributor
  • 232 Views
  • 2 replies
  • 3 kudos

Databricks asset bundles and Dashboards - pass parameters depending on bundle target

Hello everyone !Since Databricks Asset Bundles can now be used to deploy dashboards, I'm wondering how to pass parameters so that the queries for the dev dashboard query the dev catalog, and the dashboard in stg query the stg catalog etc.Is there any...

  • 232 Views
  • 2 replies
  • 3 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 3 kudos

Hi @erigaud, Here’s how you can achieve this:   Define Variables in the Bundle Configuration: You can define custom variables in your databricks.yml file to hold the catalog names for different environments. For example:   variables:    catalog_name...

  • 3 kudos
1 More Replies
devagya
by New Contributor
  • 222 Views
  • 2 replies
  • 1 kudos

Infor Data Lake to Databricks

I'm working on this project which involves moving data from Infor to Databricks.Infor is somewhat of an enterprise solution. I could not find much resources on this. I could not even find any free trial option on their site.If anyone has experience w...

  • 222 Views
  • 2 replies
  • 1 kudos
Latest Reply
michelle653burk
New Contributor III
  • 1 kudos

@devagya wrote:I'm working on this project which involves moving data from Infor to Databricks.Infor is somewhat of an enterprise solution. I could not find much resources on this. I could not even find any free trial option on their site.If anyone h...

  • 1 kudos
1 More Replies
schluca
by New Contributor II
  • 172 Views
  • 1 replies
  • 0 kudos

Error Querying Shallow Clones: Couldn't Initialize File System for Path

Hi,We are offering data products through a central catalog for our users. To minimize data duplication and to display relationships between tables, we use shallow clones to provide access to the data.However, since implementing this approach, we occa...

  • 172 Views
  • 1 replies
  • 0 kudos
Latest Reply
TakuyaOmi
Valued Contributor
  • 0 kudos

Hi @schluca ,I’ve encountered an issue where an error occurred when trying to reference a table after deleting and recreating the source table for a Shallow Clone, and then performing the Shallow Clone again. As a solution, try deleting the destinati...

  • 0 kudos
Rahman823
by New Contributor II
  • 275 Views
  • 2 replies
  • 1 kudos

Databricks table lineage

Hi,I wanted to know if it is possible to edit the lineage that we see in databricks, like the one shown below.Can I edit this lineage graph, like add other ETL tools (at the start of the tables) that I have used to get data in aws and then in databri...

Rahman823_0-1731512283889.png
  • 275 Views
  • 2 replies
  • 1 kudos
Latest Reply
chm_user_1
New Contributor II
  • 1 kudos

This will be extremely beneficial. We have certain use cases where we do not leverage Spark in our pipelines and lose the lineage. I would prefer to set an extra parameter when writing a table to specify the lineage. 

  • 1 kudos
1 More Replies
vinitkhandelwal
by New Contributor III
  • 207 Views
  • 2 replies
  • 0 kudos

Error while running a notebook job using git repo (Gitlab)

I am trying to run a notebook job with a git repo hosted on Gitlab. I have Linked my Gitlab account using Gitlab tokenYet i am getting the following error on running the job How to resolve this?

Screenshot 2024-12-02 at 7.56.10 AM.png Screenshot 2024-12-02 at 7.58.11 AM.png
  • 207 Views
  • 2 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

Hi @vinitkhandelwal, Looks like the token could be missing required permissions for the operation. Please refer to: You can clone public remote repositories without Git credentials (a personal access token and a username). To modify a public remote r...

  • 0 kudos
1 More Replies
sharukh_lodhi
by New Contributor III
  • 1078 Views
  • 3 replies
  • 2 kudos

Azure IMDS is not accesbile selecting shared compute policy

Hi, Databricks community,I recently encountered an issue while using the 'azure.identity' Python library on a cluster set to the personal compute policy in Databricks. In this case, Databricks successfully returns the Azure Databricks managed user id...

image.png
Data Engineering
azure IMDS
DefaultAzureCredential
  • 1078 Views
  • 3 replies
  • 2 kudos
Latest Reply
daisy08
New Contributor II
  • 2 kudos

I'm having a similar problem, my aim is from an Azure DataBricks notebook to invoke an AzureDataDactory pipeline I created an Access Connector for Azure Databricks to which I gave Data Factory Contributor permissions. Using these lines pythonfrom azu...

  • 2 kudos
2 More Replies

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.

If there isn’t a group near you, start one and help create a community that brings people together.

Request a New Group
Labels