cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

slimbnsalah
by New Contributor II
  • 2271 Views
  • 2 replies
  • 0 kudos

Use Salesforce Lakeflow Connector with a Salesforce Connected App

Hello, I'm trying to use the new Salesforce Lakeflow connector to ingest data into my Databricks account.However I see only the option to connect using a normal user, whereas I want to use a Salesforce App, just like how it is described here Run fede...

  • 2271 Views
  • 2 replies
  • 0 kudos
Latest Reply
Ajay-Pandey
Databricks MVP
  • 0 kudos

@slimbnsalah Please select Connection type as of Salesforce Data Cloud then you will be asked for details 

  • 0 kudos
1 More Replies
ManojkMohan
by Honored Contributor II
  • 955 Views
  • 4 replies
  • 2 kudos

Resolved! Silver to Gold Layer | Running ML - Debug Help Needed

Problem I am solving:Reads the raw sports data  IPL CSV → bronze layerCleans and aggregates → silver layerSummarizes team stats → gold layerPrepares ML-ready features and trains a Random Forest classifier to predict match winners Getting error: [PARS...

ManojkMohan_0-1756389913835.png
  • 955 Views
  • 4 replies
  • 2 kudos
Latest Reply
BS_THE_ANALYST
Databricks Partner
  • 2 kudos

@ManojkMohan thanks for sharing this, I'm looking at starting an ML project in the coming weeks, I might have to bring this forward . Feeling motivated with that confusion matrix in your output .Congrats on getting it working!All the best,BS

  • 2 kudos
3 More Replies
Srinivas5
by New Contributor II
  • 1068 Views
  • 6 replies
  • 3 kudos

Jar File Upload To Workspace

Spoiler  #dbfsI am unable to upload jar file dbfs to job cluster as it's deprecated now I need to upload it to workspace and install it to cluster, hower my jar size is 70mb i can't upload it through api or cli as max size is 50mb. Is there alternati...

  • 1068 Views
  • 6 replies
  • 3 kudos
Latest Reply
Advika
Community Manager
  • 3 kudos

Hi @Srinivas5! Were you able to find a solution or approach that worked? If so, please mark the helpful reply as the Accepted Solution, or share your approach so others can benefit as well.

  • 3 kudos
5 More Replies
ShankarM
by Databricks Partner
  • 517 Views
  • 2 replies
  • 0 kudos

Notebook exposure

i have created a notebook as per client requirement. I have to migrate the notebook in the client env for testing with live data but do not want to expose the Databricks notebook code to the testers in the client env.Is there a way to package the not...

  • 517 Views
  • 2 replies
  • 0 kudos
Latest Reply
WiliamRosa
Databricks Partner
  • 0 kudos

Hi @ShankarM,I’ve had to do something similar—packaging a Python class as a wheel. This documentation might help: https://docs.databricks.com/aws/en/dev-tools/bundles/python-wheel

  • 0 kudos
1 More Replies
DatabricksEngi1
by Contributor
  • 1700 Views
  • 2 replies
  • 1 kudos

Resolved! databricks assets bundles issue

Hii all,I’m working with Databricks Asset Bundles (DAB) and trying to move from a single repository-level bundle to a structure where each workflow (folder under resources/jobs) has its own bundle.• My repository contains:• Shared src/variables.yml a...

  • 1700 Views
  • 2 replies
  • 1 kudos
Latest Reply
DatabricksEngi1
Contributor
  • 1 kudos

I solved it.For some reason, the Terraform folder created under the bundles wasn’t set up correctly.I copied it from a working bundle, and everything completed successfully.

  • 1 kudos
1 More Replies
JPNP
by Databricks Partner
  • 1499 Views
  • 3 replies
  • 1 kudos

Not able to creare Secret scope in Azure databricks

Hello,I am trying to create the  Azure Key Vault-backed secret scope, but it failing with the below error, I have tried to clear the cache, and logged out , used incognito browser as well but not able to create a scope. Can you please help here ? 

JPNP_0-1755692310711.jpeg
  • 1499 Views
  • 3 replies
  • 1 kudos
Latest Reply
Yogesh_Verma_
Contributor II
  • 1 kudos

If the UI keeps failing with that vague error, the CLI approach suggested above is the best next step, since it usually gives a clearer error message. Also make sure that:The service principal you’re using to create the scope has Key Vault Administra...

  • 1 kudos
2 More Replies
jar
by Contributor
  • 476 Views
  • 1 replies
  • 0 kudos

Excluding job update from DAB .yml deployment

Hi.We have a range of scheduled jobs and _one_ continuous job all defined in .yml and deployed with DAB. The continuous job is paused per default and we use a scheduled job of a notebook to pause and unpause it so that it only runs during business ho...

  • 476 Views
  • 1 replies
  • 0 kudos
Latest Reply
Yogesh_Verma_
Contributor II
  • 0 kudos

You’re running into this because DAB treats the YAML definition as the source of truth — so every time you redeploy, it will reset the job state (including the paused/running status) back to what’s defined in the file. Unfortunately, there isn’t curr...

  • 0 kudos
karthik_p
by Databricks Partner
  • 16533 Views
  • 5 replies
  • 1 kudos

does delta live tables supports identity columns

we are able to test identity columns using sql/python, but when we are trying same using DLT, we are not seeing values under identity column. it is always empty for coloumn we created "id BIGINT GENERATED ALWAYS AS IDENTITY" 

  • 16533 Views
  • 5 replies
  • 1 kudos
Latest Reply
Gowrish
New Contributor II
  • 1 kudos

Hi,i see from the following databricks documentaion - https://docs.databricks.com/aws/en/dlt/limitationsit states the following which kind of giving an impression that you can define identity column to a steaming table Identity columns might be recom...

  • 1 kudos
4 More Replies
mtreigelman
by New Contributor III
  • 717 Views
  • 1 replies
  • 3 kudos

First Lakeflow (DLT) Pipeline Best Practice Question

Hi, I am writing my first streaming pipeline and trying to ensure it is setup to work as a "Lakeflow" pipeline.  It is connecting an external Oracle database with some external Azure Blob storage data (all managed in the same Unity Catalog). The pipe...

  • 717 Views
  • 1 replies
  • 3 kudos
Latest Reply
BS_THE_ANALYST
Databricks Partner
  • 3 kudos

@mtreigelmanthanks for providing the update. If you wouldn't mind, could you explain why you think the first way didn't work and why the second way did? Then you can mark your response as a solution to the question .I found this article to be useful ...

  • 3 kudos
ck7007
by Contributor II
  • 746 Views
  • 1 replies
  • 2 kudos

Cost

Reduced Monthly Databricks Bill from $47K to $12.7KThe Problem: We were scanning 2.3TB for queries needing only 8GB of data.Three Quick Wins1. Multi-dimensional Partitioning (30% savings)# Beforedf.write.partitionBy("date").parquet(path)# After-parti...

  • 746 Views
  • 1 replies
  • 2 kudos
Latest Reply
BS_THE_ANALYST
Databricks Partner
  • 2 kudos

@ck7007 thanks so much for sharing! That's such a saving, by the way. Congrats.Out of curiosity, did you consider using Liquid Clustering which was meant to replace partitioning and z-order: https://docs.databricks.com/aws/en/delta/clustering I found...

  • 2 kudos
AbhishekNakka15
by Databricks Partner
  • 702 Views
  • 1 replies
  • 1 kudos

Resolved! Unable to login to partner account

When I try to login with my office email to the partner acccount. It says, The service is currently unavailable. Please try again later. It says "You are not authorized to access https://partner-academy.databricks.com. Please select a platform you ca...

  • 702 Views
  • 1 replies
  • 1 kudos
Latest Reply
Advika
Community Manager
  • 1 kudos

Hello @AbhishekNakka15! Please raise a ticket with the Databricks Support Team, and include your email address so they can review your account and provide further assistance.

  • 1 kudos
viralpatel
by New Contributor II
  • 1229 Views
  • 2 replies
  • 1 kudos

Lakebridge Synapse Conversion to DBX and Custom transpiler

I have 2 questions about Lakebridge solution,Synapse with dedicated pool ConversionWe were conducting a PoC for Synapse to DBX migration using Lakebridge. What we have observed is that the conversions are not correct. I was anticipating all tables wi...

  • 1229 Views
  • 2 replies
  • 1 kudos
Latest Reply
yourssanjeev
Databricks Partner
  • 1 kudos

We are also checking on this use case but got it confirmed from Databricks that it does not work for this use case yet, not sure whether it is in their roadmap

  • 1 kudos
1 More Replies
vishalv4476
by New Contributor III
  • 624 Views
  • 1 replies
  • 0 kudos

Databricks job runs failures Py4JJavaError: An error occurred while calling o404.sql. : java.util.No

Hi ,We had a successful running pipeline but it started failing since 20th august , no change were published. Can you please guide me resolve this issue.I've tried increasing delta.deletedFileRetentionDuration' = 'interval 365 days' but it didn't hel...

  • 624 Views
  • 1 replies
  • 0 kudos
Latest Reply
SP_6721
Honored Contributor II
  • 0 kudos

Hi @vishalv4476 ,The error is likely due to a corrupted Delta transaction log or files deleted manually/outside of Delta. Check the table history and verify that no user or automated process removed data files. If issues are found, restore the table ...

  • 0 kudos
anazen13
by New Contributor III
  • 1805 Views
  • 9 replies
  • 2 kudos

databricks api to create a serverless job

I am trying to follow your documentation on how to create serverless job via API https://docs.databricks.com/api/workspace/jobs/create#environments-spec-environment_version So i see that sending the json request resulted for me to see serverless clus...

  • 1805 Views
  • 9 replies
  • 2 kudos
Latest Reply
siennafaleiro
New Contributor II
  • 2 kudos

It looks like you’re hitting one of the current limitations of Databricks serverless jobs. Even though the API supports passing an environments object, only certain fields are honored right now. In particular:The environment_version parameter will de...

  • 2 kudos
8 More Replies
zero234
by New Contributor III
  • 7317 Views
  • 3 replies
  • 1 kudos

i have created a materialized view table using delta live table pipeline and its not appending data

i have created a materialized view table using delta live table pipeline , for some reason it is overwriting data every day , i want it to append data to the table instead of doing full refresh suppose i had 8 million records in table and if irun the...

  • 7317 Views
  • 3 replies
  • 1 kudos
Latest Reply
UMAREDDY06
New Contributor II
  • 1 kudos

[expect_table_not_view.no_alternative] 'insert' expects a table but dim_airport_unharmonised is a view can you please help how to reslove this.thanksuma devi

  • 1 kudos
2 More Replies
Labels