cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Malthe
by New Contributor II
  • 0 Views
  • 0 replies
  • 0 kudos

Parametrize DLT pipeline

If I'm using Databricks Asset Bundles, how would I parametrize a DLT pipeline based on a static configuration file.In pseudo-code, I would have a .py-file:import dlt # Something that pulls a pipeline resource (or artifact) and parses from JSON table...

  • 0 Views
  • 0 replies
  • 0 kudos
vziog
by Visitor
  • 19 Views
  • 1 replies
  • 0 kudos

Costs from cost managem azure portal are not allligned with costs calculated from usage system table

Hello,the costs regarding the databricks service from cost management in azure portal (45,869...) are not allligned with costs calculated from usage system table (75,34). The costs from the portal are filtered based on the desired period (usage_date ...

  • 19 Views
  • 1 replies
  • 0 kudos
Latest Reply
matthew554
Visitor
  • 0 kudos

@vziog wrote:Hello,the costs regarding the databricks service from cost management in azure portal (45,869...) are not allligned with costs calculated from usage system table (75,34). The costs from the portal are filtered based on the desired period...

  • 0 kudos
dc-rnc
by New Contributor III
  • 822 Views
  • 3 replies
  • 0 kudos

DAB | Set tag based on job parameter

Hi Community.Since I wasn't able to find a way to set a job tag dynamically at runtime based on a parameter that is passed to the job, I was wondering if it is possible or if there is an equivalent way to do it.Thank you. Regards.

  • 822 Views
  • 3 replies
  • 0 kudos
Latest Reply
BigRoux
Databricks Employee
  • 0 kudos

Based on the provided context, it appears that there isn't a direct way within Databricks to dynamically set job tags at runtime based on a parameter passed to the job. However, there are alternative approaches you can consider to work around this li...

  • 0 kudos
2 More Replies
cs_de
by Visitor
  • 40 Views
  • 4 replies
  • 3 kudos

How do I deploy or run one job if I have multiple jobs in a Databricks Asset Bundle?

How do I deploy or run a single job if I have 2 or more jobs defined in my asset bundle?$databricks bundle deploy job1 #does not work I do not see a flag to identify what job to run.

  • 40 Views
  • 4 replies
  • 3 kudos
Latest Reply
mark_ott
Databricks Employee
  • 3 kudos

I haven't done it with multiple jobs, but I think under resources you name multiple jobs, then when you deploy you just call that job key.  

  • 3 kudos
3 More Replies
petitregny
by New Contributor
  • 59 Views
  • 1 replies
  • 0 kudos

Reading from an S3 bucket using boto3 on serverless cluster

Hello All,I am trying to read a CSV file from my S3 bucket in a notebook running on serverless.I am using the two standard functions below, but I get a credentials error (Error reading CSV from S3: Unable to locate credentials).I don't have this issu...

  • 59 Views
  • 1 replies
  • 0 kudos
Latest Reply
cgrant
Databricks Employee
  • 0 kudos

For use cases where you want to use cloud service credentials to authenticate to cloud services, I recommend using Unity Catalog Service Credentials. These work with serverless and class compute in Databricks. You'd create a service credential, and t...

  • 0 kudos
smpa01
by New Contributor
  • 15 Views
  • 0 replies
  • 0 kudos

Debugging jobs/run-now endpoint

I am not being able to run jobs/runnow endpoint. I am getting an error asError fetching files: 403 - {"error_code":"PERMISSION_DENIED","message":"User xxxx-dxxxx-xxx-xxxx does not have Manage Run or Owner or Admin permissions on job 437174060919465",...

smpa01_0-1744904979789.png smpa01_1-1744905123010.png
  • 15 Views
  • 0 replies
  • 0 kudos
GregTyndall
by New Contributor II
  • 637 Views
  • 3 replies
  • 0 kudos

Resolved! Materialized View Refresh - NUM_JOINS_THRESHOLD_EXCEEDED?

I have a very basic view with 3 inner joins that will only do a full refresh. Is there a limit to the number of joins you can have and still get an incremental refresh?"incrementalization_issues": [{"issue_type": "INCREMENTAL_PLAN_REJECTED_BY_COST_MO...

  • 637 Views
  • 3 replies
  • 0 kudos
Latest Reply
PotnuruSiva
Databricks Employee
  • 0 kudos

@GregTyndall Yes, the current limit is 2 by default. But we can increase up to 5 with the below flag added to the pipeline settings. pipelines.enzyme.numberOfJoinsThreshold 5

  • 0 kudos
2 More Replies
Christian_C
by Visitor
  • 11 Views
  • 0 replies
  • 0 kudos

Google Pub Sub and Delta live table

I am using delta live table and pub sub to ingest message from 30 different topics in parallel. I noticed that initialization time can be very long around 15 minutes. Does someone knows how to reduced initialization time in dlt ? Thanks You 

  • 11 Views
  • 0 replies
  • 0 kudos
Chris_sh
by New Contributor II
  • 2229 Views
  • 2 replies
  • 1 kudos

[STREAMING_TABLE_OPERATION_NOT_ALLOWED.REQUIRES_SHARED_COMPUTE]

Currently trying to refresh a Delta Live Table using a Full Refresh but an error keeps coming up saying that we have to use a shared cluster or a SQL warehouse. I've tried both a shared cluster and a SQL warehouse and the same error keeps coming up. ...

  • 2229 Views
  • 2 replies
  • 1 kudos
Latest Reply
BigRoux
Databricks Employee
  • 1 kudos

You are not using "No Isolation Shared" mode, right?  Also, can you share the chunk of code that is causing the failure? Thanks, Louis.

  • 1 kudos
1 More Replies
guest0
by New Contributor II
  • 549 Views
  • 6 replies
  • 3 kudos

Spark UI Simulator Not Accessible

Hello,The Spark UI Simulator is not accessible since the last few days. I was able to refer to it last week, at https://www.databricks.training/spark-ui-simulator/index.html. I already have access to partner academy (if that is any relevant).  <Error...

Data Engineering
simulator
spark-ui
  • 549 Views
  • 6 replies
  • 3 kudos
Latest Reply
guest0
New Contributor II
  • 3 kudos

Just a short update: the request I raised was closed saying there is no active support contract with the org (from the email I used) to look into this. Perhaps someone else could try raising a request using the link above.

  • 3 kudos
5 More Replies
HaripriyaP
by New Contributor
  • 816 Views
  • 3 replies
  • 0 kudos

Multiple Notebooks Migration from one workspace to another without using Git.

Hi all!I need to migrate multiple notebooks from one workspace to another. Is there any way to do it without using Git?Since Manual Import and Export is difficult to do for multiple notebooks and folders, need an alternate solution.Please reply as so...

  • 816 Views
  • 3 replies
  • 0 kudos
Latest Reply
RiyazAli
Valued Contributor II
  • 0 kudos

Hello @HaripriyaP , @rabia_farooq check for the databricks cli documentation here - https://docs.databricks.com/aws/en/dev-tools/cli/commandsYou can work with databricks workflow <command_such_as_export> and move the notebooks/code files around.Cheer...

  • 0 kudos
2 More Replies
Vasu_Kumar_T
by New Contributor II
  • 92 Views
  • 3 replies
  • 1 kudos

Data Migration using Bladebridge

Hi,We are planning to migrate from Teradata to Databricks using Bladebridge. Going through various portals, I am not able to conclude the component that facilitates Data movement between Teradata and databricks.Please clarify end to end tool and acti...

  • 92 Views
  • 3 replies
  • 1 kudos
Latest Reply
RiyazAli
Valued Contributor II
  • 1 kudos

I'm not aware if blade bridge has data movement tool handy with them.I don't see anything advertised by them though.Let me know if you find anything on this.

  • 1 kudos
2 More Replies
minhhung0507
by Contributor III
  • 217 Views
  • 11 replies
  • 2 kudos

API for Restarting Individual Failed Tasks within a Job?

Hi everyone,I'm exploring ways to streamline my workflow in Databricks and could really use some expert advice. In my current setup, I have a job (named job_silver) with multiple tasks (e.g., task 1, task 2, task 3). When one of these tasks fails—say...

  • 217 Views
  • 11 replies
  • 2 kudos
Latest Reply
RiyazAli
Valued Contributor II
  • 2 kudos

Hey @minhhung0507 - quick question - what is the cluster type you're using to run your workflow?I'm using a shared, interactive cluster, so I'm passing the parameter {'existing_cluster_id' : task['existing_cluster_id']}in the payload. This parameter ...

  • 2 kudos
10 More Replies
yashojha1995
by New Contributor
  • 137 Views
  • 1 replies
  • 0 kudos

Error while running update statement using delta lake linked service through ADF

Hi All, I am getting the below error while running an update query in a lookup activity using the delta lake linked service:ErrorCode=AzureDatabricksCommandError,Hit an error when running the command in Azure Databricks. Error details: <span class='a...

  • 137 Views
  • 1 replies
  • 0 kudos
Latest Reply
RiyazAli
Valued Contributor II
  • 0 kudos

Hi @yashojha1995 EOL while scanning string literal hints that there might be a syntax error in the update query.could you share your update query here, and any other info such as how are you creating a Linked service to your delta lake? Does it mean ...

  • 0 kudos

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels