- 2 Views
- 0 replies
- 0 kudos
May be I am new to Databricks that's why I have confusion.Suppose I have worker memory of 64gb in Databricks job max 12 nodes...and my job is failing due to Executor Lost due to 137 (OOM if found on internet).So, to fix this I need to increase execut...
- 2 Views
- 0 replies
- 0 kudos
- 7 Views
- 0 replies
- 0 kudos
Hi All,Please help me understand how the billing is calculated for using the Job cluster.Document says they are charged hourly basis, so if my job ran for 1hr 30mins then will be charged for the 30mins based on the hourly rate or it will be charged f...
- 7 Views
- 0 replies
- 0 kudos
- 7 Views
- 0 replies
- 0 kudos
I am having an issue where when I do a shallow clone using :create or replace table `catalog_a_test`.`schema_a`.`table_a` shallow clone `catalog_a`.`schema_a`.`table_a` I get:[TABLE_OR_VIEW_NOT_FOUND] The table or view catalog_a_test.schema_a.table_a...
- 7 Views
- 0 replies
- 0 kudos
- 394 Views
- 3 replies
- 0 kudos
Hi,I am migrating from dbx to databricks asset bundles. Previously with dbx I could work on different features in separate branches and launch jobs without issue of one job overwritting the other. Now with databricks asset bundles it seems like I can...
- 394 Views
- 3 replies
- 0 kudos
Latest Reply
Any updates here?My team is migrating from dbx to DABs and we are running into the same issue. Ideally, we would like to deploy multiple, parametrized jobs from a single bundle. If this is not possible, we have to keep dbx.Thank you!
2 More Replies
- 15 Views
- 0 replies
- 0 kudos
Good morning, I'm trying to run: databricks bundle run --debug -t dev integration_tests_job My bundle looks: bundle:
name: x
include:
- ./resources/*.yml
targets:
dev:
mode: development
default: true
workspace:
host: x
r...
- 15 Views
- 0 replies
- 0 kudos
- 12 Views
- 0 replies
- 0 kudos
Getting following error while saving a dataframe partitioned by two columns.Job aborted due to stage failure: Task 5774 in stage 33.0 failed 4 times, most recent failure: Lost task 5774.3 in stage 33.0 (TID 7736) (13.2.96.110 executor 7): ExecutorLos...
- 12 Views
- 0 replies
- 0 kudos
- 33 Views
- 0 replies
- 0 kudos
I am currently exploring testing methodologies for Databricks notebooks and would like to inquire whether it's possible to write pytest tests for notebooks that contain code not encapsulated within functions or classes.***********************a = 4b ...
- 33 Views
- 0 replies
- 0 kudos
by
Phani1
• Valued Contributor
- 158 Views
- 4 replies
- 0 kudos
Hi Team,Please provide guidance on enabling SQL cells parallel execution in a notebook containing multiple SQL cells. Currently, when we execute notebook and all the SQL cells they run sequentially. I would appreciate assistance on how to execute th...
- 158 Views
- 4 replies
- 0 kudos
Latest Reply
Hi @Phani1 Yes you can achieve this scenario with the help of Databricks Workflow jobs where you can create task and dependencies for each other.
3 More Replies
- 112 Views
- 5 replies
- 0 kudos
I need help with migrating from dbfs on databricks to workspace. I am new to databricks and am struggling with what is on the links provided.My workspace.yml also has dbfs hard-coded. Included is a full deployment with great expectations.This was don...
- 112 Views
- 5 replies
- 0 kudos
Latest Reply
There's also this KB specific to init script migration - https://kb.databricks.com/clusters/migration-guidance-for-init-scripts-on-dbfs
4 More Replies
- 119 Views
- 2 replies
- 0 kudos
There are multiple tables in the config/metadata table. These tables need to bevalidated for DQ rules.1.Natural Key / Business Key /Primary Key cannot be null orblank.2.Natural Key/Primary Key cannot be duplicate.3.Join columns missing values4.Busine...
- 119 Views
- 2 replies
- 0 kudos
Latest Reply
Hi @subha2, To dynamically validate the data quality (DQ) rules for tables configured in a metadata-driven system using PySpark, you can follow these steps:
Define Metadata for Tables:
First, create a metadata configuration that describes the rules ...
1 More Replies
- 613 Views
- 6 replies
- 0 kudos
Do you know why the userIdentity is anonymous in AWS Cloudtail's logs even though I have specified an instance profile?
- 613 Views
- 6 replies
- 0 kudos
Latest Reply
Thank you for posting your question in our community! We are happy to assist you.To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best answers your question?This...
5 More Replies
- 925 Views
- 4 replies
- 2 kudos
from pyspark.sql import functions as F
from pyspark.sql import types as T
from pyspark.sql import DataFrame, Column
from pyspark.sql.types import Row
import dlt
S3_PATH = 's3://datalake-lab/XXXXX/'
S3_SCHEMA = 's3://datalake-lab/XXXXX/schemas/'
...
- 925 Views
- 4 replies
- 2 kudos
Latest Reply
Was this resolved? I run into the same issue
3 More Replies
- 1396 Views
- 4 replies
- 1 kudos
Hello,I am attempting to configure Autoloader in File Notification mode with Delta Live Tables. I configured an instance profile, but it is not working because I immediately get AWS access denied errors. This is the same issue that is referenced here...
- 1396 Views
- 4 replies
- 1 kudos
Latest Reply
Hi, I'm running into the same issue. Was this solved?
3 More Replies
- 39 Views
- 0 replies
- 0 kudos
Currently, the bronze table ingests JSON files using @Dlt.table decorator on a spark.readStream functionA daily batch job does some transformation on bronze data and stores results in the silver table.New ProcessBronze still the same.A stream has bee...
- 39 Views
- 0 replies
- 0 kudos
- 2397 Views
- 5 replies
- 14 kudos
I see that spark fully supports Scala 2.13. I wonder why is there no databricks runtime with Scala 2.13 yet. Any plans on making this available? It would be super useful.
- 2397 Views
- 5 replies
- 14 kudos
Latest Reply
I see db runtime 14 is out, but still 2.12, when would databricks plan to support 2.13 or 3 thank you
4 More Replies