cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

jeremy98
by Contributor III
  • 448 Views
  • 0 replies
  • 0 kudos

Environment set up in serveless notebook task

Hi community,Is there a way to install dependencies inside a notebook task using serveless compute using Databricks Asset Bundle? Is there a way to avoid installing everytime for each serverless task that compose a job the dependencies (or the librar...

  • 448 Views
  • 0 replies
  • 0 kudos
Vetrivel
by Contributor
  • 436 Views
  • 0 replies
  • 0 kudos

Federate AWS Cloudwatch logs to Databricks Unity Catalog

I am looking to integrate CloudWatch logs with Databricks. Our objective is not to monitor Databricks via CloudWatch, but rather to facilitate access to CloudWatch logs from within Databricks. If anyone has implemented a similar solution, kindly prov...

  • 436 Views
  • 0 replies
  • 0 kudos
sachamourier
by New Contributor III
  • 202 Views
  • 1 replies
  • 0 kudos

Install Python libraries on Databricks job cluster

Hello,I am trying to install some wheel file and requirements.txt file from my Unity Catalog Volumes on my Databricks job cluster using an init script but the results are very inconsistent.Does anyone have ever faced that ?What's wrong with my approa...

job_cluster_issue.png job_cluster_issue_1.png job_cluster_issue_2.png
  • 202 Views
  • 1 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

Hi @sachamourier, Could you please clarify what is the inconsistency? are some packages missing or the incorrect library was loaded?

  • 0 kudos
LeenB
by New Contributor
  • 169 Views
  • 1 replies
  • 0 kudos

Running a notebook as 'Run all below' when sheduled via Azure DataFactory

We have a notebook with a lot of subsequent cells that can run independent from each other. When we execute the notebook manually via 'Run all', the runs stops when an error is thrown. When we execute manually via 'Run all below', the run proceeds ti...

  • 169 Views
  • 1 replies
  • 0 kudos
Latest Reply
PiotrMi
New Contributor III
  • 0 kudos

Hi @LeenB For example each cell execution you can build up with try except command. Example belowtry:     print("Hello world")    #your code of each cellexcept Exception as e:    print("Issue with printing hello world")For sure it is not recommended ...

  • 0 kudos
Deenar
by New Contributor II
  • 159 Views
  • 1 replies
  • 0 kudos

Dashboard Filters (Showing Description) but selecting the id value for use query parameters

HiIs there a way to have a Single or multiple filter in a Dashboard that shows titles/names but on selection passes in the identifier as query parameter filters. The cost management dashboard shipped by Databricks seems to do this for the workspace s...

Deenar_0-1740567476099.png
  • 159 Views
  • 1 replies
  • 0 kudos
Latest Reply
Deenar
New Contributor II
  • 0 kudos

HiFor those interested in the same. After digging into the Databricks cost dashboards code, I found out that this feature isn't supported yet. The Databricks dashboards use a hack to get around this by having a lookup table in each dataset definition...

  • 0 kudos
data_mifflin
by New Contributor III
  • 380 Views
  • 6 replies
  • 1 kudos

Accessing Job parameters using cluster v15.4

After upgrading databricks cluster to version 15.4, is there any way to access job parameters in notebook except the following way ?dbutils.widgets.get("parameter_name")In v15.4, dbutils.notebook.entry_point.getCurrentBindings() has been discontinued...

  • 380 Views
  • 6 replies
  • 1 kudos
Latest Reply
Pawan1979
New Contributor II
  • 1 kudos

For me it is working at 15.4 LTS (includes Apache Spark 3.5.0, Scala 2.12)

  • 1 kudos
5 More Replies
JW_99
by New Contributor II
  • 291 Views
  • 2 replies
  • 2 kudos

PySparkRuntimeError: [CONTEXT_ONLY_VALID_ON_DRIVER]

I've troubleshot this like 20+ times. I am aware that the current code is causing the spark session to be passed to the workers, where it should only be applied to the driver. Can someone please help me resolve this (the schema is defined earlier)?--...

JW_99_0-1740614786516.png JW_99_1-1740614786523.png JW_99_2-1740614786524.png JW_99_3-1740614786524.png
  • 291 Views
  • 2 replies
  • 2 kudos
Latest Reply
narasimha_reddy
New Contributor II
  • 2 kudos

You cannot use Spark session explicitly inside Executor logic. Here you are trying mapPartitions which makes the customlogic to get executed inside the executor thread. Either you need to change whole problem approach to segregate spark variable usag...

  • 2 kudos
1 More Replies
adhi_databricks
by New Contributor III
  • 340 Views
  • 4 replies
  • 0 kudos

Connect snowflake to Databricks

Hey Folks,I just want to know if there is a way to mirror the Snowflake tables in Databricks , Meaning creating a table using format snowflake and give in options of table (host,user,pwd and dbtable in snowflake). I just tried it as per this code bel...

  • 340 Views
  • 4 replies
  • 0 kudos
Latest Reply
adhi_databricks
New Contributor III
  • 0 kudos

Hi @Alberto_Umana , Just a QQ would we be able to change table properties like adding column details, column tagging and Column level masking on the snowflake tables that are under the foreign catalog created?

  • 0 kudos
3 More Replies
nikhilkumawat
by New Contributor III
  • 13036 Views
  • 11 replies
  • 8 kudos

Resolved! Get file information while using "Trigger jobs when new files arrive" https://docs.databricks.com/workflows/jobs/file-arrival-triggers.html

I am currently trying to use this feature of "Trigger jobs when new file arrive" in one of my project. I have an s3 bucket in which files are arriving on random days. So I created a job to and set the trigger to "file arrival" type. And within the no...

  • 13036 Views
  • 11 replies
  • 8 kudos
Latest Reply
Jaison
New Contributor II
  • 8 kudos

Issue with Databricks File Arrival Trigger – Missing File Name InformationThe File Arrival Trigger in Databricks is practically useless if it does not provide the file name and path of the triggering file. In Azure Blob Storage triggers (Function App...

  • 8 kudos
10 More Replies
jakub_adamik
by New Contributor II
  • 241 Views
  • 2 replies
  • 0 kudos

Resolved! Delta Live Tables - BAD_REQUEST: Pipeline cluster is not reachable.

Hi all,I have very simple pipeline: -- Databricks notebook source CREATE OR REFRESH STREAMING TABLE `catalog-prod`.default.dlt_table AS SELECT * FROM STREAM read_files('/Volumes/catalog-prod/storage/*', format=> 'json') -- COMMAND ---------- CREATE...

jakub_adamik_0-1740565875990.png
  • 241 Views
  • 2 replies
  • 0 kudos
Latest Reply
jakub_adamik
New Contributor II
  • 0 kudos

Hi,thank you for your response. In the mean time I found the bug Databricks UI which caused this behaviour. I will raise ticket to Databricks. Please see the draft of the ticket bellow for workaround:  We’re facing an issue with Delta Live Tables pip...

  • 0 kudos
1 More Replies
wilmorlserios
by New Contributor
  • 162 Views
  • 1 replies
  • 0 kudos

Using databricks-sql-connector in Notebook

I am attempting to utilse the databricks-sql-connector python package within a generalised application deployed to run within a Databricks notebook. Upon attempting to import, I am receiving a  module not found error. However, the package is visible ...

wilmorlserios_0-1740576722972.png wilmorlserios_1-1740576738825.png
  • 162 Views
  • 1 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

Hi @wilmorlserios  The import it's incorrect. It should be: from databricks import sql

  • 0 kudos
vk217
by Contributor
  • 12502 Views
  • 4 replies
  • 1 kudos

ModuleNotFoundError: No module named 'pyspark.dbutils'

I have a class in a python file like this from pyspark.sql import SparkSession from pyspark.dbutils import DBUtils class DatabricksUtils: def __init__(self‌‌): self.spark = SparkSession.getActiveSession() self.dbutils = DBUtil...

  • 12502 Views
  • 4 replies
  • 1 kudos
Latest Reply
T0M
New Contributor III
  • 1 kudos

Had the same Problem in my GitLab CI/CD Pipeline while trying to deploy: $ databricks bundle deploy -t dev Building package... Error: build failed package, error: exit status 1, output: Traceback (most recent call last): [...] File "/builds/user/...

  • 1 kudos
3 More Replies
MathewDRitch
by New Contributor II
  • 2635 Views
  • 5 replies
  • 1 kudos

Connecting from Databricks to Network Path

Hi All,Will appreciate if someone can help me with some references links on connecting from Databricks to external network path. I have Databricks on AWS and previously used to connect to files on external network path using Mount method. Now Databri...

  • 2635 Views
  • 5 replies
  • 1 kudos
Latest Reply
om_khade
New Contributor II
  • 1 kudos

Do we have any update on this?

  • 1 kudos
4 More Replies

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.

If there isn’t a group near you, start one and help create a community that brings people together.

Request a New Group
Labels