cancel
Showing results for 
Search instead for 
Did you mean: 
Get Started Discussions
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and introductory topics. Connect with beginners and experts alike to kickstart your Databricks experience.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

vijaypodili
by New Contributor III
  • 58 Views
  • 2 replies
  • 0 kudos

Azure databrics Learning tutorials ADB+SQL,ADB+PYSPARK,ADB+PYTHON

Suggest me the best learning tutorials of Azure databricks with the combinations of pyspark,python,sql is the any learning web based tutorials from Databricks  suggest me the best one from scratch to advanced

  • 58 Views
  • 2 replies
  • 0 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 0 kudos

Hi @vijaypodili ,I can recommend Data Engineering Learning Path on Databricks Academy:https://customer-academy.databricks.com/On Udemy, there's an excellent course that covers all the important aspects of working with Databricks on a daily basis:Data...

  • 0 kudos
1 More Replies
Rohan_Samariya
by New Contributor III
  • 63 Views
  • 1 replies
  • 2 kudos

Resolved! Building an End-to-End ETL Pipeline with Data from S3 in Databricks

Hey everyone I’m excited to share the progress of my Databricks learning journey! Recently, I worked on building an end-to-end ETL pipeline in Databricks, starting from data extraction from AWS S3 to creating a dynamic dashboard for insights.Here’s h...

  • 63 Views
  • 1 replies
  • 2 kudos
Latest Reply
bianca_unifeye
New Contributor III
  • 2 kudos

@Rohan_Samariya  this is fantastic work! I’m genuinely impressed with how you’ve taken the Databricks stack end-to-end: S3 ingestion → PySpark transformations → Delta optimisation → interactive SQL dashboards. This is exactly the type of hands-on, fu...

  • 2 kudos
CookDataSol
by New Contributor II
  • 88 Views
  • 2 replies
  • 1 kudos

Resolved! SQL cell v spark.sql in notebooks

I am fairly new to Databricks, and indeed Python, so apologies if this has been answered elsewhere but I've been unable to find it.I have been mainly working in notebooks as opposed to the SQL editor, but coding in SQL where possible using SQL cells ...

  • 88 Views
  • 2 replies
  • 1 kudos
Latest Reply
CookDataSol
New Contributor II
  • 1 kudos

Thanks Louis, really good explanation and helpful examples!

  • 1 kudos
1 More Replies
Judith
by New Contributor III
  • 5571 Views
  • 5 replies
  • 3 kudos

Connect to Onelake using Service Principal, Unity Catalog and Databricks Access Connector

We are trying to connect Databricks to OneLake, to read data from a Fabric workspace into Databricks, using a notebook. We also use Unity Catalog. We are able to read data from the workspace with a Service Principal like this:from pyspark.sql.types i...

Judith_0-1739892045239.png Judith_1-1739891020619.png
  • 5571 Views
  • 5 replies
  • 3 kudos
Latest Reply
Coffee77
Contributor III
  • 3 kudos

As commented you need to assign "Storage Blob Data Contributor or Storage Account Contributor to the service principal you're using in the "connection" provided to the "external location". Another more advanced and even better option would be to use ...

  • 3 kudos
4 More Replies
Radix95
by New Contributor II
  • 3526 Views
  • 1 replies
  • 0 kudos

Updating a Delta Table in Delta Live Tables (DLT) from Two Event Hubs

I am working with Databricks Delta Live Tables (DLT) and need to ingest data from two different Event Hubs. My goal is to:Ingest initial data from the first Event Hub (Predictor) and store it in a Delta Table (data_predictions).Later, update this tab...

  • 3526 Views
  • 1 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

To achieve robust, persistent CDC (Change Data Capture)–style updates in Databricks DLT with your scenario—while keeping data_predictions as a Delta Table (not a Materialized View)—you need to carefully avoid streaming joins and side effects across s...

  • 0 kudos
ChristianRRL
by Valued Contributor III
  • 3424 Views
  • 1 replies
  • 0 kudos

Databricks UMF Best Practice

Hi there, I would like to get some feedback on what are the ideal/suggested ways to get UMF data from our Azure cloud into Databricks. For context, UMF can mean either:User Managed FileUser Maintained FileBasically, a UMF could be something like a si...

Get Started Discussions
Data ingestion
UMF
User Maintained File
User Managed File
  • 3424 Views
  • 1 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

Several effective patterns exist for ingesting User Managed Files (UMF) such as CSVs from Azure into Databricks, each with different trade-offs depending on governance, user interface preferences, and integration with Microsoft 365 services. Common A...

  • 0 kudos
Ibrahim1
by New Contributor
  • 3408 Views
  • 1 replies
  • 0 kudos

DLT detecting changes but not applying them

We have three source tables used for a streaming dimension table in silver. Around 50K records are changed in one of the source tables, and the DLT pipeline shows that it has updated those 50K records, but they remain unchanged. The only way to pick ...

  • 3408 Views
  • 1 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

The most likely reason your DLT pipeline shows 50K updates but the records remain unchanged is related to how Delta Live Tables (DLT) handle streaming tables, update logic, and schema constraints. When the target table uses an auto-increment ID (espe...

  • 0 kudos
nachii_rajput
by New Contributor
  • 2695 Views
  • 1 replies
  • 0 kudos

Issue with Disabled "Repair DAG", "Repair All DAGs" Buttons in Airflow UI, functionality is working.

We are encountering an issue in the Airflow UI where the 'Repair DAG' and 'Repair All DAGs' options are disabled when a specific task fails. While the repair functionality itself is working properly (i.e., the DAGs can still be repaired through execu...

  • 2695 Views
  • 1 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

The issue with the 'Repair DAG' and 'Repair All DAGs' options being disabled in the Airflow UI when using the Databricks Workflow Operator is a known UI-specific problem that does not affect backend execution or the actual repair functionality. While...

  • 0 kudos
Sudheer2
by New Contributor III
  • 982 Views
  • 1 replies
  • 1 kudos

How to Fetch Azure OpenAI api_version and engine Dynamically After Resource Creation via Python?

Hello,I am using Python to automate the creation of Azure OpenAI resources via the Azure Management API. I am successfully able to create the resource, but I need to dynamically fetch the following details after the resource is created:API Version (a...

  • 982 Views
  • 1 replies
  • 1 kudos
Latest Reply
jamesl
Databricks Employee
  • 1 kudos

Hi Sudheer,  It's been a while since you posted, but are you still facing this issue? Here are a few things you could check if needed:  API version In Azure OpenAI, api-version is a query parameter on the data-plane (inference) requests, not a proper...

  • 1 kudos
jadhav_vikas
by New Contributor II
  • 238 Views
  • 5 replies
  • 2 kudos

#data bricks snowflake dialect

Hello,I’m encountering an issue while converting SQL code to the Lake bridge Snowflake dialect. It seems that DML and DDL statements may not be supported in the Snowflake dialect within Lake bridge.Could you please confirm whether DML and DDL stateme...

  • 238 Views
  • 5 replies
  • 2 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 2 kudos

@jadhav_vikas , I did some digging through internal docs and I have some hints/suggestions. Short answer Databricks Lakehouse Federation (often referred to as “Lakehouse Bridge”) provides read‑only access to Snowflake; DML and DDL are not supported ...

  • 2 kudos
4 More Replies
zibi
by New Contributor II
  • 151 Views
  • 3 replies
  • 0 kudos

Resolved! API call fails to initiate create Service Principal secret

Hi,I've constructed an AWS lambda function which is used to auto rotate my Service Principal Secret in the Databricks account. Authentication is setup with OAuth2, the api call for the token generation is successful but when executing the api call to...

  • 151 Views
  • 3 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

Your error message, "Invalid service principal id," typically indicates a mismatch or formatting problem with the service principal's unique identifier in your API request. Although you checked the client_id, this value is not always the one needed f...

  • 0 kudos
2 More Replies
spd_dat
by New Contributor III
  • 3616 Views
  • 2 replies
  • 0 kudos

Can AWS workspaces share subnets?

The docs state:"You can choose to share one subnet across multiple workspaces or both subnets across workspaces."as well as:"You can reuse existing security groups rather than create new ones."and on this page:"If you plan to share a VPC and subnets ...

  • 3616 Views
  • 2 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

AWS WorkSpaces can be configured with subnets that can be shared within an AWS account or across AWS accounts using resource sharing mechanisms, but this depends on the specific AWS service and context. For Databricks workspaces on AWS, documentation...

  • 0 kudos
1 More Replies
mancosta
by New Contributor
  • 2888 Views
  • 1 replies
  • 0 kudos

Joblib with optuna and SB3

Hi everyone,I am training some reinforcement learning models and I am trying to automate the hyperparameter search using optuna. I saw in the documentation that you can use joblib with spark as a backend to train in paralel. I got that working with t...

  • 2888 Views
  • 1 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

Stable Baselines 3 (SB3) models can be optimized with Optuna for hyperparameter search, but parallelizing these searches using Joblib with Spark as the backend—like the classic scikit-learn example—commonly encounters issues. The root problem is that...

  • 0 kudos
December
by New Contributor II
  • 3066 Views
  • 1 replies
  • 0 kudos

NiFi on EKS Fails to Connect to Databricks via JDBC – "Connection reset" Error

I'm using Apache Nifi (running on AWS EKS) to connect to Databricks (with compute on EC2) via JDBC. My JDBC URL is as follows: jdbc:databricks://server_hostname:443/default;transportMode=http;ssl=1;httpPath=my_httppath;AuthMech=3;UID=token;PWD=my_tok...

December_1-1741684901117.png
Get Started Discussions
Connection
JDBC
  • 3066 Views
  • 1 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

A "Connection reset" error in NiFi when connecting to Databricks via JDBC, despite successful telnet and working connectivity from DBeaver, usually points to subtle protocol or compatibility issues rather than network-level blocks.​ Common Causes JD...

  • 0 kudos

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels