cancel
Showing results for 
Search instead for 
Did you mean: 
Community Articles
Dive into a collaborative space where members like YOU can exchange knowledge, tips, and best practices. Join the conversation today and unlock a wealth of collective wisdom to enhance your experience and drive success.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

ayushbadhera1
by New Contributor III
  • 948 Views
  • 2 replies
  • 6 kudos

Databricks LLM Evolution and Future Prospects

Databricks LLM Evolution and Future ProspectsDatabricks has progressed from a big-data compute engine to a full-stack AI powerhouse that designs, trains, and serves state‐of‐the-art large language models (LLMs). This article explores two key technica...

  • 948 Views
  • 2 replies
  • 6 kudos
Latest Reply
ayushbadhera1
New Contributor III
  • 6 kudos

Thanks, @RiyazAliM, for checking out the blog post!More insights on Databricks LLM and Dolly are on the way in the next one. Stay tuned and keep learning!Best,Ayush

  • 6 kudos
1 More Replies
TejeshS
by Contributor
  • 1506 Views
  • 2 replies
  • 7 kudos

Building a Metadata Table-Driven Framework Using LakeFlow Declarative (Formerly DLT) Pipelines

IntroductionScaling data pipelines across an organization can be challenging, particularly when data sources, requirements, and transformation rules are always changing. A metadata table-driven framework using LakeFlow Declarative (Formerly DLT) enab...

TejeshS_0-1753177497326.png
  • 1506 Views
  • 2 replies
  • 7 kudos
Latest Reply
sridharplv
Valued Contributor II
  • 7 kudos

Good one Tejesh. Quick intro on DLT meta.

  • 7 kudos
1 More Replies
hozefa413
by New Contributor III
  • 2237 Views
  • 3 replies
  • 9 kudos

Modernizing Legacy Data Platforms to Lakehouse for AI-Readiness

As organizations increasingly migrate from legacy platforms—like on-prem SQL Server, Oracle Exadata, Teradata, Informatica, Cloudera, or Netezza—to modern cloud architectures, one critical question often arises:"Are we just lifting and shifting the s...

  • 2237 Views
  • 3 replies
  • 9 kudos
Latest Reply
sridharplv
Valued Contributor II
  • 9 kudos

Great article @hozefa413 , It shows all your expertise and delivery excellence

  • 9 kudos
2 More Replies
gdschld
by New Contributor
  • 1103 Views
  • 2 replies
  • 3 kudos

Establishing Trust relationship for Databricks on AWS

Hello.Our databricks is on Azure. We are trying to connect with AWS S3 as an external source from Unity Catalog.We have followed all steps given here, is there anything additional required?https://docs.databricks.com/aws/en/connect/unity-catalog/clou...

  • 1103 Views
  • 2 replies
  • 3 kudos
Latest Reply
Pat
Esteemed Contributor
  • 3 kudos

Hi @gdschld ,what ID have you used here:"sts:ExternalId": "<STORAGE-CREDENTIAL-EXTERNAL-ID>"I haven't done this for some time and got a bit confused with this STORAGE-CREDENTIAL-EXTERNAL_ID. I used to put there Databricks Account ID.I found this, it ...

  • 3 kudos
1 More Replies
DouglasMoore
by Databricks Employee
  • 5663 Views
  • 2 replies
  • 1 kudos

How to enable unity catalog system tables?

Unity Catalog system tables provide lots of metadata & log data related to the operations of Databricks. System tables are organized into separate schemas containing one to a few tables owned and updated by Databricks. The storage and the cost of the...

  • 5663 Views
  • 2 replies
  • 1 kudos
Latest Reply
AlanD
New Contributor III
  • 1 kudos

It's in the Databricks CLI Unity Catalog section Databricks CLI commands | Databricks DocumentationmetastoresCommands to manage metastores, which are the top-level container of objects in Unity Catalog:assign, create, current, delete, get, list, summ...

  • 1 kudos
1 More Replies
Pat
by Esteemed Contributor
  • 935 Views
  • 1 replies
  • 3 kudos

Building DLT Pipelines with Databricks Free Edition and Amazon Q Developer

How AI-powered development accelerated my data engineering workflow Watch the Complete Development Process YouTube Video: See the entire 30-minute development sessionThis is a screen recording without voice narration showing the complete development ...

  • 935 Views
  • 1 replies
  • 3 kudos
Latest Reply
Advika
Databricks Employee
  • 3 kudos

This is super insightful @Pat, thanks for sharing this with the Community!

  • 3 kudos
smpa01
by Contributor
  • 815 Views
  • 2 replies
  • 0 kudos

Resolved! Databricks VS code extension to add cell title

I use the databricks extension in vs code for all my work. Is there any way for me to add a cell title from the extension itself?. There is no point in adding in the server version of this notebook cause when I sync the local to sever, it will overwr...

smpa01_0-1750343946018.png smpa01_1-1750344072719.png
  • 815 Views
  • 2 replies
  • 0 kudos
Latest Reply
smpa01
Contributor
  • 0 kudos

One needs to use # DBTITLE 1,cell_title in a py file # COMMAND ---------- # DBTITLE 1,Title 1 from pyspark.sql import SparkSession from delta.tables import DeltaTable from pyspark.sql.functions import *  

  • 0 kudos
1 More Replies
RiyazAliM
by Honored Contributor
  • 777 Views
  • 1 replies
  • 4 kudos

The Databricks Python SDK

The Databricks SDK is a script (written in Python, in our case) which lets you control and automate actions on Databricks using the methods available in the WorkSpaceClient (more about this below).Why do we need Databricks SDK:- Automation: You can d...

aayrm5_1-1752841018665.png aayrm5_3-1752841258075.png aayrm5_4-1752841851794.png
  • 777 Views
  • 1 replies
  • 4 kudos
Latest Reply
sridharplv
Valued Contributor II
  • 4 kudos

Good Article @RiyazAliM.

  • 4 kudos
ilir_nuredini
by Honored Contributor
  • 2048 Views
  • 2 replies
  • 4 kudos

Apache 4.0

Missed the Apache Spark 4.0 release? It is not just a version bump, it is a whole new level for big data processing. Some of the highlights that really stood out to me:1. SQL just got way more powerful: reusable UDFs, scripting, session variables, an...

apache-4-0.jpg
  • 2048 Views
  • 2 replies
  • 4 kudos
Latest Reply
Advika
Databricks Employee
  • 4 kudos

Yeah, Spark 4.0 brings powerful enhancements while staying compatible with existing workloads.Thank you for putting this together and highlighting the key updates, @ilir_nuredini.

  • 4 kudos
1 More Replies
Harun
by Honored Contributor
  • 7874 Views
  • 3 replies
  • 2 kudos

Optimizing Costs in Databricks by Dynamically Choosing Cluster Sizes

Databricks is a popular unified data analytics platform known for its powerful data processing capabilities and seamless integration with Apache Spark. However, managing and optimizing costs in Databricks can be challenging, especially when it comes ...

  • 7874 Views
  • 3 replies
  • 2 kudos
Latest Reply
kmacgregor
New Contributor II
  • 2 kudos

How can this actually be used to choose a cluster pool for a Databricks workflow dynamically, that is, at run time? In other words, what can you actually do with the value of `selected_pool` other than printing it out?

  • 2 kudos
2 More Replies
OU_Professor
by New Contributor II
  • 1702 Views
  • 2 replies
  • 1 kudos

Resolved! Community (Legacy) Edition Question

Hello,I have been teaching my Data Warehousing class using the Databricks Community Edition.  With the change to the Databricks Free Edition, there are many aspects of my Community Edition notebooks that no longer work in the free Edition.  Is there ...

  • 1702 Views
  • 2 replies
  • 1 kudos
Latest Reply
Advika
Databricks Employee
  • 1 kudos

Hello @OU_Professor! Access to existing Community Edition accounts will remain available for the rest of the year. However, please note that new users attempting to sign up for Community Edition are now redirected to the Free Edition instead.

  • 1 kudos
1 More Replies
nathanielcooley
by New Contributor II
  • 2816 Views
  • 4 replies
  • 0 kudos

Data Modeling

Just got out of a session on Data Modeling using the Data Vault paradigm. Highly recommended to help think through complex data design. Look out for Data Modeling 101 for Data Lakehouse Demystified by Luan Medeiros. 

  • 2816 Views
  • 4 replies
  • 0 kudos
Latest Reply
sridharplv
Valued Contributor II
  • 0 kudos

Hi @BS_THE_ANALYST , please use this link with code for reference :https://www.databricks.com/blog/data-vault-best-practice-implementation-lakehouse

  • 0 kudos
3 More Replies
ilir_nuredini
by Honored Contributor
  • 920 Views
  • 0 replies
  • 1 kudos

Databricks Asset Bundles

Why Should You Use Databricks Asset Bundles (DABs)?Without proper tooling, Data Engineering and Machine Learning projects can quickly become messy.That is why we recommend leveraging DABs to solve these common challenges:1. Collaboration:Without stru...

dabs.jpg
  • 920 Views
  • 0 replies
  • 1 kudos
SumitSingh
by Contributor II
  • 7692 Views
  • 11 replies
  • 41 kudos

From Associate to Professional: My Learning Plan to ace all Databricks Data Engineer Certifications

In today’s data-driven world, the role of a data engineer is critical in designing and maintaining the infrastructure that allows for the efficient collection, storage, and analysis of large volumes of data. Databricks certifications holds significan...

SumitSingh_0-1721402402230.png SumitSingh_1-1721402448677.png SumitSingh_2-1721402469214.png
  • 7692 Views
  • 11 replies
  • 41 kudos
Latest Reply
BS_THE_ANALYST
Esteemed Contributor II
  • 41 kudos

@SumitSingh this is getting put in the favourites. Thanks a bunch for this All the best,BS

  • 41 kudos
10 More Replies

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels