cancel
Showing results for 
Search instead for 
Did you mean: 
cancel
Showing results for 
Search instead for 
Did you mean: 
BrickTalks: Serve intelligence from your Lakehouse to your Apps with Lakebase

Join us next Thursday with Product Manager Pranav Aurora for a another high-impact virtual session: how to bring the data intelligence from your Lakehouse to all your apps and users.  What we’ll cover: Using Lakebase (Lakebase Postgres is a fully man...

  • 423 Views
  • 2 replies
  • 4 kudos
a week ago
Free Edition Hackathon

We are running a Free Edition Hackathon from November 5-14, 2025 and would love for you to participate and/or help promote it to your networks. Leverage Free Edition for a project and record a five-minute demo showcasing your work. Free Edition launc...

  • 608 Views
  • 0 replies
  • 6 kudos
a week ago
Big Book of Data Engineering - Get how-tos, code snippets and real-world examples

As data volume and complexity increase, engineers are left figuring out how to manage, monitor and maintain fragile pipelines while also handling fragmented tools. The Big Book of Data Engineering equips you with cutting-edge methods for building pip...

  • 67 Views
  • 0 replies
  • 1 kudos
yesterday
Level Up with Databricks Specialist Sessions

How to Register & Prepare If you're interested in advancing your skills with Databricks through a Specialist Session, here's a clear guide on how to register and what free courses you can take to prepare effectively. How to Begin Your Learning Path S...

  • 1920 Views
  • 0 replies
  • 7 kudos
10-02-2025
🌟 Community Pulse: Your Weekly Roundup! October 31 – November 06, 2025

We’re stepping into November, and while 2025 is winding down, the Databricks Community is only turning up the heat — with fresh blogs, dynamic discussions, and events that kept the Databricks energy alive! Let’s dive into this week’s highlights To...

  • 390 Views
  • 5 replies
  • 7 kudos
Friday

Community Activity

kahrees
by > New Contributor
  • 59 Views
  • 2 replies
  • 2 kudos

DATA_SOURCE_NOT_FOUND Error with MongoDB (Suggestions in other similar posts have not worked)

I am trying to load data from MongoDB into Spark. I am using the Community/Free version of DataBricks so my Jupiter Notebook is in a Chrome browser.Here is my code:from pyspark.sql import SparkSession spark = SparkSession.builder \ .config("spar...

  • 59 Views
  • 2 replies
  • 2 kudos
Latest Reply
K_Anudeep
Databricks Employee
  • 2 kudos

Hey @kahrees , Good Day! I tested this internally, and I was able to reproduce the issue. Screenshot below:   You’re getting [DATA_SOURCE_NOT_FOUND] ... mongodb because the MongoDB Spark connector jar isn’t actually on your cluster’s classpath. On D...

  • 2 kudos
1 More Replies
Louis_Frolio
by Databricks Employee
  • 1 Views
  • 0 replies
  • 0 kudos

Why Agent Bricks Is a Game Changer

Check out Agent Bricks — Databricks’ new framework for building AI agents without writing a single line of code. Picture this: you describe your business challenge in plain English, point it at your data, and boom… a domain-specific agent comes to li...

  • 1 Views
  • 0 replies
  • 0 kudos
eyalholzmann
by > New Contributor
  • 81 Views
  • 3 replies
  • 1 kudos

Does VACUUM on Delta Lake also clean Iceberg metadata when using Iceberg Uniform feature?

I'm working with Delta tables using the Iceberg Uniform feature to enable Iceberg-compatible reads. I’m trying to understand how metadata cleanup works in this setup.Specifically, does the VACUUM operation—which removes old Delta Lake metadata based ...

  • 81 Views
  • 3 replies
  • 1 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 1 kudos

Here’s how to approach cleaning and maintaining Apache Iceberg metadata on Databricks, and how it differs from Delta workflows. First, know your table type For Unity Catalog–managed Iceberg tables, Databricks runs table maintenance for you (predicti...

  • 1 kudos
2 More Replies
ashfire
by > New Contributor II
  • 11 Views
  • 1 replies
  • 0 kudos

How to store & update a FAISS Index in Databricks

I’m currently using FAISS in a Databricks notebook to perform semantic search in text data. My current workflow looks like this:encode ~10k text entries using an embedding model.build a FAISS index in memory.run similarity searches using index.search...

  • 11 Views
  • 1 replies
  • 0 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 0 kudos

Hello @ashfire , Here’s a practical path to scale your FAISS workflow on Databricks, along with patterns to persist indexes, incrementally add embeddings, and keep metadata aligned. Best practice to persist/load FAISS indexes on Databricks Use faiss...

  • 0 kudos
pooja_bhumandla
by > New Contributor III
  • 17 Views
  • 1 replies
  • 0 kudos

Should I enable Liquid Clustering based on table size distribution?

Hi everyone,I’m evaluating whether Liquid Clustering would be beneficial for the tables based on the sizes. Below is the size distribution of tables in my environment:Size Bucket Table Count Large (> 1 TB)3Medium (10 GB – 1 TB)284Small (< 10 GB)17,26...

  • 17 Views
  • 1 replies
  • 0 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 0 kudos

Greetings @pooja_bhumandla  Based on your size distribution, enabling Liquid Clustering can provide meaningful gains—but you’ll get the highest ROI by prioritizing your medium and large tables first and selectively applying it to small tables where q...

  • 0 kudos
Naveenkumar1811
by > New Contributor
  • 10 Views
  • 1 replies
  • 0 kudos

Can we Change the ownership of Databricks Managed Secret to SP in Azure Data Bricks?

Hi Team,Earlier we faced an Issue where the jar file(Created by a old employee) in workspace directory is used as library in the cluster which is run from a SP. Since the employee left the org and the id got removed even though the SP is part of ADMI...

  • 10 Views
  • 1 replies
  • 0 kudos
Latest Reply
Coffee77
Contributor III
  • 0 kudos

That's the reason by which I try to deploy most part of resources with service principal accounts while using Databricks Asset Bundles. Avoid human identities whenever possible because they can indeed go away...I think you'll have to create another s...

  • 0 kudos
zach_goehring
by Databricks Employee
  • 28 Views
  • 0 replies
  • 0 kudos

Operationalize Your Lakehouse: Lakebase for Low-Latency Apps & APIs

Operationalize Your Lakehouse: Lakebase for Low-Latency Apps & APIs The Databricks Data Intelligence Platform unifies data, AI, and governance so organizations can put all of their data to work. Until recently, though, operational workloads still liv...

zach_goehring_0-1762961280978.png blog thumbnail.png
  • 28 Views
  • 0 replies
  • 0 kudos
maikel
by > New Contributor II
  • 91 Views
  • 3 replies
  • 0 kudos

Agent outside databricks communication with databricks MCP server

Hello Community!I have a following use case in my project:User -> AI agent -> MCP Server -> Databricks data from unity catalog.- AI agent is not created in the databricks- MCP server is created in the databricks and should expose tools to get data fr...

  • 91 Views
  • 3 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

Hopefully this helps... You can securely connect your external AI agent to a Model Context Protocol (MCP) server and Unity Catalog while maintaining strong control over authentication and resource management. The method depends on whether MCP is outs...

  • 0 kudos
2 More Replies
bidek56
by > Contributor
  • 185 Views
  • 5 replies
  • 1 kudos

Resolved! Location of spark.scheduler.allocation.file

In DBR 164.LTS, I am trying to add the following Spark config: spark.scheduler.allocation.file: file:/Workspace/init/fairscheduler.xmlBut the all purpose cluster is throwing this error Spark error: Driver down cause: com.databricks.backend.daemon.dri...

  • 185 Views
  • 5 replies
  • 1 kudos
Latest Reply
mark_ott
Databricks Employee
  • 1 kudos

Here's some solutions without using DBFS..  Yes, there are solutions for using the Spark scheduler allocation file on Databricks without DBFS, but options are limited and depend on your environment and access controls. Alternatives to DBFS for Schedu...

  • 1 kudos
4 More Replies
Sam11
by > Visitor
  • 7 Views
  • 0 replies
  • 0 kudos

Didn't Receive Databricks Gen AI Certificate

Hi Team,I've successfully passed the Databricks Gen AI Engineer Associate Certified exam, but I still have not received the certificate.Databricks Certified Data Engineer Associate on 9th November 2025.Result: PASSCongratulations! You've passed the e...

  • 7 Views
  • 0 replies
  • 0 kudos
Bhanu_Anandesi
by > New Contributor III
  • 2487 Views
  • 4 replies
  • 1 kudos

Resolved! Not received my certificate after passing Data Engineer Associate exam.

I've successfully passed Databricks Data Engineer Associate Certified exam but still have not received the certificate.Hello Bhanupriya,You have completed the assessment, Databricks Certified Data Engineer Associate on 31 July 2024.Result: PASS Congr...

  • 2487 Views
  • 4 replies
  • 1 kudos
Latest Reply
Sam11
Visitor
  • 1 kudos

Hi,I would like to know if this issue was resolved asap. 

  • 1 kudos
3 More Replies
zibi
by > New Contributor
  • 84 Views
  • 3 replies
  • 0 kudos

Resolved! API call fails to initiate create Service Principal secret

Hi,I've constructed an AWS lambda function which is used to auto rotate my Service Principal Secret in the Databricks account. Authentication is setup with OAuth2, the api call for the token generation is successful but when executing the api call to...

  • 84 Views
  • 3 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

Your error message, "Invalid service principal id," typically indicates a mismatch or formatting problem with the service principal's unique identifier in your API request. Although you checked the client_id, this value is not always the one needed f...

  • 0 kudos
2 More Replies
Yuki
by > Contributor
  • 34 Views
  • 4 replies
  • 1 kudos

Is there any way to run jobs from github actions and catch the results?

Hi all,Is there any way to run jobs from github actions and catch the results?Of course, I can do this if I use the API or CLI.But I found the actions for notebook: https://github.com/marketplace/actions/run-databricks-notebook  Compared to this, wri...

  • 34 Views
  • 4 replies
  • 1 kudos
Latest Reply
Yuki
Contributor
  • 1 kudos

OK, thank you for your advices, I will consider to use asset bundles for this.

  • 1 kudos
3 More Replies
Naveenkumar1811
by > New Contributor
  • 47 Views
  • 2 replies
  • 0 kudos

What is the Best Practice of Maintaining the Delta table loaded in Streaming?

Hi Team,We have our Bronze(append) Silver(append) and Gold(merge) Tables loaded using spark streaming continuously with trigger as processing time(3 secs).We Also Run our Maintenance Job on the Table like OPTIMIZE,VACCUM and we perform DELETE for som...

  • 47 Views
  • 2 replies
  • 0 kudos
Latest Reply
Naveenkumar1811
New Contributor
  • 0 kudos

Hi Mark,But the real problem is our streaming job runs 365 days 24 *7 and we cant afford any further latency to our data flowing to gold layer. We don't have any window to pause or slower our streaming and we continuously get the data feed actually s...

  • 0 kudos
1 More Replies
rathorer
by > New Contributor III
  • 8140 Views
  • 4 replies
  • 7 kudos

API Consumption on Databricks

In this blog, I will be talking about the building the architecture to serve the API consumption on Databricks Platform. I will be using Lakebase approach for this. It will be useful for this kind of API requirement.API Requirement: Performance:Curre...

rathorer_0-1759990971452.png rathorer_0-1760078458113.png rathorer_1-1760078513459.png
  • 8140 Views
  • 4 replies
  • 7 kudos
Latest Reply
SurajW
Visitor
  • 7 kudos

Insightful @rathorer 

  • 7 kudos
3 More Replies
Welcome to the Databricks Community!

Once you are logged in, you will be ready to post content, ask questions, participate in discussions, earn badges and more.

Spend a few minutes exploring Get Started Resources, Learning Paths, Certifications, and Platform Discussions.

Connect with peers through User Groups and stay updated by subscribing to Events. We are excited to see you engage!

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Top Kudoed Authors
Read Databricks Data Intelligence Platform reviews on G2

Latest from our Blog