Certifications
Join dynamic discussions on Databricks certifications within the Community. Exchange insights, tips,...
The forum board you are trying to access is permanently deleted.
Join dynamic discussions on Databricks certifications within the Community. Exchange insights, tips,...
Explore discussions on Databricks training programs and offerings within the Community. Get insights...
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and ...
Engage in discussions about the Databricks Free Edition within the Databricks Community. Share insig...
Ingestion from SharePoint is now available directly in PySpark. Just define a connection and use spark-read or, even better, spark-readStream with an autoloader. Just specify the file type and options for that file (pdf, csv, Excel, etc.)
Excel The big news this week is the possibility of native importing Excel files. Write operations are also possible. There is a possibility of choosing a data range. It also works with the streaming autoloader, currently in beta. GPT 5.2 The same day...
HiI am planning to take Databricks Certified Data Engineer Professional Exam. The fee is highly unaffordable for me. I am looking for discount vouchers. Please let me know if there is a way to get discount on the exam.Thank you!RegardsKrishna
@saikrishna08 Check out the upcoming January Learning Festival for 50% off vouchers - https://community.databricks.com/t5/events/self-paced-learning-festival-09-january-30-january-2026/ec-p/141503#M5768Thanks!
Hi Guys,Has anyone sat the Databricks Certified Data Analyst Associate exam so far? I’m planning to take this exam next month and would really appreciate it if you could share your experience.How was the difficulty level overall? Were the questions m...
Hi @Sadie_james,I have cleared the DA Ass. Exam a month ago. Would suggest the only guide you need is already provided by Databricks - https://www.databricks.com/learn/certification/data-analyst-associateDo take the related Data Analysis with Databri...
ZeroBus changes the game: you can now push event data directly into Databricks, even from on-prem. No extra event layer needed. Every Unity Catalog table can act as an endpoint.
Hi, all. I'm getting stuck on the second step of the "Get Started with Databricks Free Edition" course. It's obviously instructing me to download a zip file from a GitHub repository, but there is no link provided for the repository. What am I missing...
All leading LLMs are available natively in Databricks: - ChatGPT 5.2 from the day of the premiere! - System catalog with AI schema in Unity Catalog has multiple LLMs ready to serve! - OpenAI, Gemini, and Anthropic are available side by side!
I’m looking to gather insights from data engineers, architects, and developers who have experience building scalable pipelines in Databricks. Specifically, I want to understand how to design, implement, and manage reusable data engineering components...
The best strategy is to build modular, parameterized, Delta-optimized functions and package them into reusable Python modules, while keeping Databricks notebooks only for orchestration. This creates consistent, scalable, and easily maintainable data ...
Hi everyone!I’m working on getting Genie to accurately generate Statistical Process Control (SPC) charts when prompted. I'm looking for suggestions on how to best approach this. So far, I’ve tried using pre-defined SQL queries to select the data, bu...
Or here is hopefully a more elegant way to phrase my question:To visualise a control diagram in Genie for an end-user, should I a) instruct Genie how to create an SPC chart with SQL on the fly, of b) create a background job (pre-defined SQL query in ...
Real-time mode is a breakthrough that lets Spark utilize all available CPUs to process records with single-millisecond latency, while decoupling checkpointing from per-record processing.
Hello - I am following some online code to create a function as follows:-----------------------------------------CREATE OR REPLACE FUNCTION my_catalog.my_schema.insert_data_function(col1_value STRING,col2_value INT)RETURNS BOOLEANCOMMENT 'Inserts dat...
In UC, the functions must be read-only; they cannot modify state (no INSERT, DELETE, MERGE, CREATE, VACUUM, etc). So I tried to create a PROCEDURE and call it; I was able to insert data into the table successfully. Unity Catalog tools are really jus...
Databricks is positioning the platform as a full stack for LLM development — from data ingestion → feature/embedding pipelines → fine-tuning (Mosaic AI) → evaluation → deployment (Model Serving) → monitoring (Lakehouse Monitoring).I’m curious about r...
Thank You @Gecofer for taking the time to share such a clear, experience-backed breakdown of where Databricks shines and where real-world LLM Ops architectures still need supporting components. Your explanation was incredibly practical and resonates ...
I'm a software engineer and a bit new to databricks. My goal is to create a model serving endpoint, that interfaces with several ML models. Traditionally this would look like:API--> Service --> DataNow using databricks, my understanding is that it w...
Just register model and then deploy service endpoint to serve this model.
Hi Databricks,our learners wants to know when are we disclosing the results of october badge challenge - learningsBr
Hello @saurabh18cs! Have you checked with your company's Databricks Partner Admin or Databricks Support? They’re the right contacts to provide the official timeline for the Partner Learning Badge Challenge results.
Databricks goes native on Excel. You can now ingest + query .xls/.xlsx directly in Databricks (SQL + PySpark, batch and streaming), with auto schema/type inference, sheet + cell-range targeting, and evaluated formulas, no extra libraries anymore.
| User | Count |
|---|---|
| 212 | |
| 193 | |
| 94 | |
| 81 | |
| 76 |