- 129 Views
- 1 replies
- 3 kudos
Hadoop Walked So Databricks Could Run
Are you familiar with this scenario: Your data team spends 80% of their time fixing infrastructure issues instead of extracting insights.In today’s data-driven world, organisations are drowning in data but starving for actionable insights. Traditiona...
- 129 Views
- 1 replies
- 3 kudos
- 391 Views
- 2 replies
- 5 kudos
API Consumption on Databricks
In this blog, I will be talking about the building the architecture to serve the API consumption on Databricks Platform. I will be using Lakebase approach for this. It will be useful for this kind of API requirement.API Requirement: Performance:Curre...
- 391 Views
- 2 replies
- 5 kudos
- 5 kudos
Sharp design choices, @rathorer! Appreciate you sharing this detailed architecture.
- 5 kudos
- 118 Views
- 0 replies
- 1 kudos
Validating pointer-based Delta comparison architecture using flatMapGroupsWithState in Structured St
Hi everyone,I’m leading an implementation where we’re comparing events from two real-time streams — a Source and a Target — in Databricks Structured Streaming (Scala).Our goal is to identify and emit “delta” differences between corresponding records ...
- 118 Views
- 0 replies
- 1 kudos
- 10828 Views
- 6 replies
- 3 kudos
Editing value of widget parameter within notebook code
I have a notebook with a text widget where I want to be able to edit the value of the widget within the notebook and then reference it in SQL code. For example, assuming there is a text widget named Var1 that has input value "Hello", I would want to ...
- 10828 Views
- 6 replies
- 3 kudos
- 3 kudos
It seems that only way to use parameters in sql code block is to use dbutils.widget and you cannot change those parameters without removing widget and setting it up again in code
- 3 kudos
- 2088 Views
- 6 replies
- 9 kudos
Attribute-Based Access Control (ABAC) in Databricks Unity Catalog
What Is ABAC and Why Does It Matter?Attribute-Based Access Control (ABAC) is a data governance model now available in Databricks, designed to offer fine-grained, dynamic, and scalable access control for data, AI assets, and files managed through Data...
- 2088 Views
- 6 replies
- 9 kudos
- 9 kudos
Is there already support in Beta for Volumes?
- 9 kudos
- 175 Views
- 0 replies
- 1 kudos
Databricks Policies and Bundles Inheritance: Let Policies Rule Your DABS
Just the policy_id can specify the entire cluster configuration. Yes, we can inherit default and fixed values from policies. Updating runtime version for 100s of jobs, for example, is much easier this way. Read more: - https://databrickster.med...
- 175 Views
- 0 replies
- 1 kudos
- 411 Views
- 1 replies
- 2 kudos
Relationship in databricks Genie
Now you can define relations also directly in Genie. It includes options like “Many to One”, “One to Many”, “One to One”, “Many to Many”. Read more: - https://databrickster.medium.com/relationship-in-databricks-genie-f8bf59a9b578 - https://www.su...
- 411 Views
- 1 replies
- 2 kudos
- 375 Views
- 1 replies
- 3 kudos
Real-Time Mode in Apache Spark Structured Streaming
Real-Time Mode in Spark StreamingApache Spark™ Structured Streaming has been the backbone of mission-critical pipelines for years — from ETL to near real-time analytics and machine learning.Now, Databricks has introduced something game-changing: Real...
- 375 Views
- 1 replies
- 3 kudos
- 3 kudos
And now in Public Preview! Thank you for writing this up, @Yogesh_Verma_.
- 3 kudos
- 1464 Views
- 16 replies
- 26 kudos
(Episode 1: Getting Data In) - Learning Databricks one brick at a time, using the Free Edition
Episode 1: Getting Data InLearning Databricks one brick at a time, using the Free Edition.Project IntroWelcome to everyone reading. My name’s Ben, a.k.a BS_THE_ANALYST, and I’m going to share my experiences as I learn the world of Databricks. My obje...
- 1464 Views
- 16 replies
- 26 kudos
- 26 kudos
Thank you, @BS_THE_ANALYST , for sharing this. I didn’t have much time to read it last week since I was preparing for the Databricks Professional exam, but today I finally had the chance, and I have to say - it’s a great article.I really appreciate t...
- 26 kudos
- 389 Views
- 2 replies
- 3 kudos
(Teaser CHALLENGE - Community Data Pull) - Upcoming Challenge For the Community
Hey everybody, I've been dying to share this with the community. Over the last few weeks, I've been thinking about how I can do a Data Pull from the Community to highlight some of the cool stuff we all do! 拾. Below is a snippet of visual from the Dat...
- 389 Views
- 2 replies
- 3 kudos
- 3 kudos
I'll aim to have the data for the challenge sorted and ready for next week! . I want to strip out some of the columns and figure out where is best to host the data . Potentially I could have it on the Databricks Marketplace or Github.All the best,BS
- 3 kudos
- 676 Views
- 1 replies
- 2 kudos
9 Powerful 🚀 Spark Optimization Techniques in Databricks (With Real Examples)
IntroductionOne of our ETL pipelines used to take 10 hours to complete. After tuning and scaling in Databricks, it finished in just about 1 hour — a 90% reduction in runtime.That’s the power of Spark tuning.Databricks, built on Apache Spark, is a po...
- 676 Views
- 1 replies
- 2 kudos
- 2 kudos
This is a fantastic breakdown of Spark optimization techniques, @savlahanish27!Definitely helpful for anyone working on performance tuning in Databricks.
- 2 kudos
- 445 Views
- 2 replies
- 12 kudos
(Episode 2: Reading Excel Files) - Learning Databricks one brick at a time, using the Free Edition
Episode 2: Reading Excel FilesLearning Databricks one brick at a time, using the Free Edition.You can download the accompanying Notebook and Excel files used in the demonstration over on my GitHub:Excel Files & Notebook: https://github.com/BSanalyst...
- 445 Views
- 2 replies
- 12 kudos
- 12 kudos
@BS_THE_ANALYST that final snippet of code looks very clean! I saw that "sheet name = None" part and was a bit confused why you'd written that as I assumed that was just the default. Turns out the default is "sheet name = 0", which is simply the firs...
- 12 kudos
- 343 Views
- 0 replies
- 4 kudos
Semantic Modelling and Data Products using Agent
Hey Community!We have built something cool for Data Engineers in Databricks!Raw Files -> Semantic Model -> Data Products without writing ETL/ELT Code.Demo/ Guide - https://youtu.be/wjQYXrBwA-oNotebook - https://github.com/Intugle/data-tools/blob/main...
- 343 Views
- 0 replies
- 4 kudos
- 556 Views
- 2 replies
- 7 kudos
Data Quality at Scale: My Experience Using Databricks and AWS
Over the past few years working as a data engineer, I’ve seen how quickly companies are moving their platforms to Databricks and AWS. The flexibility and scale these platforms provide are amazing, but one challenge always comes up again and again: ho...
- 556 Views
- 2 replies
- 7 kudos
- 7 kudos
Hi @Brahmareddy very good insights , I can summarize this as follows: Area Best Practice ExampleSchema ManagementDefine schemas in JSON/YAML, enforce with Delta LakeGovernanceUse Unity Catalog for access, lineage, and ownershipMonitoringSet up Lakeho...
- 7 kudos
- 4345 Views
- 4 replies
- 6 kudos
From Pilots to Production: Unlocking Enterprise AI Agents with Agent Bricks
Many enterprises today launch AI agents with high hopes but more often than not, those pilots never reach production. The culprit? Complexity, poor evaluations, ballooning costs, and governance gaps.Why do so many AI agent pilots never make it to pro...
- 4345 Views
- 4 replies
- 6 kudos
- 6 kudos
Agent Bricks looks solid for scaling AI, and I’ve seen platforms like Agentra.io also tackle the enterprise workflow side of this challenge.
- 6 kudos
Join Us as a Local Community Builder!
Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!
Sign Up Now-
Access Data
1 -
ADF Linked Service
1 -
ADF Pipeline
1 -
Advanced Data Engineering
3 -
AI Agents
1 -
AI Readiness
1 -
ApacheSpark
1 -
Associate Certification
1 -
Automation
1 -
AWSDatabricksCluster
1 -
Azure
1 -
Azure databricks
3 -
Azure devops integration
1 -
AzureDatabricks
2 -
Big data
1 -
Blog
1 -
Caching
2 -
CICDForDatabricksWorkflows
1 -
Cluster
1 -
Cluster Policies
1 -
Cluster Pools
1 -
Community Event
1 -
Cost Optimization Effort
1 -
custom compute policy
1 -
CustomLibrary
1 -
Data
1 -
Data Analysis with Databricks
1 -
Data Engineering
3 -
Data Governance
1 -
Data Mesh
1 -
Data Processing
1 -
Databricks Assistant
1 -
Databricks Community
1 -
Databricks Delta Table
1 -
Databricks Demo Center
1 -
Databricks Job
1 -
Databricks Migration
2 -
Databricks Mlflow
1 -
Databricks Notebooks
1 -
Databricks Support
1 -
Databricks Unity Catalog
2 -
Databricks Workflows
1 -
DatabricksML
1 -
DBR Versions
1 -
Declartive Pipelines
1 -
DeepLearning
1 -
Delta Live Table
1 -
Delta Live Tables
1 -
Delta Time Travel
1 -
Devops
1 -
DimensionTables
1 -
DLT
2 -
DLT Pipelines
3 -
DLT-Meta
1 -
Dns
1 -
Dynamic
1 -
Free Databricks
2 -
GenAI agent
1 -
GenAI and LLMs
2 -
GenAIGeneration AI
1 -
Generative AI
1 -
Genie
1 -
Governance
1 -
Hive metastore
1 -
Lakeflow Pipelines
1 -
Lakehouse
1 -
Lakehouse Migration
1 -
Lazy Evaluation
1 -
Learning
1 -
Library Installation
1 -
Llama
1 -
Medallion Architecture
1 -
Migrations
1 -
MSExcel
2 -
Multiagent
1 -
Networking
2 -
Partner
1 -
Performance
1 -
Performance Tuning
1 -
Private Link
1 -
Pyspark Code
1 -
Pyspark Databricks
1 -
Pytest
1 -
Python
1 -
Reading-excel
1 -
Scala Code
1 -
SDK
1 -
Serverless
2 -
Spark Caching
1 -
SparkSQL
1 -
SQL Serverless
1 -
Support Ticket
1 -
Sync
1 -
Tutorial
1 -
Unit Test
1 -
Unity Catalog
4 -
Unity Catlog
1 -
Warehousing
1 -
Workflow Jobs
1 -
Workflows
3
- « Previous
- Next »
User | Count |
---|---|
56 | |
43 | |
35 | |
28 | |
22 |