- 2817 Views
- 7 replies
- 9 kudos
From Associate to Professional: My Learning Plan to ace all Databricks Data Engineer Certifications
In today’s data-driven world, the role of a data engineer is critical in designing and maintaining the infrastructure that allows for the efficient collection, storage, and analysis of large volumes of data. Databricks certifications holds significan...
- 2817 Views
- 7 replies
- 9 kudos
- 9 kudos
As an additional tip for those working towards both the Associate and Professional certifications, I recommend avoiding a long gap between the two exams to maintain your momentum. If possible, try to schedule them back-to-back with just a few days in...
- 9 kudos
- 712 Views
- 1 replies
- 3 kudos
Notification destination API
You can now create, delete, and update notification destinations through the Databricks API.The notification destinations API lets you programmatically manage a workspace's notification destinations. Notification destinations are used to send notific...
- 712 Views
- 1 replies
- 3 kudos
- 3 kudos
Hi,Thank you for sharing this Ajay. We appreciate you keeping the community informed!Thanks,Anushree
- 3 kudos
- 1701 Views
- 2 replies
- 4 kudos
Onboarding your new Databricks AI/BI Genie
The integration of AI and BI into the modern data stack has been a game-changer for businesses seeking to leverage data-driven insights. Databricks, a leader in this innovative frontier, has introduced the AI/BI Genie, a tool designed to democratize ...
- 1701 Views
- 2 replies
- 4 kudos
- 4 kudos
Hi,The AI/BI Genie is a fantastic innovation! By enabling natural language queries and learning from user interactions, it makes data analytics more accessible and insightful. It’s a powerful tool for businesses looking to enhance their data-driven d...
- 4 kudos
- 1106 Views
- 1 replies
- 2 kudos
The Importance of Databricks in Today's Data Market
Unlocking the Power of Data with Databricks In the rapidly evolving landscape of data and analytics, Databricks has emerged as a transformative force, reshaping how organizations handle big data, data engineering, and machine learning. As we naviga...
- 1106 Views
- 1 replies
- 2 kudos
- 2 kudos
Hi @Rishabh-Pandey,Thank you for sharing this! Databricks is really making a difference with its unified platform and powerful features. Excited to see how it will continue to shape the future of data! Thanks,
- 2 kudos
- 6250 Views
- 1 replies
- 1 kudos
Databricks on Databricks: Kicking off the Journey to Governance with Unity Catalog
In the world of data engineering and analytics, governance is paramount. Databricks has taken a significant step forward in this arena with the introduction of Unity Catalog, a unified governance solution for all data and AI assets. This journey to e...
- 6250 Views
- 1 replies
- 1 kudos
- 1 kudos
Hi Rishabh ,Unity Catalog is a major advancement in data governance, enhancing how we manage and secure data and AI assets in Databricks. Exciting to see these improvements!Thanks,Anushree
- 1 kudos
- 782 Views
- 1 replies
- 2 kudos
How to detect the Risks in Claims Data Using Databricks and PySpark
As a data engineer with experience in Databricks and other data engineering tools, I know that processing claims data and detecting risks early can really help in insurance claims processing. In this article, I’ll show you how to use Databricks and P...
- 782 Views
- 1 replies
- 2 kudos
- 2 kudos
Hi, Thanks for sharing this! Kudos for breaking it down so clearly. I’m sure, it will help other community members. Thanks,Anushree
- 2 kudos
- 431 Views
- 0 replies
- 2 kudos
September 2024 - Databricks SQL fixes
Added customization options for number formats in dashboard widgets, including currencies and percentages.The width of the Values column in pivot tables now auto-adjusts based on the width of the cell value names.The time displayed in dashboard widge...
- 431 Views
- 0 replies
- 2 kudos
- 503 Views
- 1 replies
- 0 kudos
SQL code for appending a notebook result into an existing database table
I am attempting to append the results from a notebook query results table into an existing databricks database table. By chance would someone share an example of the sql code with me?
- 503 Views
- 1 replies
- 0 kudos
- 0 kudos
Hi @wheersink ,So let's say you created following table with some sample values.%sql CREATE TABLE dev.default.employee ( id INT, name STRING, age INT, department STRING ); INSERT INTO dev.default.employee VALUES (1, 'John Doe', 30, 'Financ...
- 0 kudos
- 1277 Views
- 1 replies
- 2 kudos
🚀 Boost Your Data Pipelines with Dynamic, Data-Driven Databricks Workflows (For Each Task)! 💡
Unlock the power of the For Each task in Databricks to seamlessly iterate over collections—whether it's a list of table names or any value—and dynamically run tasks with specific parameter values. This powerful feature lets you automate repetitive pr...
- 1277 Views
- 1 replies
- 2 kudos
- 745 Views
- 0 replies
- 2 kudos
Databricks August Update
August 2024 | Databricks on AWS
- 745 Views
- 0 replies
- 2 kudos
- 759 Views
- 0 replies
- 2 kudos
Code patterns / best practices in SparkSQL, pyspark and Scala
Hi all,I am looking for collections of Code patterns for:SparkSQLpySparkScalaI'm sure there are at least a few repos in Github with snippets and will share them as I find them in this thread. If you come across any good collections, please post your...
- 759 Views
- 0 replies
- 2 kudos
- 1630 Views
- 0 replies
- 2 kudos
Cross-filtering for AI/BI dashboards
AI/BI dashboards now support cross-filtering, which allows you to click on an element in one chart to filter and update related data in other charts.Cross-filtering allows users to interactively explore relationships and patterns across multiple visu...
- 1630 Views
- 0 replies
- 2 kudos
- 522 Views
- 0 replies
- 2 kudos
SHOW CREATE TABLE for materialized views and streaming tables
Added support for SHOW CREATE TABLE for materialized views and streaming tables. This will show the complete CREATE command used at creation time, including properties and schedulesCatalog Explorer "Overview" tab now shows the full CREATE command for...
- 522 Views
- 0 replies
- 2 kudos
- 2332 Views
- 0 replies
- 2 kudos
LakeFlow Connect
LakeFlow Connect introduces simple ingestion connectors for databases, enterprise applications, and file sources. Set up efficient, low-maintenance pipelines in just a few clicks or via an API. Current sources include Salesforce, Workday, and SQL Ser...
- 2332 Views
- 0 replies
- 2 kudos
- 2336 Views
- 2 replies
- 3 kudos
Resolved! Timeout handling with JDBC connection to SQL Warehouse
We have tried to build a connection test logic to our software to try out the reachability of the SQL Warehouse, yet the connection parameters do not seem to function in expected manner.When the SQL Warehouse is running, the connection test functions...
- 2336 Views
- 2 replies
- 3 kudos
- 3 kudos
Hello, To create the connection you would need an endpoint, I would suggest you to give Serverless warehouse a try so that you don not have to wait, and for the suggestion on the product you may also submit a feedback and share the details of use cas...
- 3 kudos
- 876 Views
- 0 replies
- 1 kudos
Databricks java.util.NoSuchElementException: None.get Error on "SHOW TABLES IN" command
This is not a question, this is just the solution to a problem we encountered in case someone from the community finds it useful.Recently we encountered an issue, where our users' jobs started failing out of nowhere on the following command, with the...
- 876 Views
- 0 replies
- 1 kudos
Connect with Databricks Users in Your Area
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.
If there isn’t a group near you, start one and help create a community that brings people together.
Request a New Group-
ADF Linked Service
1 -
ADF Pipeline
1 -
Advanced Data Engineering
1 -
AI
1 -
ApacheSpark
1 -
API
1 -
Append blob
1 -
Automation
1 -
AWS
1 -
Azure databricks
1 -
Azure DevOps
1 -
Azure devops integration
1 -
ChangingSchema
1 -
CICD
1 -
CICDForDatabricksWorkflows
1 -
Clone
1 -
Cluster
1 -
Cluster Pools
1 -
compute policies
1 -
compute policy
1 -
Cost
1 -
Cost Optimization Effort
1 -
custom compute policy
1 -
CustomLibrary
1 -
Data Engineering
1 -
Data Mesh
1 -
Data Processing
1 -
Databricks Delta Table
1 -
Databricks Demo Center
1 -
Databricks jobs
1 -
Databricks Migration
1 -
Databricks Mlflow
1 -
Databricks Support
1 -
Databricks Unity Catalog
2 -
Databricks Workflows
1 -
DatabricksJobsAPI
1 -
DatabricksML
1 -
DatabricksWorkflowsCICD
1 -
Date
1 -
Delta Lake
3 -
Devops
1 -
DimensionTables
1 -
Dns
1 -
Dynamic
1 -
Governance
1 -
Hive metastore
1 -
Jobs & Workflows
1 -
LakeFlow
1 -
Library Installation
1 -
Medallion Architecture
1 -
Mlops
1 -
MSExcel
1 -
Networking
1 -
Partner
1 -
Private Link
1 -
Pyspark Code
1 -
Pytest
1 -
Python
1 -
Question
1 -
Scala Code
1 -
Schema
1 -
Schema Evaluation
1 -
Serverless
1 -
Serverless SQL Datawarehouse
1 -
Spark
4 -
SparkSQL
1 -
SQL Serverless
1 -
Support Ticket
1 -
Sync
1 -
ucx
1 -
Unit Test
1 -
Unity Catalog
3 -
Unity Catlog
1 -
Workflow Jobs
1 -
Workflows
1
- « Previous
- Next »
User | Count |
---|---|
40 | |
14 | |
10 | |
9 | |
8 |