Dive into a collaborative space where members like YOU can exchange knowledge, tips, and best practices. Join the conversation today and unlock a wealth of collective wisdom to enhance your experience and drive success.
Added support for SHOW CREATE TABLE for materialized views and streaming tables. This will show the complete CREATE command used at creation time, including properties and schedulesCatalog Explorer "Overview" tab now shows the full CREATE command for...
LakeFlow Connect introduces simple ingestion connectors for databases, enterprise applications, and file sources. Set up efficient, low-maintenance pipelines in just a few clicks or via an API. Current sources include Salesforce, Workday, and SQL Ser...
We have tried to build a connection test logic to our software to try out the reachability of the SQL Warehouse, yet the connection parameters do not seem to function in expected manner.When the SQL Warehouse is running, the connection test functions...
Hello,
To create the connection you would need an endpoint, I would suggest you to give Serverless warehouse a try so that you don not have to wait, and for the suggestion on the product you may also submit a feedback and share the details of use cas...
This is not a question, this is just the solution to a problem we encountered in case someone from the community finds it useful.Recently we encountered an issue, where our users' jobs started failing out of nowhere on the following command, with the...
AI/BI dashboards can now be managed through Terraform.Dashboard using serialized_dashboard attribute: data "databricks_sql_warehouse" "starter" {
name = "Starter Warehouse"
}
resource "databricks_dashboard" "dashboard" {
display_name = "...
Hello Everyone!!!I hope everyone knows about the exciting new launch from Databricks: LakeFlow. This innovative tool makes data engineering simpler and more efficient. Let’s understand what LakeFlow offers and how it can help data teams. Let’s unders...
AI/BI dashboards:Improvements:The assistant can now help you edit chart axis titles and control the visibility of chart data labels.You can now assign multiple measures to pivot tables.Added more relative datetime options, such as Last 10 seconds and...
Does anyone know how to do DevOps CICD pipeline to deploy changes to a Databricks Catalog? I've added a couple tables in DEV and need to deploy in Prod. Could'nt find any info on this anywhere.
Thanks @Retired_mod , the links refer mostly deployments of workspace objects rather than changes in database objects, eg. a new schema, table, or new column added to a table etc. I'll keep looking...
Databricks, a unified data analytics platform, is widely recognized for its ability to process big data and run complex algorithms. When it comes to infrastructure as code, Terraform stands out as a tool that can manage Databricks resources efficient...
Hi @Ajay-Pandey ,
Thank you for sharing this comprehensive guide on managing Databricks resources with Terraform. The detailed explanation of the different authentication methods and examples is very helpful. I appreciate you taking the time to put t...
Hello Everyone, 3 type of reading mode into pyspark:1.Permissive Mode : its a default mode. if spark is not able to parse it due to dattype mismatch then make it as null without impacting other result, i.e only issue record are null others are fine.2...
Lakehouse Monitoring provides automated profiling and a dashboard that visualizes trends and anomalies over time. Track key metrics such as data volume, percent nulls, numerical distribution changes, and categorical distribution. For inference tables...
Reference the result of SQL tasks in Workflows and use it for control flow (If/else condition, For each task) or as a parameter of another task.Enable it today from your Previews page.
Thank you for sharing this valuable information with the community! @Ajay-Pandey This will surely help many users manage their data workflows more efficiently. We appreciate your contribution and encourage more insights like this.
Best practices for utilizing Databricks Asset Bundles effectively streamline your data and AI projects, ensuring they are robust, scalable, and maintainable. Here are some key practices to consider:1. Version Control: Always maintain a versioned hist...
Thank you for sharing these best practices. @Ajay-Pandey We appreciate your contribution to helping the community members streamline their projects effectively.