Engage in discussions on data warehousing, analytics, and BI solutions within the Databricks Community. Share insights, tips, and best practices for leveraging data for informed decision-making.
Here's your Data + AI Summit 2024 - Warehousing & Analytics recap as you use intelligent data warehousing to improve performance and increase your organization’s productivity with analytics, dashboards and insights.
Keynote: Data Warehouse presente...
We need to build some BI dashboards for a client. We are having issues to find a way to Download a dataset behind a chart. Ideally the dataset subset after appliying a set of filters. The documentation says here that the full dataset could also be do...
I don't think promoting them to Admins is the best approach. You should be able to assign them to appropriate groups and set permissions to they data you want devs and guests to have query access to, respectively. Then, in terms of sharing the dashbo...
Hey everyone,So I managed to create a Streamlit app on Databricks, works fine deployed in the cloud.However, when I try to run it locally, it complains about "assert os.getenv('DATABRICKS_WAREHOUSE_ID'), "DATABRICKS_WAREHOUSE_ID must be set in app.ya...
I figured it out this morning. You have to use an underscore in the yaml file instead of a dash. This works for me: - name: "DATABRICKS_WAREHOUSE_ID" valueFrom: "sql_warehouse"
Hi all!Need help configuring ODBC connection on a Mac. I'm receiving the error "The web server failed to start, please verify if port 8020 is being used by another connection or annother application".I've downloaded and installed Databricks ODBC con...
Is there any official word as to when we can expect Chloropleth or Marker Map visuals in AI/BI Dashboards? I realize Legacy Dashboards are still supported, but it feels uncertain to build new ones with AI/BI Dashboards in GA.
Definitely. Marker Maps are now live. You can find documentation for that here. Choropleth maps are in actively development and should land in a few months!
Hi folks, I am creating a SQL dashboard and want to know if I can re-use datasets within the same dashboard. The screenshot below captures what I would like to do pretty well, but to summarize... I need to run an computationally expensive query and w...
@rskuntoji24 My current workaround has been to create the first query as a materialized view in my catalog, and then run a workflow daily to update the view. The downsides to this strategy is that it might not be practical for streaming data (not an ...
Hello all (and, if you are watching these posts, Databricks employees/representatives ),I hope you take this idea into account - I haven't been able to find a place where one can vote on features, but I believe that this idea would make very real im...
Although this is not getting any visibility, I've got another feature idea to add:* Colour coding or otherwise marking columns to reflect table relationIn a simple join between e.g. table "cart" and table "cartItems", columns from "cart" would be mar...
Hello,Right now I have chosen the free tier mode and region is east US. But I am unable to create a all purpose compute cluster. Any config I chose it says not supported. Any help is appreciated for pointing which config for worker and driver I shoul...
It seems to be an issue with instance availability on the zone, it could be that at the given time there were no infrastructure available or a limitation with the zone itself. Have you checked the link available in the error message?
Hi-I have a spark DF that I create a visual from, I added a dbutils widget to filter the visual. I want to display the visual in our current IU dashboards (that were created with SQL datasets and calculated measures) but when I click on the drop down...
Hello @HeathDG1!
It looks like your visualization was created in a Databricks Notebook using the dbutils widget for filtering. However, Databricks only allows SQL-based visualizations to be added to AI/BI dashboards, while notebook visuals can only b...
Hello,When I do some DML statements using a view "the_source_view" direcly in SQL as a basis, all works fine. I can do inserts in another Delta table based on the data coming out of this view.It works too from a Job: WorkFlow -> Task -> Notebook con...
Hi all, Does Delta Lake table guarantee linearizability for the following operations on a single record:- SELECT- UPDATE and DELETE with condition (WHERE clauses) If linearizability is not guaranteed, what consistency model does Delta Lake provide fo...
Hello @Torch3333!
Delta Lake does not guarantee linearizability for single record operations. In Delta Lake, isolation levels ensure consistency guarantees in transactions. By default, SELECT operations follow snapshot isolation, ensuring that reads ...
I've been trying to run a dbt project (sourced in Azure DevOps) in Databricks Workflows, but I get this error message: CalledProcessError: Command 'b'\nmkdir -p "/tmp/tmp-dbt-run-1124228490001263"\nunexpected_errors="$(cp -a -u "/Workspace/Repos/.in...
Hey Support Team,We are experiencing an issue with our Databricks warehouse where the auto-stop feature does not seem to be working as expected. Despite setting an idle timeout, the warehouse continues running after the configured auto-stop time has ...
We were able to diagnose and resolve the problem. The problem was caused due to a cube.js JDBC connection repeatedly connecting to our Databricks SQL Warehouse. Databricks SQL Warehouse does not scale down when there are repeated new connections made...
I have applied row-level security (RLS) on the department column in the table so that users can only see data related to their own department. The security policy works perfectly when I query the table in Databricks SQL.Now, I have built a Databricks...
Hi @Akshay_Petkar , which credential mode did you use to publish the dashboard?
In order to make access control work based on the viewer's credential, "Don't embed credentials" mode should be used. If you published the dashboard with "Embed credentia...