Engage in discussions on data warehousing, analytics, and BI solutions within the Databricks Community. Share insights, tips, and best practices for leveraging data for informed decision-making.
Here's your Data + AI Summit 2024 - Warehousing & Analytics recap as you use intelligent data warehousing to improve performance and increase your organization’s productivity with analytics, dashboards and insights.
Keynote: Data Warehouse presente...
Users get confused when querying data with timestamps because UTC is not intuitive for many. It is possible to set TIME ZONE at query level or at SQL Warehouse level, but those options fail to address the need of multiple users working on the same wa...
it is possible to set the time zone at the session level using the SET TIME ZONE statement in Databricks SQL. This allows users to control the local timezone used for timestamp operations within their session. However, there is no direct option of us...
I'm trying to use Databricks Apps to host a Streamlit app to serve some interactive applicaton. I face two limitations:1. In environment for App I see it using certain python version, but how to update it to use any other version?It is already set to...
Hi @atikiwala ,
Good Day!
Python 3.11 is currently the only version we support. We are thinking of adding additional options in the future.
Would love to hear your feedback on this - https://docs.databricks.com/en/resources/ideas.html#submit-product-...
@ImranA and @igorstar
I repost my response here again:to create materialized views, you could use CREATE OR REFRESH LIVE TABLE however according to the official docs:
The CREATE OR REFRESH LIVE TABLE syntax to create a materialized view is deprecat...
Hi, I'd appreciate it if anyone could help!We are using offical ODBC driver (Simba Spark ODBC Driver 64-bit 2.08.02.1013) in our application. C++ APIs.All the following SQL statements are passed through ODBC API to Databricks:successully executing: C...
@AnnaP Hey,Did you try below:To disable the SQL Connector feature, select the Use Native Query check box.Important:l When this option is enabled, the connector cannot executeparameterized queries.l By default, the connector applies transformations to...
I am working on a use case where I have streaming data that needs to be displayed in real-time on a live dashboard. The goal is for any new data arriving in the stream to instantly reflect on the dashboard. Is this possible on Databricks? If yes, how...
@Akshay_Petkar wrote:I am working on a use case where I have streaming data that needs to be displayed in real-time on a live dashboard. The goal is for any new data arriving in the stream to instantly reflect on the dashboard. Is this possible on Da...
While using cx_oracle python library, it returns the below error:
error message: Cannot locate a 64-bit Oracle Client library: "libclntsh.so: cannot open shared object file: No such file or directory
The cx_oracle library is dependent on native...
Hi @AshvinManoj I used your script and still get same error. sudo echo 'LD_LIBRARY_PATH="/dbfs/databricks/instantclient_23_6"' >> /databricks/spark/conf/spark-env.shsudo echo 'ORACLE_HOME="/dbfs/databricks/instantclient_23_6"' >> /databricks/spark...
We're creating a report with Power BI using data from our AWS Databricks workspace. Currently, I can view the report on Power BI (service) after publishing. Is there a way to change the data source connection, e.g. if I want to change the data source...
In the Power BI transform data view, you should be able to access the M-Query code and actually change the server and the host directly. My recommendation is to create a couple of parameters to keep this info for all your queries. Then you can just c...
I am embedding a dashboard in my react application using an iframe.https://www.databricks.com/blog/how-embed-aibi-dashboards-your-websites-and-applicationshttps://docs.databricks.com/en/dashboards/index.html#embed-a-dashboard1. How can i let my users...
Hi , Has anyone gotten this working? I am facing the same issue when trying to get the embedded URL working from an iframe in my React application. I think I have all the necessary permissions in place, but still, the Databricks sign-in page opens in...
What is the optimal number of cluster/nodes in a warehouse?
Depends on your workload. Our DBSQL guide suggests a size range on two main things: time to execute your query and bytes spilled from it. This link can help you understand better optimizatio...
I have an on-premises Power BI Report Server that uses the Simba Spark ODBC Driver (2.8) to connect to Databricks. It can connect to a serverless warehouse successfully and run its queries, but it never seems to disconnect the session, and so the war...
Hi @bradleyjamrozik ,The problem is that the current designed behavior of Power BI can lead to the connection not being closed under certain conditions.In short, since Databricks is Async, if an error occurs while fetching data, the connection may no...
I am preparing to migrate from EMR to Databricks and would like to know the best practices for this process. Is there a direct connector, such as a JDBC connector, available to facilitate the migration? Alternatively, would it be more effective to ex...
@Akshay_Petkar if you're planning to migrate from your current technology to Databricks, Travinto Technologies' Code Converter Tool is here to make the process seamless. This powerful tool enables you to migrate data, ETL workflows, and reports acros...
I cancel my subscription plan (account created via accounts.databricks.com), without noticing that by doing this, all of my workspaces will be deleted. Is this possible to recover those workspace, and if possible, what should I do to recover them?
Hello San2.
To recover a workspace, could you please raise a case with support? They will be helping you to recover these workspaces, otherwise you can reach your account team, who will help you for further steps, for more details you see disaster ...
Hi all, working on this project, my team plans to migrate some data from some databases to Databricks. We plan to run this migration by submitting queries to a warehouse through python on a local machine.Now I was wondering what would be the best app...
@Ismail1 if you're planning to migrate from your current technology to Databricks, Travinto Technologies' Code Converter Tool is here to make the process seamless. This powerful tool enables you to migrate data, ETL workflows, and reports across pla...