cancel
Showing results for 
Search instead for 
Did you mean: 
Warehousing & Analytics
Engage in discussions on data warehousing, analytics, and BI solutions within the Databricks Community. Share insights, tips, and best practices for leveraging data for informed decision-making.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

MadelynM
by Databricks Employee
  • 3107 Views
  • 0 replies
  • 0 kudos

[Recap] Data + AI Summit 2024 - Warehousing & Analytics | Improve performance and increase insights

Here's your Data + AI Summit 2024 - Warehousing & Analytics recap as you use intelligent data warehousing to improve performance and increase your organization’s productivity with analytics, dashboards and insights.  Keynote: Data Warehouse presente...

Screenshot 2024-07-03 at 10.15.26 AM.png
Warehousing & Analytics
AI BI Dashboards
AI BI Genie
Databricks SQL
  • 3107 Views
  • 0 replies
  • 0 kudos
User16826992666
by Databricks Employee
  • 2431 Views
  • 1 replies
  • 0 kudos

Resolved! Should I enable Photon on my SQL Endpoint?

I see the option to enable Photon when creating a new SQL Endpoint. The description says that enabling it helps speed up up queries, which sounds good, but are there any downsides I need to be aware of?

  • 2431 Views
  • 1 replies
  • 0 kudos
Latest Reply
Ryan_Chynoweth
Databricks Employee
  • 0 kudos

Generally, yes you should enable photon. The majority of functionality is available and will perform extremely well. There are some limitations with it that can be found here. Limitations: Works on Delta and Parquet tables only for both read and writ...

  • 0 kudos
User16826992666
by Databricks Employee
  • 3036 Views
  • 1 replies
  • 0 kudos

Resolved! How can I see the performance of individual queries in Databricks SQL?

If I want to get more information about how an individual query is performing in the Databricks SQL environment, is there anywhere I can see that?

  • 3036 Views
  • 1 replies
  • 0 kudos
Latest Reply
sajith_appukutt
Databricks Employee
  • 0 kudos

You could see details on different queries that ran against an endpoint under the query history section

  • 0 kudos
User16826992666
by Databricks Employee
  • 6085 Views
  • 1 replies
  • 0 kudos

Resolved! Can you run Structured Streaming on a job cluster?

Need to know if I can use job clusters to start and run streaming jobs or if it has to be interactive

  • 6085 Views
  • 1 replies
  • 0 kudos
Latest Reply
sajith_appukutt
Databricks Employee
  • 0 kudos

Yes. Here is a doc containing some info on running Structured Streaming in production using Databricks jobs

  • 0 kudos
User16788316451
by Databricks Employee
  • 1516 Views
  • 1 replies
  • 0 kudos

How to troubleshoot SSL certificate errors while connecting Business Intelligence (BI) tools to Databricks in a Private Cloud (PVC) environment?

How to troubleshoot SSL certificate errors while connecting Business Intelligence (BI) tools to Databricks in a Private Cloud (PVC) environment?

  • 1516 Views
  • 1 replies
  • 0 kudos
Latest Reply
User16788316451
Databricks Employee
  • 0 kudos

See attached for steps to inspect the certificate chain using openssl

  • 0 kudos
Srikanth_Gupta_
by Databricks Employee
  • 2581 Views
  • 1 replies
  • 1 kudos

Reading bulk CSV files from Spark

While trying to read 100GB of csv.gz file from Spark which is taking forever to read, what are best options to read this file faster?

  • 2581 Views
  • 1 replies
  • 1 kudos
Latest Reply
sean_owen
Databricks Employee
  • 1 kudos

Part of the problem here is that .gz files are not splittable. If you have 1 huge 100GB .gz file, it can only be processed by one task. Can you change your input to use a splittable compression like .bz2? it'll work much better.

  • 1 kudos
User16826992666
by Databricks Employee
  • 1560 Views
  • 1 replies
  • 0 kudos
  • 1560 Views
  • 1 replies
  • 0 kudos
Latest Reply
sean_owen
Databricks Employee
  • 0 kudos

No. If you use %pip or %conda to attach a library, then it will only affect the execution of the notebook. A separate virtualenv is created for each notebook and its dependencies, even on a shared cluster.If you create a Library in the workspace and ...

  • 0 kudos
User16753724663
by Databricks Employee
  • 6424 Views
  • 1 replies
  • 0 kudos

Unable to use JDBC/ODBC url with sql workbench

  SQL Workbench is not able to connect to Cluster using JDBC/ODBC connection. Getting the following error. I used the configuration provided by the cluster (jdbc:spark://<host>.cloud.databricks.com:443/default;transportMode=http;ssl=1;httpPath=sql/pr...

  • 6424 Views
  • 1 replies
  • 0 kudos
Latest Reply
User16753724663
Databricks Employee
  • 0 kudos

As we are getting 401 error that means an authentication issue. We should use Personal access token for password.The username should be "token" and the password should be PAT token.

  • 0 kudos
User16753724663
by Databricks Employee
  • 2350 Views
  • 1 replies
  • 0 kudos

Unable to install kneed library in cluster with DBR version 5.5 LTS

I have an issue to install and use kneed python libary.    https://pypi.org/project/kneed/   I can install it and check it from log.   [Install command] %sh pip install kneed   [log] Installing collected packages: kneed Successfully installed kneed-0...

  • 2350 Views
  • 1 replies
  • 0 kudos
Latest Reply
User16753724663
Databricks Employee
  • 0 kudos

The kneed library has a dependency and we need to install them as well in order to work:numpy==1.18scipy==1.1.0scikit-learn==0.21.3Once we install the above libraries using GUI, we can run the below command to check the installed library with the cor...

  • 0 kudos
User16753724663
by Databricks Employee
  • 4068 Views
  • 1 replies
  • 0 kudos

Unable to construct the sql url as the password is having special characters.

while using the sqlalchemy, unable to connect with sql server from databricks:   user='user@host.mysql.database.azure.com' password='P@test' host="host.mysql.database.azure.com" database = "db" connect_args={'ssl':{'fake_flag_to_enable_tls': True}} c...

  • 4068 Views
  • 1 replies
  • 0 kudos
Latest Reply
User16753724663
Databricks Employee
  • 0 kudos

We can use urllib.parse to handle special characters. Here is an example:import urllib.parse user='user@host.mysql.database.azure.com'   password=urllib.parse.quote_plus("P@test") host="host.mysql.database.azure.com" database = "db"   connect_args={'...

  • 0 kudos
User16826992666
by Databricks Employee
  • 3061 Views
  • 1 replies
  • 0 kudos
  • 3061 Views
  • 1 replies
  • 0 kudos
Latest Reply
User16826992666
Databricks Employee
  • 0 kudos

SQL Analytics actually uses several layers of caching. Some documentation about the different layers can be found here in the documentation. There are two primary layers that users will experience. 1) The first is that the actual data results of spec...

  • 0 kudos
User16826992666
by Databricks Employee
  • 1694 Views
  • 1 replies
  • 0 kudos
  • 1694 Views
  • 1 replies
  • 0 kudos
Latest Reply
User16826992666
Databricks Employee
  • 0 kudos

At this time user credential passthrough is not supported on SQL Endpoints. Instead Databricks recommends using Table ACL's for data security.

  • 0 kudos
User16826992666
by Databricks Employee
  • 1740 Views
  • 1 replies
  • 0 kudos

Can you run data manipulation queries in Databricks SQL?

I am wondering if you can run queries that manipulate the actual data from within the Databricks SQL environment, or is it the case that you can only query tables?

  • 1740 Views
  • 1 replies
  • 0 kudos
Latest Reply
User16826992666
Databricks Employee
  • 0 kudos

Yes, as long as proper permissions are in place you can run data manipulation through the SQL interface.

  • 0 kudos