Engage in discussions on data warehousing, analytics, and BI solutions within the Databricks Community. Share insights, tips, and best practices for leveraging data for informed decision-making.
Hi, I'm currently starting to use SQL Warehouse, and we have most of our lake in a compression different than snappy.
How can I set the SQL warehouse to use a compression like gzip, zstd, on CREATE, INSERT, etc?
Tried this:
set spark.sql.parquet.comp...
Hi @Alejandro Martinez Great to meet you, and thanks for your question! Let's see if your peers in the community have an answer to your question. Thanks.
Availability of SQL Warehouse to Data Science and Engineering persona
Hi All,
Now we can use SQL Warehouse in our notebook execution.
It's in preview now and soon will be GA.
Refreshing SQL DashboardYou can schedule the dashboard to automatically refresh at an interval.At the top of the page, click Schedule.If the dashboard already has a schedule, you see Scheduled instead of Schedule.Select an interval, such as Every 1 h...
We are going to be a databricks customer and did some PoC tests. Our one test contains dataset in one partitioned table (15colums) is roughly 250M rows, each partition is ~50K-150K rows.
Occasionally we have hundreds (up to one thousand) concurrent u...
Hi @Marian Kovac Hope all is well! Just wanted to check in if you were able to resolve your issue and would you be happy to share the solution or mark an answer as best? Else please let us know if you need more help. We'd love to hear from you.Thank...
Hi @vinay kumar Hope all is well! Just wanted to check in if you were able to resolve your issue and would you be happy to share the solution or mark an answer as best? Else please let us know if you need more help. We'd love to hear from you.Thanks...