Explore discussions on Databricks administration, deployment strategies, and architectural best practices. Connect with administrators and architects to optimize your Databricks environment for performance, scalability, and security.
Hi, I am new to Databricks and I am exploring its CE. I have data science background and I am building an analytics webapp, instead of connecting to Supabase or other back-end services, is it possible for me to connect to Databricks directly? I could...
How about the Databricks REST API? DB offers a comprehensive REST API that allows you to programmatically interact with Databricks resources. You can use this API to submit jobs, run queries, manage clusters, and perform various other operations.
Hi,Has anybody tried using Duo for SSO (Single Sign On)?We used OIDC on Duo, and getting this error when we test. Basically, after we get authenticated by Duo, the redirect page shows up like this:Any tips would be greatly appreciated. Thanks!
Hi @yubin-apollo We are reaching out to inform you that Databricks Community has migrated to a new service provider. As a result, all users and published content are migrated on June 24, 2023.This change will enhance your overall Databricks Community...
Hi, I am looking for a way to better migrate our workspace to Unity Catalog. We have several tables created as external tables and created using mount points, but I am worried that migrating to unity catalog will break those tables if we migrate usin...
@Rodrigo_Mohr you need to mainly consider below things1. Number of managed tables2. Number of external tablesin case external tables upgrade tool will help which is built in with UC, where as for managed you need to either go CTAS/Deep clone. but we ...
The company is getting a high expensive cost in data traffic after setting up the VPC, now we notice that we have to set up a s3 gateway, I wonder what is the best practices
@Ayun In terms of s3 gateway wrt to aws that looks promising in terms of price, but i have not seen anything related to s3 gateway mentioned in databricks articles. for most of implementations we went with standard approach that databricks mentioned....
I hope this message finds you all well.I followed the necessary steps to enable the Unity Catalog in our workspace.However, upon saving the configuration, I noticed that the Unity Catalog feature is automatically being disabled.I have double-checked ...
I'm working on a tool to visualise who accessed which (unity) catalogs, schemas and tables. To do that, I would like to access the query plan of the queries in the query history. Unfortunately, It seems like the REST api doesn't support accessing tho...
You can create either a managed or an unmanaged (external) table in the bronze layer, depending on your preference and use case. If you choose to create a managed table, Databricks will manage both the metadata and the data for the table. If you choo...
Redis offers various data structures such as strings, lists, sets, and hashes. Depending on your use case, select the appropriate data structure for storing the crm enrich data. For example, if you need to store key-value pairs, Redis hashes (HSET, H...
Currently we use AWS Athena and aws glue catalog for our data lake. We would like to evaluate delta lake for data management. How shall we try this with the existing setup?
Actually the Glue Hive Metastore integration with Unity Catalog as been announced today at the Databricks Summit.Give it a try herehttps://docs.databricks.com/archive/external-metastores/aws-glue-metastore.html
Hello anybody tried to attach an internal Git Hub Enterprise server through databricks whiteout having to whitelist the public op of the control plane ?both on Azure or AWS ?
Hello! Is it possible to retrieve data from a SQL Endpoint in the Databricks SQL persona using the Data Science and Engineering persona? In other words, I would like to use pyspark in DS&E to query a table in Databricks SQL.#DatabricksSQL#Databricks...
you do not need the sql warehouse itself for that. for ds & e you need a classic cluster (not a sql endpoint) anyway so you can easily read the tables from the metastore using spark.read.table().Connecting the sql endpoint to the ds cluster seems od...
AWS Fleet instance types are now GA and available for clusters and instance pools. You can see more details here: https://docs.databricks.com/compute/aws-fleet-instances.html