Delivery audit logs to multiple S3 buckets
Hi!Am I able to configure delivery of Databricks audit logs to multiple S3 buckets (on different AWS accounts)? Thanks in Advance!
- 2121 Views
- 1 replies
- 1 kudos
Hi!Am I able to configure delivery of Databricks audit logs to multiple S3 buckets (on different AWS accounts)? Thanks in Advance!
import requestsimport json# Databricks workspace API URLdatabricks_url = "https://dbc-ab846cbe-f48b.cloud.databricks.com/api/2.0/workspace/import"# Databricks API token (generate one from your Databricks account)databricks_token = "xxxxxxxxxxxxxxxxxx...
Hello, Thanks for contacting Databricks Support. It appears you're employing a CloudFormation template to establish a Databricks workspace. The recommended method for creating workspaces is through the AWS Quick Start. Please refer to the documenta...
Hi, i am facing below error while accessing Microosfot sql server. Please suggest what permissions I need to check at database level. I have the scope and secret created and key vault set up as expected. I feel some DB permission issue.Error: com.mi...
Hi Team, I scheduled my exam today and I showed the room in proctor. They said dull light. But I turned on with a better place. They again wanted to share the room and suspended the exam. Please help me asapWebassessor Id: npt.senthil@gmail.com
Thanks I got the rescheduled invite. Thanks much!
Hi, I'm creating an mlflow pyfunc serving endpoint and I would like to connect to a delta table to retrieve some information within the pyfunc. Is this possible?I ask because I don't think that serving endpoint environment has access to spark, and we...
Hello we need to change the billing account associated with our Databricks subscription. Is there any documentation available describing the procedure to be followed ? ThanksHoratiu
Start by logging into the Google Cloud Platform. If you are a new user, you need to create an account before you subscribe to Data bricks. Once in the console, start by selecting an existing Google Cloud project, or create a new project, and confirm ...
We've noticed that the GKE worker nodes which are automatically created when Databricks workspace is created inside GCP project are using the default compute engine SA which's not the best security approach, even Google doesn't recommend using defaul...
Hello Team, I have a databricks workspace running on an AWS environment. I have a requirement where the team wanted to add a few customized tags as per the docs I see below the recommendationTagValue:Description: All new AWS objects get a tag with t...
"run failed with error message Driver of the cluster (0307-***-gpbwt) was restarted during the run.", "effectiveIntegrationRuntime": "vnet-ir-*-**** (East US) while performing merge operation.This error is not consistant. DB runtime: 11.3 LTS (inclu...
Hi all! I am working on a DevOps project to automate the creation of UC objects through different environments (dev-test-prod). Each time we deploy our code to a different environment (using a Github workflow, not really relevant) we want to also cre...
Hi Team,I've recently begun working with Databricks and I'm exploring options for setting up a CI/CD pipeline to pull the latest code from GitHub.I have to pull latest code(.sql) from Github whenever push is done to main branch and update .sql notebo...
FWIW:we pull manually, but it is possible to automate that without any cost if you use Azure Devops. There is a free tier (depending on the number of pipelines/duration).
I have been using the Azure Databricks Workspace Editor for a few weeks to put together a python script as well as a notebook.All was well, until yesterday evening. Since then I suddenly have the following issuethe cursor in the editor is misbehaving...
Thanks @Chibberto - I will try the zoom level to see if it makes a difference.In the meantime, the latest issue is that the Autosave is not kicking in sometimes for several minutes. So, if I make a change and then re-run the job - the latest code is ...
Hi Team,Can you please direct me to any content on Databricks on Virtualization?Regards,Phanindra
Platform You can now use Structured Streaming to Stream Data from Apache Pulsar on Databricks. For more information : https://docs.databricks.com/en/structured-streaming/pulsar.html (DBR 14.1 required)Databricks Runtime 14.1 and 14.1 ML are now avail...
Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!
Sign Up NowUser | Count |
---|---|
133 | |
114 | |
56 | |
42 | |
31 |