cancel
Showing results for 
Search instead for 
Did you mean: 
Administration & Architecture
Explore discussions on Databricks administration, deployment strategies, and architectural best practices. Connect with administrators and architects to optimize your Databricks environment for performance, scalability, and security.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

cuhlmann
by New Contributor
  • 839 Views
  • 1 replies
  • 0 kudos

data ingestion from external system - auth via client certificate

Hi Community,we have the requirement to ingest data in azure databricks from external systems.Our customer ask us to use Client Certificate as authentication method.Requests - https://requests.readthedocs.io/en/latest/user/advanced/Aiohttp - https://...

  • 839 Views
  • 1 replies
  • 0 kudos
Latest Reply
filipniziol
Esteemed Contributor
  • 0 kudos

Hi @cuhlmann ,As I understand you need to ingest data into Azure Databricks from external systems, and your customer requires using client certificate authentication. The challenge is that the client certificate is stored in Azure Key Vault, but the ...

  • 0 kudos
JianfengHuang
by New Contributor
  • 527 Views
  • 2 replies
  • 1 kudos

Bill for Premium subscription

hi, there, I have subscribed the Premium plan of databricks, How can I get the bills for this subscription? I didn't find it from the account settings. Anyone can help?

  • 527 Views
  • 2 replies
  • 1 kudos
Latest Reply
gchandra
Databricks Employee
  • 1 kudos

AWS https://docs.databricks.com/en/admin/account-settings/account.html Azure https://learn.microsoft.com/en-us/azure/databricks/admin/account-settings/account GCP https://docs.gcp.databricks.com/en/admin/account-settings/account.html  

  • 1 kudos
1 More Replies
ossinova
by Contributor II
  • 4895 Views
  • 4 replies
  • 0 kudos

Override default Personal Compute policy using terraform / disable Personal Compute policy

I want to programmatically do some adjustments to the default personal compute resource or preferably create my own custom one based on the same configuration or policy family (in which all users can gain access to) when deploying a new workspace usi...

  • 4895 Views
  • 4 replies
  • 0 kudos
Latest Reply
Clowa
New Contributor II
  • 0 kudos

Only way I got it working was by importing the pre-existing policy into terraform and do an overwrite as already mentioned by @jsimonovic . The full code example looks like this:import { id = "001BF0AC280610B4" # Polcy ID of the pre-existing person...

  • 0 kudos
3 More Replies
mattiags
by New Contributor II
  • 853 Views
  • 1 replies
  • 1 kudos

Resolved! Retention for hive_metastore tables

HiI have a notebook that creates tables in the hive_metastore with the following code: df.write.format("delta").mode("overwrite").saveAsTable(output_table_name) Which is the retantion for the data saved in the hive metastore? is there any configurati...

  • 853 Views
  • 1 replies
  • 1 kudos
Latest Reply
Stefan-Koch
Valued Contributor II
  • 1 kudos

Hi mattiagsAs long as you do not delete the data via notebook or in the data lake, it will not be deleted in any other way. This means that there is no retention time in this sense, or conversely, it is infinite until you deliberately delete the data...

  • 1 kudos
abhishekdas
by New Contributor II
  • 2708 Views
  • 3 replies
  • 0 kudos

Resolved! Databricks on AWS - Changes to your Unity Catalog storage credentials

Hi Context: On June 30, 2023, AWS updated its IAM role trust policy, which requires updating Unity Catalog storage credentials. Databricks previously sent an email communication to customers in March 2023 on this topic and updated the documentation a...

  • 2708 Views
  • 3 replies
  • 0 kudos
Latest Reply
abhishekdas
New Contributor II
  • 0 kudos

Thank you for the response @MoJaMa - we will try it out tomorrow and post an update here.

  • 0 kudos
2 More Replies
NadithK
by Contributor
  • 3009 Views
  • 0 replies
  • 1 kudos

Pre-loading docker images to cluster pool instances still requires docker URL at cluster creation

I am trying to pre-load a docker image to a Databricks cluster pool instance.As per this article I used the REST API to create the cluster pool and defined a custom Azure container registry as the source for the docker images.https://learn.microsoft....

  • 3009 Views
  • 0 replies
  • 1 kudos
AlbertWang
by Valued Contributor
  • 3612 Views
  • 4 replies
  • 1 kudos

Resolved! How to use Databricks CLI as a service principal?

Hi all,I have a question about how to use Databricks CLI on my local environment as a service principal?I have installed Databricks CLI and configured the file `.databrickscfg` as shown below. [DEFAULT] host = https://adb-123123123.1.azuredatabr...

  • 3612 Views
  • 4 replies
  • 1 kudos
Latest Reply
Stefan-Koch
Valued Contributor II
  • 1 kudos

got you.I found a working solution. Try this one:[devsp] azure_workspace_resource_id = /subscriptions/bc0cd1..././.../Databricks/workspaces/my-workspace azure_tenant_id = bc0cd1... azure_client_id = fa0cd1... azure_client_secr...

  • 1 kudos
3 More Replies
Mayank1
by New Contributor
  • 1904 Views
  • 4 replies
  • 0 kudos
  • 1904 Views
  • 4 replies
  • 0 kudos
Latest Reply
case-k
New Contributor III
  • 0 kudos

Thank you so much!! I solve this by reinstalling chrome browser. I got this issue last week and can not solve it even if wait, clear cache, restart..etc. But worked in another type of browser. So I reinstall Chrome browser and it worked. Thank you 

  • 0 kudos
3 More Replies
adil-shiva
by New Contributor II
  • 1236 Views
  • 4 replies
  • 0 kudos

Interface for Databricks CE

My Databricks CE interface does not have the quick guides or toggles for Data Science & Engineering/Machine Learning. this is what it looks like and I want to see the quick guides. 

IMG_1294.png
  • 1236 Views
  • 4 replies
  • 0 kudos
Latest Reply
LauJohansson
Contributor
  • 0 kudos

What cloud provider do you use? 

  • 0 kudos
3 More Replies
GuyPerson
by New Contributor
  • 1941 Views
  • 1 replies
  • 0 kudos

Resolved! Docs/Info for metastore, artifact blob storage endpoints etc for Azure Databricks

Hi! This is probably a very newbie type of question and my google skills seems to be lacking but is there any in depth documentation/explanations of the metastore, artifact blob storage, system tables storage, log blob storage and event hubs services...

  • 1941 Views
  • 1 replies
  • 0 kudos
Latest Reply
shashank853
Databricks Employee
  • 0 kudos

Hi,You can check below: System Tables Storage Purpose: System tables storage is used to store system-level metadata and configuration data for the Azure Databricks workspace.Data Stored: This includes metadata related to the Unity Catalog, cluster c...

  • 0 kudos
Fiabane
by New Contributor III
  • 3230 Views
  • 2 replies
  • 3 kudos

Resolved! Databricks Asset Bundles + Artifacts + Poetry

Hello,I've configured the DABs on our project successfully. Moreover, I could switch from setuptools to poetry almost successfully. In the project's databricks.yml I configured it as the documentation suggested, I've just changed the name of the arti...

Fiabane_0-1729111836436.png
  • 3230 Views
  • 2 replies
  • 3 kudos
Latest Reply
filipniziol
Esteemed Contributor
  • 3 kudos

Hi @Fiabane ,Could you first check:Do you see your .whl file in your artifacts folder?Could you try to install the package by running the code in your notebook :         %pip install  <path to your wheel>As far as I understand you want to have a job ...

  • 3 kudos
1 More Replies
UnaiUribarri
by New Contributor II
  • 1437 Views
  • 2 replies
  • 0 kudos

Databricks Kryo setup

I would like to consolidate all our Spark jobs in Databricks. One of those jobs that are currently running in Azure HDInsight is not properly working using a Databricks JAR job.It uses Spark 3.3 RDDs and requires configuring Kryo serialisation. There...

  • 1437 Views
  • 2 replies
  • 0 kudos
Latest Reply
dilsan77
New Contributor II
  • 0 kudos

Integrating Spark tasks with Databricks can greatly improve your workflow. For tasks that require Kryo serialization, make sure you configure your Spark session correctly. You may need to adjust the serialization settings in your Spark configuration....

  • 0 kudos
1 More Replies
Jim-Shady
by New Contributor II
  • 751 Views
  • 1 replies
  • 2 kudos

Azure Databricks Classic Compute Plane Firewall

I’m designing a compute plane configuration that will align our data platform with internal policies from a security perspective. As part of this exercise I'm documenting how the permissible traffic inbound and outbound is controlled using NSG rules,...

  • 751 Views
  • 1 replies
  • 2 kudos
Latest Reply
michael569gardn
New Contributor III
  • 2 kudos

@Jim-Shady wrote:I’m designing a compute plane configuration that will align our data platform with internal policies from a security perspective. As part of this exercise I'm documenting how the permissible traffic inbound and outbound is controlled...

  • 2 kudos
PabloCSD
by Valued Contributor II
  • 4781 Views
  • 1 replies
  • 0 kudos

Resolved! How to deploy to Databricks Assets Bundle from Azure DevOps using Service Principal?

I have a CI/CD process that after a Pull Request (PR) to main it deploys to staging.It works using a Personal Access Token using Azure Pipelines.From local, deploying using Service Principal works (https://community.databricks.com/t5/administration-a...

  • 4781 Views
  • 1 replies
  • 0 kudos
Latest Reply
PabloCSD
Valued Contributor II
  • 0 kudos

I needed to deploy a job using CI/CD Azure Pipelines without using the OAuth, this is the way:First you need to have configured the Service Principal, for that you need to generate it in your workspace with this you will have:A host: Which is your wo...

  • 0 kudos