cancel
Showing results for 
Search instead for 
Did you mean: 
Get Started Discussions
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and introductory topics. Connect with beginners and experts alike to kickstart your Databricks experience.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

AbhishekNegi
by New Contributor
  • 2655 Views
  • 1 replies
  • 1 kudos

New Cluster 90% memory already consumed

Hi, seeing this on all new clusters (single or multi-node) I am creating. As soon as the metrics start showing up, the memory consumption shows 90% already consumed between Used and Cached (something like below). This is the case with higher or lower...

AbhishekNegi_0-1725911074420.png AbhishekNegi_1-1725911119189.png
  • 2655 Views
  • 1 replies
  • 1 kudos
Latest Reply
saikumar246
Databricks Employee
  • 1 kudos

Hi @AbhishekNegi I understand your concern. The reason for you to see memory consumption before initiating any task and regarding the comment taking time to execute. This is how Spark internally works. The memory consumption observed in a Spark clust...

  • 1 kudos
RobsonNLPT
by Contributor III
  • 10994 Views
  • 15 replies
  • 3 kudos

Delta Live Tables Permissions

Hi allI'm the owner of delta live tables pipelines but I don't see the option described on documentation to grant permissions for different users. The options available are "settings" and "delete"In the sidebar, click Delta Live Tables.Select the nam...

  • 10994 Views
  • 15 replies
  • 3 kudos
Latest Reply
Walter_C
Databricks Employee
  • 3 kudos

Ok might be that the version of the workspaces could be different and the new patch will be implemented soon.

  • 3 kudos
14 More Replies
Nandhini_Kumar
by New Contributor III
  • 4608 Views
  • 1 replies
  • 0 kudos

How the Scale up process done in the databricks cluster?

For my AWS databricks cluster, i configured shared computer with 1min worker node and 3 max worker node, initailly only one worker node and driver node instance is created in the AWS console. Is there any rule set by databricks for scale up the next ...

  • 4608 Views
  • 1 replies
  • 0 kudos
Latest Reply
NandiniN
Databricks Employee
  • 0 kudos

Databricks uses autoscaling to manage the number of worker nodes in a cluster based on the workload. When you configure a cluster with a minimum and maximum number of worker nodes, Databricks automatically adjusts the number of workers within this ra...

  • 0 kudos
gb
by New Contributor
  • 3554 Views
  • 2 replies
  • 0 kudos

Write stream to Kafka topic with DLT

Hi,Is it possible to write stream to Kafka topic with Delta Live Table?I would like to do something like this:@dlt.view(name="kafka_pub",comment="Publish to kafka")def kafka_pub():return (dlt.readStream("source_table").selectExpr("to_json (struct (*)...

  • 3554 Views
  • 2 replies
  • 0 kudos
Latest Reply
dmytro
New Contributor III
  • 0 kudos

@shashas , is a Kafka sink now available? Where can we find information on setting it up, if yes?

  • 0 kudos
1 More Replies
tt_mmo
by New Contributor II
  • 2174 Views
  • 1 replies
  • 0 kudos

SQL table convert to R dataframe

I have a table with ~6 million rows. I am attempting to convert this from a sql table on my catalog to an R dataframe to use the tableone package. I separate my table into 3 tables each containing about 2 million rows then ran it through tbl() and as...

  • 2174 Views
  • 1 replies
  • 0 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 0 kudos

To handle a large SQL table (~6 million rows) and convert it into an R dataframe without splitting it into smaller subsets, you can use more efficient strategies and tools that are optimized for large datasets. Here are some recommendations: 1. Use `...

  • 0 kudos
hari-prasad
by Valued Contributor II
  • 1100 Views
  • 2 replies
  • 1 kudos

How to merge stats from my customer-academy to partner-academy Databricks

Hi,I have been using my customer-academy account from long time, and I recently for a partner-academy account to which I want to sync my stats.It is possible?

  • 1100 Views
  • 2 replies
  • 1 kudos
Latest Reply
hari-prasad
Valued Contributor II
  • 1 kudos

I have mailed to training-support, but no response yet. Just received confirmation email.

  • 1 kudos
1 More Replies
hari-prasad
by Valued Contributor II
  • 1482 Views
  • 1 replies
  • 1 kudos

When Databricks Enabling Support for Rust and Go in Notebook

Now #Rust and #GoLang are trending for their efficiency and speed. When can databricks enthusiasts can leverage the power of Rust and Golang in Databricks notebook to create data/ETL pipelines.  Any plan at #databricks ? 

  • 1482 Views
  • 1 replies
  • 1 kudos
Latest Reply
Walter_C
Databricks Employee
  • 1 kudos

Rust is an allowed language at Databricks if you must avoid a JVM process. I can see that the teams are working to provide additional support for Rust which might be available in the near future.

  • 1 kudos
shwetamagar
by New Contributor II
  • 2033 Views
  • 1 replies
  • 1 kudos

UC migration : Mount Points in Unity Catalog

Hi All,In my existing notebooks we have used mount points url as /mnt/ and we have notebooks where we have used the above url to fetch the data/file from the container. Now as we are upgrading to unity catalog these url will no longer be supporting a...

  • 2033 Views
  • 1 replies
  • 1 kudos
Latest Reply
Walter_C
Databricks Employee
  • 1 kudos

Unfortunately no, mount points are no longer supported with UC, so you will need to modify the URL manually on your notebooks.

  • 1 kudos
Thekenal
by New Contributor II
  • 627 Views
  • 1 replies
  • 0 kudos
  • 627 Views
  • 1 replies
  • 0 kudos
Latest Reply
hari-prasad
Valued Contributor II
  • 0 kudos

Hi @Thekenal , You can following link to first connect to Azure SQL server from databricks https://learn.microsoft.com/en-us/azure/databricks/connect/external-systems/sql-server.Then follow dashboard creation within Databricks https://docs.databricks...

  • 0 kudos
Sudheer2
by New Contributor III
  • 1252 Views
  • 3 replies
  • 0 kudos

Issue with Validation After DBFS to Volume Migration in Databricks Workspace

Hello Databricks Community,I have successfully migrated my DBFS (Databricks File System) from a source workspace to a target workspace, moving it from a path in Browse DBFS -> Folders to a Catalog -> Schema -> Volume.Now, I want to validate the migra...

  • 1252 Views
  • 3 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

Hi @Sudheer2, thanks for your comments, you can try using %sh magic to list the folder and sub-directores using unix-like commands  for example:  

  • 0 kudos
2 More Replies
User1234
by New Contributor II
  • 6751 Views
  • 5 replies
  • 2 kudos

Cluster compute metrics

I want to fetch compute metrics(hardware, gpu and spark) and use them in certain dashboard on Databricks, however i'm not able to fetch them. i have tried GET API request and system tables. The system tables only have CPU utilization and memory utili...

Get Started Discussions
cluster
compute
metrics
  • 6751 Views
  • 5 replies
  • 2 kudos
Latest Reply
171442
New Contributor II
  • 2 kudos

How can we store the CPU & Memory metrics for GCP databricks centrally and setup some alerts incase if the usage is high and monitor the performance.

  • 2 kudos
4 More Replies
MartinK
by New Contributor II
  • 3379 Views
  • 3 replies
  • 1 kudos

Resolved! Constantly Running Interactive Clusters Best Practices

Hello there, I’ve been creating an ETL/ELT Pipeline with Azure Databricks Workflows, Spark and Azure Data Lake. It should process in Near Real Time changes  (A Change Data Capture process) from an Azure SQL Database. For that purpose, I will have sev...

  • 3379 Views
  • 3 replies
  • 1 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 1 kudos

No problem! let me know if you have any other questions.

  • 1 kudos
2 More Replies
ludovic_hall
by New Contributor II
  • 1527 Views
  • 3 replies
  • 0 kudos

Dashboard sharing in Databriks with Unity Catalog enabled

Hello,I am planning to deploy a workspace with Unity Catalog enabled. Deploying permissions in one place sounds like a good solution. It can even simplify dataset architecture by masking rows and columns.As an architect, I’m concerned about the user’...

  • 1527 Views
  • 3 replies
  • 0 kudos
Latest Reply
Walter_C
Databricks Employee
  • 0 kudos

I will suggest for you to submit a feature request for this through https://docs.databricks.com/en/resources/ideas.html#ideas 

  • 0 kudos
2 More Replies
RohithChippa
by New Contributor III
  • 1502 Views
  • 1 replies
  • 0 kudos

Clean Room Per Collaborator Pricing

For Clean Room Pricing it is mentioned50$/ collaborator per day.1) If I keep the clean room for may be 5 hrs and delete it later will I get charged for those 5hrs(10.41$)/collaborator or will I be charged 50$ regardless of active hrs of clean room fo...

  • 1502 Views
  • 1 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

Hi @RohithChippa, Thanks for your questions! Please find below feedback: 1) If you keep the clean room for 5 hours and delete it later, you will be charged $50 per collaborator for that day, regardless of the number of active hours the clean room was...

  • 0 kudos
Labels