cancel
Showing results for 
Search instead for 
Did you mean: 
Get Started Discussions
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and introductory topics. Connect with beginners and experts alike to kickstart your Databricks experience.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

AbhishekNegi
by New Contributor
  • 769 Views
  • 1 replies
  • 1 kudos

New Cluster 90% memory already consumed

Hi, seeing this on all new clusters (single or multi-node) I am creating. As soon as the metrics start showing up, the memory consumption shows 90% already consumed between Used and Cached (something like below). This is the case with higher or lower...

AbhishekNegi_0-1725911074420.png AbhishekNegi_1-1725911119189.png
  • 769 Views
  • 1 replies
  • 1 kudos
Latest Reply
saikumar246
Databricks Employee
  • 1 kudos

Hi @AbhishekNegi I understand your concern. The reason for you to see memory consumption before initiating any task and regarding the comment taking time to execute. This is how Spark internally works. The memory consumption observed in a Spark clust...

  • 1 kudos
RobsonNLPT
by Contributor III
  • 3587 Views
  • 15 replies
  • 3 kudos

Delta Live Tables Permissions

Hi allI'm the owner of delta live tables pipelines but I don't see the option described on documentation to grant permissions for different users. The options available are "settings" and "delete"In the sidebar, click Delta Live Tables.Select the nam...

  • 3587 Views
  • 15 replies
  • 3 kudos
Latest Reply
Walter_C
Databricks Employee
  • 3 kudos

Ok might be that the version of the workspaces could be different and the new patch will be implemented soon.

  • 3 kudos
14 More Replies
DebIT2011
by New Contributor II
  • 3048 Views
  • 3 replies
  • 7 kudos

Choosing between Azure Data Factory (ADF) and Databricks PySpark notebooks

I’m working on a project where I need to pull large datasets from Cosmos DB into Databricks for further processing, and I’m trying to decide whether to use Azure Data Factory (ADF) or Databricks PySpark notebooks for the extraction and processing tas...

  • 3048 Views
  • 3 replies
  • 7 kudos
Latest Reply
BlankRichards
New Contributor II
  • 7 kudos

Hi @DebIT2011,Deciding between Azure Data Factory (ADF) and Databricks PySpark Notebooks for data extraction and processing depends on several factors specific to your use case. Let’s address each aspect raised:Advantages of ADF over Databricks PySpa...

  • 7 kudos
2 More Replies
Nandhini_Kumar
by New Contributor III
  • 2636 Views
  • 1 replies
  • 0 kudos

How the Scale up process done in the databricks cluster?

For my AWS databricks cluster, i configured shared computer with 1min worker node and 3 max worker node, initailly only one worker node and driver node instance is created in the AWS console. Is there any rule set by databricks for scale up the next ...

  • 2636 Views
  • 1 replies
  • 0 kudos
Latest Reply
NandiniN
Databricks Employee
  • 0 kudos

Databricks uses autoscaling to manage the number of worker nodes in a cluster based on the workload. When you configure a cluster with a minimum and maximum number of worker nodes, Databricks automatically adjusts the number of workers within this ra...

  • 0 kudos
gb
by New Contributor
  • 1865 Views
  • 2 replies
  • 0 kudos

Write stream to Kafka topic with DLT

Hi,Is it possible to write stream to Kafka topic with Delta Live Table?I would like to do something like this:@dlt.view(name="kafka_pub",comment="Publish to kafka")def kafka_pub():return (dlt.readStream("source_table").selectExpr("to_json (struct (*)...

  • 1865 Views
  • 2 replies
  • 0 kudos
Latest Reply
dmytro
New Contributor III
  • 0 kudos

@shashas , is a Kafka sink now available? Where can we find information on setting it up, if yes?

  • 0 kudos
1 More Replies
tt_mmo
by New Contributor II
  • 661 Views
  • 1 replies
  • 0 kudos

SQL table convert to R dataframe

I have a table with ~6 million rows. I am attempting to convert this from a sql table on my catalog to an R dataframe to use the tableone package. I separate my table into 3 tables each containing about 2 million rows then ran it through tbl() and as...

  • 661 Views
  • 1 replies
  • 0 kudos
Latest Reply
BigRoux
Databricks Employee
  • 0 kudos

To handle a large SQL table (~6 million rows) and convert it into an R dataframe without splitting it into smaller subsets, you can use more efficient strategies and tools that are optimized for large datasets. Here are some recommendations: 1. Use `...

  • 0 kudos
hari-prasad
by Valued Contributor II
  • 514 Views
  • 2 replies
  • 1 kudos

How to merge stats from my customer-academy to partner-academy Databricks

Hi,I have been using my customer-academy account from long time, and I recently for a partner-academy account to which I want to sync my stats.It is possible?

  • 514 Views
  • 2 replies
  • 1 kudos
Latest Reply
hari-prasad
Valued Contributor II
  • 1 kudos

I have mailed to training-support, but no response yet. Just received confirmation email.

  • 1 kudos
1 More Replies
hari-prasad
by Valued Contributor II
  • 505 Views
  • 1 replies
  • 1 kudos

When Databricks Enabling Support for Rust and Go in Notebook

Now #Rust and #GoLang are trending for their efficiency and speed. When can databricks enthusiasts can leverage the power of Rust and Golang in Databricks notebook to create data/ETL pipelines.  Any plan at #databricks ? 

  • 505 Views
  • 1 replies
  • 1 kudos
Latest Reply
Walter_C
Databricks Employee
  • 1 kudos

Rust is an allowed language at Databricks if you must avoid a JVM process. I can see that the teams are working to provide additional support for Rust which might be available in the near future.

  • 1 kudos
shwetamagar
by New Contributor II
  • 349 Views
  • 1 replies
  • 1 kudos

UC migration : Mount Points in Unity Catalog

Hi All,In my existing notebooks we have used mount points url as /mnt/ and we have notebooks where we have used the above url to fetch the data/file from the container. Now as we are upgrading to unity catalog these url will no longer be supporting a...

  • 349 Views
  • 1 replies
  • 1 kudos
Latest Reply
Walter_C
Databricks Employee
  • 1 kudos

Unfortunately no, mount points are no longer supported with UC, so you will need to modify the URL manually on your notebooks.

  • 1 kudos
Thekenal
by New Contributor II
  • 229 Views
  • 1 replies
  • 0 kudos
  • 229 Views
  • 1 replies
  • 0 kudos
Latest Reply
hari-prasad
Valued Contributor II
  • 0 kudos

Hi @Thekenal , You can following link to first connect to Azure SQL server from databricks https://learn.microsoft.com/en-us/azure/databricks/connect/external-systems/sql-server.Then follow dashboard creation within Databricks https://docs.databricks...

  • 0 kudos
Sudheer2
by New Contributor III
  • 443 Views
  • 3 replies
  • 0 kudos

Issue with Validation After DBFS to Volume Migration in Databricks Workspace

Hello Databricks Community,I have successfully migrated my DBFS (Databricks File System) from a source workspace to a target workspace, moving it from a path in Browse DBFS -> Folders to a Catalog -> Schema -> Volume.Now, I want to validate the migra...

  • 443 Views
  • 3 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

Hi @Sudheer2, thanks for your comments, you can try using %sh magic to list the folder and sub-directores using unix-like commands  for example:  

  • 0 kudos
2 More Replies
User1234
by New Contributor II
  • 4739 Views
  • 5 replies
  • 2 kudos

Cluster compute metrics

I want to fetch compute metrics(hardware, gpu and spark) and use them in certain dashboard on Databricks, however i'm not able to fetch them. i have tried GET API request and system tables. The system tables only have CPU utilization and memory utili...

Get Started Discussions
cluster
compute
metrics
  • 4739 Views
  • 5 replies
  • 2 kudos
Latest Reply
171442
New Contributor II
  • 2 kudos

How can we store the CPU & Memory metrics for GCP databricks centrally and setup some alerts incase if the usage is high and monitor the performance.

  • 2 kudos
4 More Replies
MartinK
by New Contributor II
  • 981 Views
  • 3 replies
  • 1 kudos

Resolved! Constantly Running Interactive Clusters Best Practices

Hello there, I’ve been creating an ETL/ELT Pipeline with Azure Databricks Workflows, Spark and Azure Data Lake. It should process in Near Real Time changes  (A Change Data Capture process) from an Azure SQL Database. For that purpose, I will have sev...

  • 981 Views
  • 3 replies
  • 1 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 1 kudos

No problem! let me know if you have any other questions.

  • 1 kudos
2 More Replies
ludovic_hall
by New Contributor II
  • 507 Views
  • 3 replies
  • 0 kudos

Dashboard sharing in Databriks with Unity Catalog enabled

Hello,I am planning to deploy a workspace with Unity Catalog enabled. Deploying permissions in one place sounds like a good solution. It can even simplify dataset architecture by masking rows and columns.As an architect, I’m concerned about the user’...

  • 507 Views
  • 3 replies
  • 0 kudos
Latest Reply
Walter_C
Databricks Employee
  • 0 kudos

I will suggest for you to submit a feature request for this through https://docs.databricks.com/en/resources/ideas.html#ideas 

  • 0 kudos
2 More Replies

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels