cancel
Showing results for 
Search instead for 
Did you mean: 
Get Started Discussions
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and introductory topics. Connect with beginners and experts alike to kickstart your Databricks experience.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

valjas
by New Contributor III
  • 4723 Views
  • 3 replies
  • 1 kudos

Clusters are really slow

We have two environments for our Azure Databricks. Dev and Prod. We had clusters created and tested in Dev environment, then they were exported to the prod environment through APIs. The clusters in Dev are performing as expected. Whereas, the cluster...

  • 4723 Views
  • 3 replies
  • 1 kudos
Latest Reply
Kaniz_Fatma
Community Manager
  • 1 kudos

Hi @valjas, Workspace Creation and Cluster Performance: Actions taken during the creation of a workspace can indeed impact cluster performance. When setting up a workspace, consider the following factors: Configuration Settings: Ensure that the wor...

  • 1 kudos
2 More Replies
NC
by New Contributor III
  • 3545 Views
  • 3 replies
  • 0 kudos

Resolved! Unable to use job cluster for task in workflows

Hi,I have a workflow setup in Databricks using 12.2 LTS ML.I am trying to use a job cluster for the task but i am getting the following error: Spark Conf: ‘spark.databricks.acl.enabled’ is not allowed when choosing an access modeAs a result I have to...

  • 3545 Views
  • 3 replies
  • 0 kudos
Latest Reply
Kaniz_Fatma
Community Manager
  • 0 kudos

Hi @NC ,  The error message you're encountering, "Spark Conf: ‘spark.databricks.acl.enabled’ is not allowed when choosing an access mode," is likely due to the job cluster's access mode not being set to "assigned" or "no isolation shared" as required...

  • 0 kudos
2 More Replies
aayusha3
by New Contributor II
  • 963 Views
  • 2 replies
  • 0 kudos

Internal error: Attach your notebook to a different compute or restart the current compute.

I am currently using a personal computer cluster [13.3 LTS (includes Apache Spark 3.4.1, Scala 2.12)] on GCP attached to a notebook. After running a few command lines without an issue, I end up getting this errorInternal error. Attach your notebook t...

aayusha3_0-1700256903637.png
  • 963 Views
  • 2 replies
  • 0 kudos
Latest Reply
Kaniz_Fatma
Community Manager
  • 0 kudos

Hi @aayusha3, Could you help us with the resource allocation for your cluster?  

  • 0 kudos
1 More Replies
Labels