cancel
Showing results for 
Search instead for 
Did you mean: 
Get Started Discussions
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and introductory topics. Connect with beginners and experts alike to kickstart your Databricks experience.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

lin
by New Contributor
  • 1149 Views
  • 0 replies
  • 0 kudos

Facing UNKNOWN_FIELD_EXCEPTION.NEW_FIELDS_IN_FILE

[UNKNOWN_FIELD_EXCEPTION.NEW_FIELDS_IN_FILE] Encountered unknown fields during parsing: [<field_name>], which can be fixed by an automatic retry: trueI am using Azure Databricks, and write with python code. Want to catch the error and raise. Tried wi...

  • 1149 Views
  • 0 replies
  • 0 kudos
mano7438
by New Contributor III
  • 8793 Views
  • 6 replies
  • 1 kudos

Resolved! Unable to create table with primary key

Hi Team,Getting below error while creating a table with primary key,"Table constraints are only supported in Unity Catalog."Table script : CREATE TABLE persons(first_name STRING NOT NULL, last_name STRING NOT NULL, nickname STRING,CONSTRAINT persons_...

  • 8793 Views
  • 6 replies
  • 1 kudos
Latest Reply
Debayan
Databricks Employee
  • 1 kudos

Hi, this needs further investigation, could you please raise a support case with Databricks? 

  • 1 kudos
5 More Replies
od
by New Contributor
  • 1173 Views
  • 1 replies
  • 0 kudos

How do I regulate notebook cache

I am experiencing error of over caching in databricks notebook. If i display different dfs one of the dfs get cache which after the result of others afterwards. Please how can I avoid the cache memory while using the notebook?

  • 1173 Views
  • 1 replies
  • 0 kudos
Latest Reply
-werners-
Esteemed Contributor III
  • 0 kudos

I don't exactly understand what your issue is. Can you elaborate more?

  • 0 kudos
inesandres567
by New Contributor
  • 1704 Views
  • 1 replies
  • 0 kudos

Problem starting cluster

I try to start cluster that i used to start it 7 times before and it gave me this error Cloud provider is undergoing a transient resource throttling. This is retryable. 1 out of 2 pods scheduled. Failed to launch cluster in kubernetes in 1800 seconds...

  • 1704 Views
  • 1 replies
  • 0 kudos
Latest Reply
Debayan
Databricks Employee
  • 0 kudos

Hi, This error "GCE out of resources" typically means that Google compute engine is out of resources as in out of nodes (can be a quota issue or can be node issues in that particular region in GCP). Could you please raise a google support case on thi...

  • 0 kudos
inesandres567
by New Contributor
  • 1159 Views
  • 1 replies
  • 0 kudos

Fail start cluster

I try to start cluster that i used to start it 7 times before and it gave me this error Cloud provider is undergoing a transient resource throttling. This is retryable. 1 out of 2 pods scheduled. Failed to launch cluster in kubernetes in 1800 seconds...

  • 1159 Views
  • 1 replies
  • 0 kudos
Latest Reply
Debayan
Databricks Employee
  • 0 kudos

Hi, This error "GCE out of resources" typically means that Google compute engine is out of resources as in out of nodes (can be a quota issue or can be node issues in that particular region in GCP). Could you please raise a google support case on thi...

  • 0 kudos
mderela
by New Contributor II
  • 6493 Views
  • 3 replies
  • 1 kudos

inegstion time clustering

Hello, in rerence to https://www.databricks.com/blog/2022/11/18/introducing-ingestion-time-clustering-dbr-112.htmlI have a silly question how to use it. So let's assume that I have a few TB of not partitioned data. So, if I would like to query on dat...

  • 6493 Views
  • 3 replies
  • 1 kudos
Simon_T
by New Contributor III
  • 5530 Views
  • 0 replies
  • 0 kudos

Databricks Terraform Cluster Issue.

Error: default auth: cannot configure default credentials. Config: token=***. Env: DATABRICKS_TOKENon cluster.tf line 27, in data “databricks_spark_version” “latest_lts”:27: data “databricks_spark_version” “latest_lts” {

  • 5530 Views
  • 0 replies
  • 0 kudos
diego_poggioli
by Contributor
  • 2814 Views
  • 1 replies
  • 1 kudos

Resolved! Run tasks conditionally "Always" condition missing?

Does the new feature 'Run If' that allows you to run tasks conditionally lack the 'ALWAYS' option? In order to execute the task both when there is OK and error from the dependencies

  • 2814 Views
  • 1 replies
  • 1 kudos
Latest Reply
Lakshay
Databricks Employee
  • 1 kudos

You can choose the All Done option to run the task in both the scenarios

  • 1 kudos
kurtrm
by New Contributor III
  • 7018 Views
  • 2 replies
  • 3 kudos

How to send alert when cluster is running for too long

Hello,Our team recently experienced an issue where a teammate started a new workflow job then went on vacation. This job ended up running continuously without failing for 4.5 days. The usage of the cluster did not seem out of place during the workday...

  • 7018 Views
  • 2 replies
  • 3 kudos
Latest Reply
kurtrm
New Contributor III
  • 3 kudos

@Retired_mod,I ended up creating a job leveraging the Databricks Python SDK to check cluster and active job run times. The script will raise an error and notify the team if the cluster hasn't terminated or restarted in the past 24 hours or if a job h...

  • 3 kudos
1 More Replies

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels