cancel
Showing results for 
Search instead for 
Did you mean: 
Get Started Discussions
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and introductory topics. Connect with beginners and experts alike to kickstart your Databricks experience.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Godhuli
by New Contributor II
  • 2014 Views
  • 0 replies
  • 0 kudos

Unable to login to Databricks Community edition

I signed up to Databricks Community Edition with gmail and verified my account and used it to create a notebook as well but having issue in re-logging in even though email and password provided are correct. I did  try  "Forget Password" step but I am...

  • 2014 Views
  • 0 replies
  • 0 kudos
arun949290
by New Contributor II
  • 4006 Views
  • 3 replies
  • 0 kudos

Unable to login to community edition

Hello there,I have successfully created a databricks account and went to login to the community edition with the exact same login credentials as my account, but it tells me that the email/password are invalid. I can login with these same exact creden...

  • 4006 Views
  • 3 replies
  • 0 kudos
Latest Reply
Erik_L
Contributor II
  • 0 kudos

Databricks Community Edition and Databricks are separate services. You have to create an account specific to community.

  • 0 kudos
2 More Replies
Omri
by New Contributor
  • 1329 Views
  • 1 replies
  • 0 kudos

Is it possible to create a scratchpad ui?

https://jupyter-contrib-nbextensions.readthedocs.io/en/latest/nbextensions/scratchpad/README.htmlIs something like this available on Databrick's notebooks ui?

  • 1329 Views
  • 1 replies
  • 0 kudos
Latest Reply
Debayan
Databricks Employee
  • 0 kudos

Hi, it would be great if you can submit the idea on this and the progress can be tracked by the ideas portal. https://docs.databricks.com/en/resources/ideas.html 

  • 0 kudos
SivaPK
by New Contributor II
  • 3585 Views
  • 1 replies
  • 1 kudos

Generated Access Token is deleted/Expired after lifetime 90 days? How to use old token now?

Hello Team,I have generated a new token via Admin Settings --> Developer --> Access Token -- > Manage.Now my token is deleted/Expired after 90 days. I know what is my token and generated alphanumeric one.Now how can i set or reuse the same token in d...

Get Started Discussions
access_token
generate_token
restore_token
settings
  • 3585 Views
  • 1 replies
  • 1 kudos
Latest Reply
Debayan
Databricks Employee
  • 1 kudos

Hi, To change the default lifetime of 90 days ,you can leave the Lifetime (days) box empty (blank). Refer: https://docs.databricks.com/en/dev-tools/auth/pat.html#databricks-personal-access-tokens-for-workspace-users

  • 1 kudos
Sujitha
by Databricks Employee
  • 7525 Views
  • 1 replies
  • 0 kudos

Exciting Announcement: Launch of New Course - Data Analysis with Databricks!

Welcome to the world of Data Analysis with Databricks! We are thrilled to introduce our latest course, providing a comprehensive journey into data analysis on the Databricks platform. Whether you're a beginner or looking to enhance your skills, this...

Screenshot 2024-01-16 at 1.57.40 PM.png
  • 7525 Views
  • 1 replies
  • 0 kudos
Latest Reply
hthiru
New Contributor II
  • 0 kudos

is this course made available in Home - Databricks Learning?

  • 0 kudos
Phani1
by Databricks MVP
  • 4602 Views
  • 1 replies
  • 2 kudos

Databricks API using the personal access token

We can access the Azure databricks API using the personal access token which is created by us manually.The objective is that client don’t want to store the personal access token which may not be secure .Do we have option to generate the token during ...

  • 4602 Views
  • 1 replies
  • 2 kudos
Latest Reply
Ajay-Pandey
Databricks MVP
  • 2 kudos

Hi @Phani1 ,Yes, now you can use databricks Create a user token API for create access token via automated API call.Please refer below doc - Create a user token | Token API | REST API reference | Azure Databricks

  • 2 kudos
Eldar_Dragomir
by New Contributor II
  • 6234 Views
  • 3 replies
  • 0 kudos

Databricks Volume. Not able to read a file from Scala.

I used to use dbfs with mounted directories and now I want to switch to Volumes for storing my jars and application.conf for pipelines. I see the file my application.conf in Data Explorer > Catalog > Volumes, I also see the file with dbutils.fs.ls("/...

Get Started Discussions
Databricks
Unity Catalog
  • 6234 Views
  • 3 replies
  • 0 kudos
Latest Reply
argus7057
New Contributor II
  • 0 kudos

Volumes mount are accessible using scala code only on a shared cluster. On single user mode this features is not supported yet. We use init scripts to move contents from Volumes to clusters local drive, when we need to access files from Native Scala ...

  • 0 kudos
2 More Replies
ChristianRRL
by Valued Contributor III
  • 6201 Views
  • 2 replies
  • 1 kudos

Resolved! DLT Notebook and Pipeline Separation vs Consolidation

Super basic question. For DLT pipelines I see there's an option to add multiple "Paths". Is it generally best practice to completely separate `bronze` from `silver` notebooks? Or is it more recommended to bundle both raw `bronze` and clean `silver` d...

ChristianRRL_1-1705597040187.png
  • 6201 Views
  • 2 replies
  • 1 kudos
Latest Reply
ChristianRRL
Valued Contributor III
  • 1 kudos

This is great! I completely missed the list view before.

  • 1 kudos
1 More Replies
ChristianRRL
by Valued Contributor III
  • 3764 Views
  • 3 replies
  • 0 kudos

Resolved! Meta Question: Databricks Community Posting Issue

Can someone please tell me what "errors" I need to correct to post??? 

ChristianRRL_0-1704906270658.png
  • 3764 Views
  • 3 replies
  • 0 kudos
Latest Reply
ChristianRRL
Valued Contributor III
  • 0 kudos

Yeah, it worked. I'll need to use the "code sample" more going forward!

  • 0 kudos
2 More Replies
manoj_2355ca
by New Contributor III
  • 2515 Views
  • 1 replies
  • 1 kudos

Azure devops pipeline throwing databricks provider bug when trying to read the metastore

Hi I created Terraform script to add the existing workspace to exisitng unity catlaog by reading the metastore id . When i created pipleina nd tried to use it . This is failing with below error.Planning failed. Terraform encountered an error while ge...

Get Started Discussions
databricks api bug
  • 2515 Views
  • 1 replies
  • 1 kudos
Latest Reply
manoj_2355ca
New Contributor III
  • 1 kudos

Please can anyone provide help me on this ?

  • 1 kudos
salib
by New Contributor II
  • 1810 Views
  • 1 replies
  • 1 kudos

Failed to create Cluster on GCP

I am getting following error while trying to create a Cluster for my Workspace Cluster creation failed: Constraint constraints/compute.disableSerialPortLogging violated for projectCloud ENV is GCP and we can't turn off the constraint mentioned above....

  • 1810 Views
  • 1 replies
  • 1 kudos
Latest Reply
salib
New Contributor II
  • 1 kudos

Hi, Haven't found any solution so far. What I hoping for is to create a cluster in a way it doesn't require SerialPortLogging so that the Policy Constraint we have e.g. disableSerialPortLogging doesn't come in the way. Not sure how can we do that. Ma...

  • 1 kudos
Phani1
by Databricks MVP
  • 8404 Views
  • 0 replies
  • 0 kudos

Cloudera SQL

Hi Team,Could you please help me how efficiently/quickly can convert Cloudera SQL and Hive SQL Scripts to Pyspark Script.Regards,Phanindra

  • 8404 Views
  • 0 replies
  • 0 kudos
marcusfox
by New Contributor
  • 2226 Views
  • 1 replies
  • 0 kudos

Databricks setup with Azure storage

Hi ,We have an issue with our intial setup and design.We are using a single Azure, Premium, Block blob storage account with hierarchical namespace and LRS enabled.We have three containers within it, one for each environment – Dev -Test – ProdBut the ...

  • 2226 Views
  • 1 replies
  • 0 kudos
Latest Reply
Debayan
Databricks Employee
  • 0 kudos

Hi, could you please check https://community.databricks.com/t5/data-governance/metastore-one-per-account-region-limitation/td-p/41097  and let us know if this discussion helps? 

  • 0 kudos
chrisf_sts
by New Contributor II
  • 6706 Views
  • 0 replies
  • 0 kudos

How to handle complex json schema

I have a mounted external directory that is an s3 bucket with multiple subdirectories containing call log files in json format.  The files are irregular and complex, when i try to use spark.read.json or spark.sql (SELECT *) i get the UNABLE_TO_INFER_...

Get Started Discussions
json
pyspark
schema
  • 6706 Views
  • 0 replies
  • 0 kudos
dvmentalmadess
by Valued Contributor
  • 13889 Views
  • 9 replies
  • 1 kudos

Resolved! Terraform databricks_storage_credential has wrong External ID

We create storage credentials using Terraform. I don't see any way to specify a given External ID (DBR Account ID) when creating the credentials via Terraform or in the web UI console. However, today when I tried creating a new set of credentials usi...

Screenshot 2023-11-28 at 6.37.50 PM.png
  • 13889 Views
  • 9 replies
  • 1 kudos
Latest Reply
Mathias_Peters
Contributor II
  • 1 kudos

I tried the proposed solution using an account provider like this provider "databricks" { account_id = "ACCOUNT_ID" host = "https://accounts.cloud.databricks.com" } for creating the storage credential. However, that did not work. I got an e...

  • 1 kudos
8 More Replies

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels