cancel
Showing results for 
Search instead for 
Did you mean: 
Get Started Discussions
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and introductory topics. Connect with beginners and experts alike to kickstart your Databricks experience.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

thethirtyfour
by New Contributor III
  • 3707 Views
  • 2 replies
  • 1 kudos

Resolved! Configure Databricks in VSCode through WSL

Hi,I am having a hard time configuring my Databricks workspace when working in VSCode via WSL. When following the steps to setup Databricks authentication I am receiving the following error on the Step 5 of "Step 4: Set up Databricks authentication"....

  • 3707 Views
  • 2 replies
  • 1 kudos
Latest Reply
RaulMoraM
New Contributor II
  • 1 kudos

What worked for me was NOT opening the browser using the pop-up (which generated the 3-legged-OAuth flow error), but clicking on the link provided by the CLI (or copy paste the link on the browser)

  • 1 kudos
1 More Replies
JohnJustus
by New Contributor III
  • 8447 Views
  • 3 replies
  • 0 kudos

Space in Column names when writing to Hive

All,I have the following code.df_Warehouse_Utilization = (    spark.table("hive_metastore.dev_ork.bin_item_detail")    .join(df_DIM_Bins,col('bin_tag')==df_DIM_Bins.BinKey,'right')    .groupby(col('BinKey'))    .agg(count_distinct(when(col('serial_lo...

  • 8447 Views
  • 3 replies
  • 0 kudos
Latest Reply
KandyKad
New Contributor III
  • 0 kudos

Hi,I have faced this issue a few times. When we are overwriting the dataframes to hive catalog in databricks, it doesn't naturally allow for column names to have spaces or special characters. However, you can add an option statement to bypass that ru...

  • 0 kudos
2 More Replies
EricMa
by New Contributor III
  • 5029 Views
  • 18 replies
  • 4 kudos

Mounting Data IOException

Hello,I am currently taking a course from Coursera for data science using SQL. For one of our assignments we need to mount some data by running a script that has been provided to us by the class. When I run the script I receive the following error. I...

IOException.jpg IOException_Classroom-Setup.jpg
  • 5029 Views
  • 18 replies
  • 4 kudos
Latest Reply
YunHai
New Contributor II
  • 4 kudos

I have resolved this issue. Please see the Databricks notebook below. Thanks to Tawfeeq for the helpful method.https://databricks-prod-cloudfront.cloud.databricks.com/public/4027ec902e239c93eaaa8714f173bcfc/2108650195107345/4043517972039275/874892524...

  • 4 kudos
17 More Replies
NagarajuT
by New Contributor
  • 661 Views
  • 1 replies
  • 0 kudos

Connect to SQL Developer using Custom JDBC

Hello,I'm trying to connect databricks SQL to the SQL Developer using custom JDBC.I'm getting errorjdbc:databricks:<server>:443;HttpPath=<HttpPath>;UID=token;PWD=<password> RegardsNaga

  • 661 Views
  • 1 replies
  • 0 kudos
Latest Reply
ShaliniC
New Contributor II
  • 0 kudos

Hi,We are trying to test if we can connect sql developer to databricks. did it work for you?Regards, shalini

  • 0 kudos
tyorisoo
by New Contributor III
  • 707 Views
  • 6 replies
  • 0 kudos

Unity Catalog About Metastore

Registered on 2024/10 from AWS marketplace.We have created a customer management VPC and manually created the workspace.No specific metastore settings were made when the workspace was created.In the catalog screen of the account console,unity catalog...

  • 707 Views
  • 6 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

Hello @tyorisoo,I hope you are doing well!Metastore manages metadata, not catalog information, schema information, table information, function information, access control information, etc. In the current state, the metastore configuration is not done...

  • 0 kudos
5 More Replies
chris0991
by New Contributor III
  • 580 Views
  • 3 replies
  • 0 kudos

Best practices for optimizing Spark jobs

What are some best practices for optimizing Spark jobs in Databricks, especially when dealing large datasets? Any tips or resources would be greatly appreciated! I’m trying to analyze data on restaurant menu prices so that insights would be especiall...

  • 580 Views
  • 3 replies
  • 0 kudos
Latest Reply
mo4
New Contributor II
  • 0 kudos

 Optimizing Spark jobs in Databricks can significantly enhance performance. Here are some strategies to consider:Efficient Partitioning: Proper partitioning reduces shuffle times, leading to faster data processing.Caching: Utilize Delta caching inste...

  • 0 kudos
2 More Replies
Surajv
by New Contributor III
  • 1414 Views
  • 2 replies
  • 1 kudos

What is the quota limit for using create user token api?

Hi Community, I was going through this doc: https://docs.databricks.com/api/workspace/tokens/create to and got to know, that there is a quota limit to how many token one can generate using the api: POST /api/2.0/token/create, having breached the thre...

  • 1414 Views
  • 2 replies
  • 1 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 1 kudos

Hello@Surajv, Q1: What is and how to find out the quota limit? The quota limit for creating user tokens via the API (POST /api/2.0/token/create) is essential to manage token usage. Each user can have multiple personal access tokens in a Databricks wo...

  • 1 kudos
1 More Replies
benito
by New Contributor
  • 435 Views
  • 1 replies
  • 0 kudos

Databricks Initial Costs AWS

I have a new premium account. I set up a cost dashboard (see attached) after I create a new workspace using AWS Quickstart, where I see some costs. Why do I have this If I am not using Databricks at all? How can I save the costs?

databricks.png
  • 435 Views
  • 1 replies
  • 0 kudos
Latest Reply
Walter_C
Databricks Employee
  • 0 kudos

Are you seeing this data from the Usage tab in the Account console? Does it allow you to filter it by SKU?

  • 0 kudos
Phani1
by Valued Contributor II
  • 9023 Views
  • 4 replies
  • 2 kudos

Masking techniques for more PII columns

Hi Databricks Team,We would appreciate it if you could inform us about the situations when Column-Masking, Row-Level Filtering, and Attributed-Based Masking should be utilized, as well as the recommended technique for handling large data volumes cont...

  • 9023 Views
  • 4 replies
  • 2 kudos
Latest Reply
NandiniN
Databricks Employee
  • 2 kudos

Agree with @Meghla-C , https://databricks.aha.io/ideas/ideas/DB-I-7941 this was the feature request and if you see the status, it is in preview.

  • 2 kudos
3 More Replies
NC
by New Contributor III
  • 3337 Views
  • 4 replies
  • 0 kudos

GDAL on Databricks Cluster Runtime 12.2 LTS

I need gdal in my course work.After reading this post, I used init script as follows to install gdal into runtime 12.2 LTS  dbutils.fs.put("/databricks/scripts/gdal_install.sh",""" #!/bin/bash sudo add-apt-repository ppa:ubuntugis/ppa sudo apt-get up...

  • 3337 Views
  • 4 replies
  • 0 kudos
Latest Reply
Matt_C
New Contributor II
  • 0 kudos

Hi, in case anyone is still struggling here. I found I could not get the init script approach to work, but if I just run a shell command to install gdal at the start of my notebook it works fine. You might note, however, that this installs gdal versi...

  • 0 kudos
3 More Replies
juliemoore
by New Contributor
  • 582 Views
  • 1 replies
  • 0 kudos

Problems with Big Data Solutions and Databricks- Any advice?

Hello everyone,I am currently facing several challenges related to big data solutions, particularly with the Databricks. As many of you may know, Databricks is a powerful platform for data engineering and analytics, but I have encountered some signif...

  • 582 Views
  • 1 replies
  • 0 kudos
Latest Reply
gchandra
Databricks Employee
  • 0 kudos

Your problem statement is too generic. If your company is facing this, you can reach out to your SA; they will help you. If it's a personal project, then mention what you are trying in detail with cluster size, what you are trying to integrate with, ...

  • 0 kudos
JolM
by New Contributor II
  • 648 Views
  • 1 replies
  • 1 kudos

Resolved! is there a way for us to see billing usage per catalog?

is there a way for us to see billing usage per catalog? I'm using 14days trial period for now.. would it be available in Premium?

  • 648 Views
  • 1 replies
  • 1 kudos
Latest Reply
gchandra
Databricks Employee
  • 1 kudos

Billing system table provides cost by notebook, jobs, and clusters. If catalog-to-job/cluster/notebook relation is maintained, then catalog-based usage can be determined. https://docs.databricks.com/en/admin/system-tables/billing.html

  • 1 kudos
hetrasol
by New Contributor III
  • 2372 Views
  • 7 replies
  • 0 kudos

Resolved! Unable to start browser for databricks certification

Hello, I  have registered for databricks certified data engineering associate exam. One of the requirements to give the exam is The exam is set for Sunday 6th October, 2024 but the browser installation (psi secure bridge browser) does not work. .Reac...

hetrasol_0-1728079492073.png
  • 2372 Views
  • 7 replies
  • 0 kudos
Latest Reply
TaiNguyen
New Contributor II
  • 0 kudos

Hi @hetrasol ,I'm a Windows user. After installation, I just got the Lockdown Browser OEM instead of the PSI browser, as you mentioned above. Can you help to instruct again on how to install these browsers

  • 0 kudos
6 More Replies

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.

If there isn’t a group near you, start one and help create a community that brings people together.

Request a New Group
Labels
Top Kudoed Authors