cancel
Showing results for 
Search instead for 
Did you mean: 
Get Started Discussions
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and introductory topics. Connect with beginners and experts alike to kickstart your Databricks experience.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Surajv
by New Contributor III
  • 356 Views
  • 1 replies
  • 0 kudos

Difference between delete token API and revoke token API Databricks

Hi Community, I am trying to understand the difference between:Delete token API: DELETE /api/2.0/token-management/tokens/{token_id}Revoke token API: POST /api/2.0/token/deleteAs, when I create more than 600 tokens - I am getting QUOTA_EXCEEDED error....

  • 356 Views
  • 1 replies
  • 0 kudos
Latest Reply
Surajv
New Contributor III
  • 0 kudos

Delete token API doc link: https://docs.databricks.com/api/workspace/tokenmanagement/deleteRevoke token API doc link: https://docs.databricks.com/api/workspace/tokens/revoketoken 

  • 0 kudos
NC
by New Contributor III
  • 335 Views
  • 1 replies
  • 0 kudos

Using libpostal in Databricks

Hi,I am trying to work on address parsing and would like to use libpostal in Databricks.I have used the official python bindings: GitHub - openvenues/pypostal: Python bindings to libpostal for fast international address parsing/normalizationpip insta...

  • 335 Views
  • 1 replies
  • 0 kudos
Latest Reply
NC
New Contributor III
  • 0 kudos

I managed to install pylibpostal via the Cluster Library. but I cannot seem to download the data needed to run it.Please help. Thank you.

  • 0 kudos
hpicatto
by New Contributor III
  • 531 Views
  • 1 replies
  • 0 kudos

Download event and run logs

how can I download the run and event logs? spark UI is loading them from somewhere but I couldn't find them in dbfs nor on s3

  • 531 Views
  • 1 replies
  • 0 kudos
Latest Reply
Kaniz
Community Manager
  • 0 kudos

Hi @hpicatto,  The Spark Web UI provides a suite of user interfaces (UIs) for monitoring your Spark cluster. You can access it by navigating to http://<driver-node-ip>:18080 in your web browser.On the appropriate application page, click the Download ...

  • 0 kudos
super7admin
by New Contributor
  • 973 Views
  • 2 replies
  • 0 kudos

unable to see AI playground in Machine Learning in Dashboard

unable to see AI playground in Machine Learning in Dashboard

  • 973 Views
  • 2 replies
  • 0 kudos
Latest Reply
Kaniz
Community Manager
  • 0 kudos

Hi @super7admin, Please check this document - https://docs.databricks.com/en/large-language-models/ai-playground.html

  • 0 kudos
1 More Replies
Surajv
by New Contributor III
  • 488 Views
  • 1 replies
  • 0 kudos

What is the quota limit for using create user token api?

Hi Community, I was going through this doc: https://docs.databricks.com/api/workspace/tokens/create to and got to know, that there is a quota limit to how many token one can generate using the api: POST /api/2.0/token/create, having breached the thre...

  • 488 Views
  • 1 replies
  • 0 kudos
Latest Reply
Kaniz
Community Manager
  • 0 kudos

Hi @Surajv, Let’s dive into the details of token quotas in Databricks. Quota Limit for Token Creation: The quota limit for creating user tokens via the API (specifically, using POST /api/2.0/token/create) is essential to manage token usage.Each u...

  • 0 kudos
Surajv
by New Contributor III
  • 412 Views
  • 1 replies
  • 0 kudos

Number of tokens generated for a service principal

Hi community, Is there any API or option to view all PAT tokens generated by a Databricks service principal?

  • 412 Views
  • 1 replies
  • 0 kudos
Latest Reply
Kaniz
Community Manager
  • 0 kudos

Hi @Surajv, When working with Databricks service principals, you can manage and view personal access tokens (PATs) associated with them. Here’s how you can achieve this: What is a Service Principal? A service principal is an identity created in ...

  • 0 kudos
DatabricksGuide
by Community Manager
  • 725 Views
  • 0 replies
  • 0 kudos

Join Our Databricks Free Trial Experience feedback AMA on Friday March 29, 2024!

We're looking for feedback on the Databricks free trial experience, and we need your help! Whether you've used it for data engineering, data science, or analytics, Sujit Nair, our Product Manager on the free trial experience, and our journey archite...

  • 725 Views
  • 0 replies
  • 0 kudos
Еmil
by New Contributor III
  • 2320 Views
  • 3 replies
  • 1 kudos

Resolved! source set to GIT for Databricks Asset Bundle notebook_task - git authentication fails on run

My post was marked as Spam after trying to post the description of my issue so now I have posted the question on stackoverflow.

  • 2320 Views
  • 3 replies
  • 1 kudos
Latest Reply
Kaniz
Community Manager
  • 1 kudos

Hi @Еmil, I've read through your question and believe I have a solution for you. Here's a response to your question: Since your job runs as a service principal, consider using OAuth M2M authentication for accessing your Azure DevOps Git repository.En...

  • 1 kudos
2 More Replies
Frustrated_DE
by New Contributor II
  • 942 Views
  • 1 replies
  • 0 kudos

DLT SQL demo pipeline issue

Hi,   First foray into DLT and following code exerts from the sample-DLT-notebook.I'm creating a notebook with the SQL below:CREATE STREAMING LIVE TABLE sales_orders_rawCOMMENT "The raw sales orders, ingested from /databricks-datasets."TBLPROPERTIES ...

  • 942 Views
  • 1 replies
  • 0 kudos
Latest Reply
Frustrated_DE
New Contributor II
  • 0 kudos

If you change the notebook default language as opposed to using magic command. I normally have it set to Python, I've wrongly assumed DLT would transpose as can't use magic command but have to change default in order for it to work. 

  • 0 kudos
hpicatto
by New Contributor III
  • 1077 Views
  • 3 replies
  • 0 kudos

using the api for getting cost in usd

I'm trying to use the API of billable usage and I do get a report but I have not been able to get the usd cost report, only the dbuHours. I guess I've to change the meter_name but I cannot find the key for that parameter anywhere

  • 1077 Views
  • 3 replies
  • 0 kudos
Latest Reply
Kaniz
Community Manager
  • 0 kudos

Hi @hpicatto, AWS Usage Reports: AWS provides detailed usage and cost reports through the AWS Cost and Usage Report. You can access this report via the AWS Management Console. Here are the steps: Log in to the AWS Management Console.Navigate to the B...

  • 0 kudos
2 More Replies
Floody
by New Contributor II
  • 1076 Views
  • 1 replies
  • 0 kudos

New draft for every post I visit

When I visit my profile page, under the drafts section I see an entry for every post I visit in the discussions. Is this normal?

  • 1076 Views
  • 1 replies
  • 0 kudos
Latest Reply
Kaniz
Community Manager
  • 0 kudos

Hi @Floody, Yes, it is normal to see an entry for every post you visit in the discussions under the drafts section of your profile page. This feature allows you to easily access and continue working on drafts of posts that you have started or viewed ...

  • 0 kudos
NoviKamayana
by New Contributor
  • 1132 Views
  • 1 replies
  • 1 kudos

Database: Delta Lake or PostgreSQL

Hey all,I am searching for a non-political answer to my database questions. Please know that I am a data newbie and litteraly do not know anything about this topic, but I want to learn, so please be gentle.  Some context: I am working for an OEM that...

  • 1132 Views
  • 1 replies
  • 1 kudos
Latest Reply
Kaniz
Community Manager
  • 1 kudos

Hi @NoviKamayana, Let’s dive into the world of data storage and explore the differences between Delta Lake and PostgreSQL. As a data newbie, you’re on the right track to learn more about these concepts. Data Lake vs. Delta Lake Benefits and Limit...

  • 1 kudos
mohaimen_syed
by New Contributor III
  • 3691 Views
  • 5 replies
  • 1 kudos

Fuzzy Match on PySpark using UDF/Pandas UDF

I'm trying to do fuzzy matching on two dataframes by cross joining them and then using a udf for my fuzzy matching. But using both python udf and pandas udf its either very slow or I get an error. @pandas_udf("int")def core_match_processor(s1: pd.Ser...

  • 3691 Views
  • 5 replies
  • 1 kudos
Latest Reply
Kaniz
Community Manager
  • 1 kudos

Hi @mohaimen_syed, Could you please help me with these details:-  - Cluster details, and - Check if Apache Apache Arrow optimization is enabled in your cluster.

  • 1 kudos
4 More Replies
ntvdatabricks
by New Contributor II
  • 1476 Views
  • 2 replies
  • 1 kudos

Resolved! Okta and Unified login

Hey Folks anyone put Databricks behind Okta and enabled Unified Login with workspaces that have a Unity Catalog metastore applied and some that don't?There are some workspaces we can't move over yet and it isn't clear in documentation if Unity Catalo...

  • 1476 Views
  • 2 replies
  • 1 kudos
Latest Reply
Walter_C
Valued Contributor II
  • 1 kudos

Yes, users should be able to use a single Okta application for all workspaces, regardless of whether the Unity Catalog metastore has been applied or not. The Unity Catalog is a feature that allows you to manage and secure access to your data across a...

  • 1 kudos
1 More Replies
Khalil
by Contributor
  • 3516 Views
  • 5 replies
  • 7 kudos

Incremental ingestion of Snowflake data with Delta Live Table (CDC)

Hello,I have some data which are lying into Snowflake, so I want to apply CDC on them using delta live table but I am having some issues.Here is what I am trying to do:  @dlt.view() def table1(): return spark.read.format("snowflake").options(**opt...

  • 3516 Views
  • 5 replies
  • 7 kudos
Latest Reply
-werners-
Esteemed Contributor III
  • 7 kudos

The CDC for delta live works fine for delta tables, as you have noticed.  However it is not a full blown CDC implementation/software.If you want to capture changes in Snowflake, you will have to implement some CDC method on Snowflake itself, and read...

  • 7 kudos
4 More Replies
Labels
Top Kudoed Authors