- 11123 Views
- 4 replies
- 0 kudos
My notebooks running in parallel no longer include jobids
Hey Databricks - what happened to the jobids that used to be returned from parallel runs? We used them to identify which link matched the output. See attached. How are we supposed to match up the links?
- 11123 Views
- 4 replies
- 0 kudos
- 0 kudos
Hey @Databricks, @Retired_mod - waccha doing? Yesterday's "newer version of the app" that got rolled out seems to have broken the parallel runs. The ephemeral notebook is missing. The job ids are missing. What's up? Benedetta
- 0 kudos
- 2075 Views
- 0 replies
- 0 kudos
Unable to build LZO-codec
Hi Community i am try to create lzo-codec in my dbfs using:https://docs.databricks.com/en/_extras/notebooks/source/init-lzo-compressed-files.htmlbut i am facing the errorCloning into 'hadoop-lzo'... The JAVA_HOME environment variable is not defined c...
- 2075 Views
- 0 replies
- 0 kudos
- 2185 Views
- 1 replies
- 1 kudos
Looking for a promo code for Databricks certification
I went through the webinar suggested and the courses mentioned, but I have not received any voucher code for the certification. Can anyone please help. Thank you so much.
- 2185 Views
- 1 replies
- 1 kudos
- 1 kudos
Thank you. The link is broken in the response for ticketing portal. I have added a ticket with the help center. Kindly let me know if anything else is needed. Appreciate the help.
- 1 kudos
- 2739 Views
- 1 replies
- 0 kudos
LLM Chatbot With Retrieval Augmented Generation (RAG)
When executing the second block under 01-Data-Preparation-and-IndexI get the following error. Please help. AnalysisException: [RequestId=c9625879-339d-45c6-abb5-f70d724ddb47 ErrorClass=INVALID_STATE] Metastore storage root URL does not exist.Please p...
- 2739 Views
- 1 replies
- 0 kudos
- 0 kudos
I fixed this issue by providing a MANAGED LOCATION for the catalog - meant updating the _resources/00-init file as follows - spark.sql(f"CREATE CATALOG IF NOT EXISTS {catalog} MANAGED LOCATION '<location path>'")
- 0 kudos
- 5580 Views
- 3 replies
- 0 kudos
Resolved! My Global Init Script doesn't run with the new(er) LTS version 13.3. It runs great on 12.2LTS
Hey Databricks, Seems like you changed the way Global Init Scripts work. How come you changed it? My Global Init Script runs great on 12.2LTS but not on the new(er) LTS version 13.3. We don't have Unity Catalog turned on. What's up with that? Ar...
- 5580 Views
- 3 replies
- 0 kudos
- 0 kudos
Thank you ChloeBors - I tried upgrading the ubuntu version per your suggestion but got a new error "Can t open lib ODBC Driver 17 for SQL Server file not found 0 SQLDriverConnect". I tried modifying this line to 18: ACCEPT_EULA=Y apt-get install mso...
- 0 kudos
- 4722 Views
- 0 replies
- 0 kudos
How to save a catalog table as a spark or pandas dataframe?
HelloI have a table in my catalog, and I want to have it as a pandas or spark df. I was using this code to do that before, but I don't know what is happened recently that the code is not working anymore. from pyspark.sql import SparkSession spark = S...
- 4722 Views
- 0 replies
- 0 kudos
- 3141 Views
- 2 replies
- 0 kudos
A problem has occured
I am using the Community version of Databricks, and even though how much I clear and delete my items, notebooks, the below message appear when I want to insert a new cell in the notebook. Where can I see my usage and quota? A problem has occuredServe...
- 3141 Views
- 2 replies
- 0 kudos
- 0 kudos
sorry, I do not see any Trash button on my side. Can you printscreen?
- 0 kudos
- 1811 Views
- 1 replies
- 1 kudos
Model logging issues - Azure Databricks
I am running job with service principal ( principal e1077c1b-1dcf-4975-8164-d14f748d4774). Mentioned job has a task that will log (mlflow.pyfunc.log_model) newly trained model to unity catalog, but I get PERMISSION DENIED error. I do not get why prin...
- 1811 Views
- 1 replies
- 1 kudos
- 3645 Views
- 2 replies
- 3 kudos
Resolved! Can we get the all notebooks attached to an all purpose compute at a particular instant ?
For an all purpose compute, we can see the notebooks attached to a cluster in the UI. Is there a way to get the same details using some api or other method in program.
- 3645 Views
- 2 replies
- 3 kudos
- 1408 Views
- 1 replies
- 0 kudos
Difference between delete token API and revoke token API Databricks
Hi Community, I am trying to understand the difference between:Delete token API: DELETE /api/2.0/token-management/tokens/{token_id}Revoke token API: POST /api/2.0/token/deleteAs, when I create more than 600 tokens - I am getting QUOTA_EXCEEDED error....
- 1408 Views
- 1 replies
- 0 kudos
- 0 kudos
Delete token API doc link: https://docs.databricks.com/api/workspace/tokenmanagement/deleteRevoke token API doc link: https://docs.databricks.com/api/workspace/tokens/revoketoken
- 0 kudos
- 1695 Views
- 1 replies
- 0 kudos
Urgent: Issue with Databricks Certified Machine Learning Engineer Associate Exam
Hello Team,I encountered an issue during my Databricks Certified Machine Learning Engineer Associate exam, which took place on 30- 03- 2024, at 11:00 AM IST. Despite being present at my seat and viewing the camera, the proctor unexpectedly suspended...
- 1695 Views
- 1 replies
- 0 kudos
- 0 kudos
@himashu-singh thank you for filing at ticket with the support team, it looks like they have already responded to you.
- 0 kudos
- 2040 Views
- 0 replies
- 0 kudos
How to enable custom metrics for databricks EC2 instance?
Case 1: I have an AWS Databricks Instance, followed some steps for enabling CloudWatch for the Databricks EC2 instance. However, memory metrics is not available in the cloud watch. For that followed below steps for enabling custom metrics for the EC2...
- 2040 Views
- 0 replies
- 0 kudos
- 3885 Views
- 1 replies
- 0 kudos
How to install and use tkinter in databricks notebooks?
I have access to databricks cloud UI. I use the notebooks for my experiments (not MLFlow).I install libraries using the notebooks by using this command - !pip install LIBRARY_NAMEI am unable to install tkinter and use it in notebook as. Is there a wa...
- 3885 Views
- 1 replies
- 0 kudos
- 0 kudos
Hi @Retired_mod Thanks for your response. I believe the "Create Library" is depreciated. Although there is a way to install libraries in the cluster configurations which I found and can try to install libraries through there. After using that, still ...
- 0 kudos
- 2348 Views
- 1 replies
- 0 kudos
Using libpostal in Databricks
Hi,I am trying to work on address parsing and would like to use libpostal in Databricks.I have used the official python bindings: GitHub - openvenues/pypostal: Python bindings to libpostal for fast international address parsing/normalizationpip insta...
- 2348 Views
- 1 replies
- 0 kudos
- 0 kudos
I managed to install pylibpostal via the Cluster Library. but I cannot seem to download the data needed to run it.Please help. Thank you.
- 0 kudos
- 1847 Views
- 1 replies
- 0 kudos
Exam suspended
Hi Team,My databricks exam got suspended on 28h march 2024 and it is still in the suspended state. I have raised a support request using the below linkhttps://help.databricks.com/s/contact-us?ReqType=trainingand my request Id is #00454905I would real...
- 1847 Views
- 1 replies
- 0 kudos
Join Us as a Local Community Builder!
Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!
Sign Up Now-
.CSV
1 -
Access Data
2 -
Access Databricks
3 -
Access Delta Tables
2 -
Account reset
1 -
ADF Pipeline
1 -
ADLS Gen2 With ABFSS
1 -
Advanced Data Engineering
2 -
AI
4 -
Analytics
1 -
Apache spark
1 -
Apache Spark 3.0
1 -
Api Calls
1 -
API Documentation
3 -
App
2 -
Application
1 -
Architecture
1 -
asset bundle
1 -
Asset Bundles
3 -
Auto-loader
1 -
Autoloader
4 -
Aws databricks
1 -
AWS security token
1 -
AWSDatabricksCluster
1 -
Azure
6 -
Azure data disk
1 -
Azure databricks
15 -
Azure Databricks SQL
6 -
Azure databricks workspace
1 -
Azure Unity Catalog
6 -
Azure-databricks
1 -
AzureDatabricks
1 -
AzureDevopsRepo
1 -
Big Data Solutions
1 -
Billing
1 -
Billing and Cost Management
2 -
Blackduck
1 -
Bronze Layer
1 -
Certification
3 -
Certification Exam
1 -
Certification Voucher
3 -
CICDForDatabricksWorkflows
1 -
Cloud_files_state
1 -
CloudFiles
1 -
Cluster
3 -
Cluster Init Script
1 -
Comments
1 -
Community Edition
3 -
Community Event
1 -
Community Group
2 -
Community Members
1 -
Compute
3 -
Compute Instances
1 -
conditional tasks
1 -
Connection
1 -
Contest
1 -
Credentials
1 -
Custom Python
1 -
CustomLibrary
1 -
Data
1 -
Data + AI Summit
1 -
Data Engineer Associate
1 -
Data Engineering
3 -
Data Explorer
1 -
Data Ingestion & connectivity
1 -
Data Processing
1 -
Databrick add-on for Splunk
1 -
databricks
2 -
Databricks Academy
1 -
Databricks AI + Data Summit
1 -
Databricks Alerts
1 -
Databricks App
1 -
Databricks Assistant
1 -
Databricks Certification
1 -
Databricks Cluster
2 -
Databricks Clusters
1 -
Databricks Community
10 -
Databricks community edition
3 -
Databricks Community Edition Account
1 -
Databricks Community Rewards Store
3 -
Databricks connect
1 -
Databricks Dashboard
3 -
Databricks delta
2 -
Databricks Delta Table
2 -
Databricks Demo Center
1 -
Databricks Documentation
4 -
Databricks genAI associate
1 -
Databricks JDBC Driver
1 -
Databricks Job
1 -
Databricks Lakehouse Platform
6 -
Databricks Migration
1 -
Databricks Model
1 -
Databricks notebook
2 -
Databricks Notebooks
4 -
Databricks Platform
2 -
Databricks Pyspark
1 -
Databricks Python Notebook
1 -
Databricks Repo
1 -
Databricks Runtime
1 -
Databricks SQL
5 -
Databricks SQL Alerts
1 -
Databricks SQL Warehouse
1 -
Databricks Terraform
1 -
Databricks UI
1 -
Databricks Unity Catalog
4 -
Databricks Workflow
2 -
Databricks Workflows
2 -
Databricks workspace
3 -
Databricks-connect
1 -
databricks_cluster_policy
1 -
DatabricksJobCluster
1 -
DataCleanroom
1 -
DataDays
1 -
Datagrip
1 -
DataMasking
2 -
DataVersioning
1 -
dbdemos
2 -
DBFS
1 -
DBRuntime
1 -
DBSQL
1 -
DDL
1 -
Dear Community
1 -
deduplication
1 -
Delt Lake
1 -
Delta Live Pipeline
3 -
Delta Live Table
5 -
Delta Live Table Pipeline
5 -
Delta Live Table Pipelines
4 -
Delta Live Tables
7 -
Delta Sharing
2 -
deltaSharing
1 -
Deny assignment
1 -
Development
1 -
Devops
1 -
DLT
10 -
DLT Pipeline
7 -
DLT Pipelines
5 -
Dolly
1 -
Download files
1 -
Dynamic Variables
1 -
Engineering With Databricks
1 -
env
1 -
ETL Pipelines
1 -
External Sources
1 -
External Storage
2 -
FAQ for Databricks Learning Festival
2 -
Feature Store
2 -
Filenotfoundexception
1 -
Free Edition
1 -
Free trial
1 -
GCP Databricks
1 -
GenAI
1 -
Getting started
2 -
Google Bigquery
1 -
HIPAA
1 -
Hubert Dudek
2 -
import
2 -
Integration
1 -
JDBC Connections
1 -
JDBC Connector
1 -
Job Task
1 -
JSON Object
1 -
Learning
1 -
Lineage
1 -
LLM
1 -
Login
1 -
Login Account
1 -
Machine Learning
3 -
MachineLearning
1 -
Materialized Tables
2 -
Medallion Architecture
1 -
meetup
2 -
Metadata
1 -
Migration
1 -
ML Model
2 -
MlFlow
2 -
Model Training
1 -
Module
1 -
Monitoring
1 -
Networking
2 -
Notebook
1 -
Onboarding Trainings
1 -
OpenAI
1 -
Pandas udf
1 -
Permissions
1 -
personalcompute
1 -
Pipeline
2 -
Plotly
1 -
PostgresSQL
1 -
Pricing
1 -
Pyspark
1 -
Python
5 -
Python Code
1 -
Python Wheel
1 -
Quickstart
1 -
Read data
1 -
Repos Support
1 -
Reset
1 -
Rewards Store
2 -
Sant
1 -
Schedule
1 -
Serverless
3 -
serving endpoint
1 -
Session
1 -
Sign Up Issues
2 -
Software Development
1 -
Spark Connect
1 -
Spark scala
1 -
sparkui
2 -
Speakers
1 -
Splunk
2 -
SQL
8 -
Summit23
7 -
Support Tickets
1 -
Sydney
2 -
Table Download
1 -
Tags
3 -
terraform
1 -
Training
2 -
Troubleshooting
1 -
Unity Catalog
4 -
Unity Catalog Metastore
2 -
Update
1 -
user groups
1 -
Venicold
3 -
Vnet
1 -
Voucher Not Recieved
1 -
Watermark
1 -
Weekly Documentation Update
1 -
Weekly Release Notes
2 -
Women
1 -
Workflow
2 -
Workspace
3
- « Previous
- Next »
| User | Count |
|---|---|
| 133 | |
| 129 | |
| 57 | |
| 42 | |
| 42 |