- 8392 Views
- 2 replies
- 0 kudos
Lab for generative-ai-engineering-with-databricks.dbc doesn't work
It keeps on failing with this error:HTTPError: 503 Server Error: Service Unavailable for url: https://community.cloud.databricks.com/api/2.0/feature-store/feature-tables/search?max_results=200 Response from server: { 'error_code': 'TEMPORARILY_UNAVAI...
- 8392 Views
- 2 replies
- 0 kudos
- 0 kudos
Any update with this issue?@abhi-verma have you received some support?
- 0 kudos
- 942 Views
- 0 replies
- 2 kudos
Customizing class imbalance handling in databricks.automl
Hi Databricks! I'd like to make a feature request for end users to be able to customize class imbalance behavior in `databricks.automl`, specifically `databricks.automl.classify`. This public documentation describes the default procedure, but my team...
- 942 Views
- 0 replies
- 2 kudos
- 1555 Views
- 1 replies
- 0 kudos
Bug: MLflow connection fails after 2d
Hi there, After exactly 2d of training, the following error is raised after an API call to MLflow: ValueError: Enum ErrorCode has no value defined for name '403' --------------------------------------------------------------------------- ValueError ...
- 1555 Views
- 1 replies
- 0 kudos
- 0 kudos
Hello @Alex42 !The error message indicates that access is forbidden due to an expired access token. This occurs when a notebook or job runs for an extended period, exceeding the default 48-hour threshold set for security reasons. The Databricks acces...
- 0 kudos
- 3130 Views
- 2 replies
- 0 kudos
Deploy ML
How deploy a model using community account version?
- 3130 Views
- 2 replies
- 0 kudos
- 0 kudos
From your workspace page are you able to see the option to create Models?
- 0 kudos
- 2797 Views
- 4 replies
- 0 kudos
AssertionError Failed to create the catalog
getting this error when trying to setup the get-started-with-databricks-for-machine-learning LAB . Unity catalog is enabled. Validating the locally installed datasets: | listing local files...(0 seconds) | validation completed...(0 seconds total) C...
- 2797 Views
- 4 replies
- 0 kudos
- 0 kudos
EDIT: I found the issue in my Databricks metastore was that it did not have a default storage location set. The academy setup scripts don't explicitly set MANAGED LOCATION and therefore without a default location it fails.I tried to establish a defau...
- 0 kudos
- 1602 Views
- 1 replies
- 0 kudos
Does Databricks Container Services (DCS) support for GPU containers with Databricks Runtime 11.3 LTS and higher?
I have been trying to start a cluster using DCS with GPU containers (https://github.com/databricks/containers/tree/master/ubuntu/gpu), but was only successful with Databricks Runtime 10.4 LTS and lower. With Databricks Runtime 11.3 LTS and higher, I ...
- 1602 Views
- 1 replies
- 0 kudos
- 0 kudos
Hello @ppang ! Since you posted your question, the repository you shared has received an update, which includes the following warning: "Using conda in DCS images is no longer supported starting Databricks Runtime 9.0. We highly recommend users to ext...
- 0 kudos
- 1956 Views
- 1 replies
- 0 kudos
Generative AI
I am working on generative AI. Does Databricks offer databases for storing vector embeddings, if yes, how efficient are they.
- 1956 Views
- 1 replies
- 0 kudos
- 0 kudos
Hello @RA_DATA !Yes, Databricks provides the Databricks Vector Search solution, which is a vector database optimized for storing and retrieving embeddings. With Vector Search, you can create a vector search index from a Delta table, which includes em...
- 0 kudos
- 2931 Views
- 1 replies
- 0 kudos
RuntimeError: Expected to mark a variable ready only once error
I'm using a Single Node machine with g5-2x-large to fine tune a LLaMa-2 model. My Come Notebook runs very smoothly on Google Col but when I try to run it on `Databricks`, it throws me the exact error given below:RuntimeError: Expected to mark a varia...
- 2931 Views
- 1 replies
- 0 kudos
- 0 kudos
Hello @saleem_shady! Have you tried including the parameter ddp_find_unused_parameters=False in your TrainingArguments? Here's an example of how to include it: https://github.com/databricks/databricks-ml-examples/blob/master/llm-models/llamav2/llamav...
- 0 kudos
- 1118 Views
- 1 replies
- 0 kudos
Specify extras of library in asset bundles yml file
Hello,I am using asset bundles to build, deploy, create a cluster and install a package in it. The package comes with some extra dependencies that I also want to include in the cluster. Do you know how to specify this extras in the databricks.yml fil...
- 1118 Views
- 1 replies
- 0 kudos
- 0 kudos
Hello @pablobd! Have you tried declaring these additional dependencies of your wheel file before the wheel file itself? Here's an example snippet for clarification: resources: jobs: train: name: train-ml tasks: - task_key: train-ml ...
- 0 kudos
- 1239 Views
- 3 replies
- 0 kudos
How to I select an 80/10/10 split when doing AutoML
Headline says it all. I am doing a regression and want to select a testvaltrain split that is not 60/20/20. Anyone know how to do this?
- 1239 Views
- 3 replies
- 0 kudos
- 0 kudos
You'd need to put 80% of your data with the earliest timestamp, then 10% with another one and 10% with another.
- 0 kudos
- 1282 Views
- 0 replies
- 1 kudos
Machine Learning Practitioner learning Plan Notebook demos
I am enrolled on the Machine Learning Practitioner learning Plan free version, I can't get the notebook demos to run on databricks community edition. How can I do the demo practices of these courses? Is there another alternative?
- 1282 Views
- 0 replies
- 1 kudos
- 1684 Views
- 1 replies
- 0 kudos
Machine learning accuracy depends on execution plans
I'm using Databricks for a machine learning project -- a fairly standard text classification problem, where I want to use the description of an item (i.e. AXELTTNING KOLKERAMIK MM) to predict which of n product categories the item belongs to ('Bushin...
- 1684 Views
- 1 replies
- 0 kudos
- 0 kudos
that is weird.The regression algorithm should just do a prediction on a dataframe. Such a huge difference in accuracy seems very suspicious.I would test the algorithm on a reference dataset, for which you know the accuracy beforehand.Perhaps your tr...
- 0 kudos
- 4170 Views
- 3 replies
- 0 kudos
java.lang.ClassNotFoundException: com.johnsnowlabs.nlp.DocumentAssembler
I am trying to serve a pyspark model using an endpoint. I was able to load and register the model normally. I could also load that model and perform inference but while serving the model, I am getting the following error: [94fffqts54] ERROR StatusLog...
- 4170 Views
- 3 replies
- 0 kudos
- 0 kudos
I'm having the same problem and have tried various solutions with no luck. I found some potentially relevant information on the following link: https://www.johnsnowlabs.com/serving-spark-nlp-via-api-3-3-databricks-jobs-and-mlflow-serve-apis/ In the ...
- 0 kudos
- 1530 Views
- 1 replies
- 0 kudos
Notebook cell gets hung up but code completes
Have been running into an issue when running a pymc-marketing model in a Databricks notebook. The cell that fits the model gets hung up and the progress bar stops moving, however the code completes and dumps all needed output into a folder. After the...
- 1530 Views
- 1 replies
- 0 kudos
- 0 kudos
Hi @Retired_mod , Thanks for the feedback here as well as on the other discussion forum. I've commented on your trouble shooting tips on that board. One thing to touch upon hereVerify that the output files are correct and contain the expected results...
- 0 kudos
- 1384 Views
- 0 replies
- 0 kudos
Content Type error legacy serving
Hi,I have deployed an endpoint in Databricks using legacy serving. I am using the custom pyfunc in mlflow to deploy the custom code. This code uses Machine Learning to parse out the table of contents in some pdf files then returns the table of conten...
- 1384 Views
- 0 replies
- 0 kudos
Join Us as a Local Community Builder!
Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!
Sign Up Now-
Access control
3 -
Access Data
2 -
AccessKeyVault
1 -
ADB
2 -
Airflow
1 -
Amazon
2 -
Apache
1 -
Apache spark
3 -
APILimit
1 -
Artifacts
1 -
Audit
1 -
Autoloader
6 -
Autologging
2 -
Automation
2 -
Automl
32 -
AWS
7 -
Aws databricks
1 -
AWSSagemaker
1 -
Azure
32 -
Azure active directory
1 -
Azure blob storage
2 -
Azure data lake
1 -
Azure Data Lake Storage
3 -
Azure data lake store
1 -
Azure databricks
32 -
Azure event hub
1 -
Azure key vault
1 -
Azure sql database
1 -
Azure Storage
2 -
Azure synapse
1 -
Azure Unity Catalog
1 -
Azure vm
1 -
AzureML
2 -
Bar
1 -
Beta
1 -
Better Way
1 -
BI Integrations
1 -
BI Tool
1 -
Billing and Cost Management
1 -
Blob
1 -
Blog
1 -
Blog Post
1 -
Broadcast variable
1 -
Business Intelligence
1 -
CatalogDDL
1 -
Centralized Model Registry
1 -
Certification
2 -
Certification Badge
1 -
Change
1 -
Change Logs
1 -
Chatgpt
2 -
Check
2 -
Classification Model
1 -
Cloud Storage
1 -
Cluster
10 -
Cluster policy
1 -
Cluster Start
1 -
Cluster Termination
2 -
Clustering
1 -
ClusterMemory
1 -
CNN HOF
1 -
Column names
1 -
Community Edition
1 -
Community Edition Password
1 -
Community Members
1 -
Company Email
1 -
Condition
1 -
Config
1 -
Configure
3 -
Confluent Cloud
1 -
Container
2 -
ContainerServices
1 -
Control Plane
1 -
ControlPlane
1 -
Copy
1 -
Copy into
2 -
CosmosDB
1 -
Cost
1 -
Courses
2 -
Csv files
1 -
Dashboards
1 -
Data
8 -
Data Engineer Associate
1 -
Data Engineer Certification
1 -
Data Explorer
1 -
Data Ingestion
2 -
Data Ingestion & connectivity
11 -
Data Quality
1 -
Data Quality Checks
1 -
Data Science & Engineering
2 -
databricks
5 -
Databricks Academy
3 -
Databricks Account
1 -
Databricks AutoML
9 -
Databricks Cluster
3 -
Databricks Community
5 -
Databricks community edition
4 -
Databricks connect
1 -
Databricks dbfs
1 -
Databricks Feature Store
1 -
Databricks Job
1 -
Databricks Lakehouse
1 -
Databricks Mlflow
4 -
Databricks Model
2 -
Databricks notebook
10 -
Databricks ODBC
1 -
Databricks Platform
1 -
Databricks Pyspark
1 -
Databricks Python Notebook
1 -
Databricks Runtime
9 -
Databricks SQL
8 -
Databricks SQL Permission Problems
1 -
Databricks Terraform
1 -
Databricks Training
2 -
Databricks Unity Catalog
1 -
Databricks V2
1 -
Databricks version
1 -
Databricks Workflow
2 -
Databricks Workflows
1 -
Databricks workspace
2 -
Databricks-connect
1 -
DatabricksContainer
1 -
DatabricksML
6 -
Dataframe
3 -
DataSharing
1 -
Datatype
1 -
DataVersioning
1 -
Date Column
1 -
Dateadd
1 -
DB Notebook
1 -
DB Runtime
1 -
DBFS
5 -
DBFS Rest Api
1 -
Dbt
1 -
Dbu
1 -
DDL
1 -
DDP
1 -
Dear Community
1 -
DecisionTree
1 -
Deep learning
4 -
Default Location
1 -
Delete
1 -
Delt Lake
4 -
Delta
24 -
Delta lake table
1 -
Delta Live
1 -
Delta Live Tables
6 -
Delta log
1 -
Delta Sharing
3 -
Delta-lake
1 -
Deploy
1 -
DESC
1 -
Details
1 -
Dev
1 -
Devops
1 -
Df
1 -
Different Notebook
1 -
Different Parameters
1 -
DimensionTables
1 -
Directory
3 -
Disable
1 -
Distribution
1 -
DLT
6 -
DLT Pipeline
3 -
Dolly
5 -
Dolly Demo
2 -
Download
2 -
EC2
1 -
Emr
2 -
Ensemble Models
1 -
Environment Variable
1 -
Epoch
1 -
Error handling
1 -
Error log
2 -
Eventhub
1 -
Example
1 -
Experiments
4 -
External Sources
1 -
Extract
1 -
Fact Tables
1 -
Failure
2 -
Feature Lookup
2 -
Feature Store
52 -
Feature Store API
2 -
Feature Store Table
1 -
Feature Table
6 -
Feature Tables
4 -
Features
2 -
FeatureStore
2 -
File Path
2 -
File Size
1 -
Fine Tune Spark Jobs
1 -
Forecasting
2 -
Forgot Password
2 -
Garbage Collection
1 -
Garbage Collection Optimization
1 -
Github
2 -
Github actions
2 -
Github Repo
2 -
Gitlab
1 -
GKE
1 -
Global Init Script
1 -
Global init scripts
4 -
Governance
1 -
Hi
1 -
Horovod
1 -
Html
1 -
Hyperopt
4 -
Hyperparameter Tuning
2 -
Iam
1 -
Image
3 -
Image Data
1 -
Inference Setup Error
1 -
INFORMATION
1 -
Input
1 -
Insert
1 -
Instance Profile
1 -
Int
2 -
Interactive cluster
1 -
Internal error
1 -
INVALID STATE
1 -
Invalid Type Code
1 -
IP
1 -
Ipython
1 -
Ipywidgets
1 -
JDBC Connections
1 -
Jira
1 -
Job
4 -
Job Parameters
1 -
Job Runs
1 -
JOBS
5 -
Jobs & Workflows
6 -
Join
1 -
Jsonfile
1 -
Kafka consumer
1 -
Key Management
1 -
Kinesis
1 -
Lakehouse
1 -
Large Datasets
1 -
Latest Version
1 -
Learning
1 -
Limit
3 -
LLM
3 -
Local computer
1 -
Local Machine
1 -
Log Model
2 -
Logging
1 -
Login
1 -
Logs
1 -
Long Time
2 -
Low Latency APIs
2 -
LTS ML
3 -
Machine
3 -
Machine Learning
24 -
Machine Learning Associate
1 -
Managed Table
1 -
Max Retries
1 -
Maximum Number
1 -
Medallion Architecture
1 -
Memory
3 -
Metadata
1 -
Metrics
3 -
Microsoft azure
1 -
ML Lifecycle
4 -
ML Model
4 -
ML Practioner
3 -
ML Runtime
1 -
MlFlow
75 -
MLflow API
5 -
MLflow Artifacts
2 -
MLflow Experiment
6 -
MLflow Experiments
3 -
Mlflow Model
10 -
Mlflow project
4 -
Mlflow registry
3 -
Mlflow Run
1 -
Mlflow Server
5 -
MLFlow Tracking Server
3 -
MLModels
2 -
Model Deployment
4 -
Model Lifecycle
6 -
Model Loading
2 -
Model Monitoring
1 -
Model registry
5 -
Model Serving
1 -
Model Serving Cluster
2 -
Model Serving REST API
6 -
Model Training
2 -
Model Tuning
1 -
Models
8 -
Module
3 -
Modulenotfounderror
1 -
MongoDB
1 -
Mount Point
1 -
Mounts
1 -
Multi
1 -
Multiline
1 -
Multiple users
1 -
Nested
1 -
New
2 -
New Feature
1 -
New Features
1 -
New Workspace
1 -
Nlp
3 -
Note
1 -
Notebook
6 -
Notebooks
5 -
Notification
2 -
Object
3 -
Onboarding
1 -
Online Feature Store Table
1 -
OOM Error
1 -
Open Source MLflow
4 -
Optimization
2 -
Optimize Command
1 -
OSS
3 -
Overwatch
1 -
Overwrite
2 -
Packages
2 -
Pandas udf
4 -
Pandas_udf
1 -
Parallel
1 -
Parallel processing
1 -
Parallel Runs
1 -
Parallelism
1 -
Parameter
2 -
PARAMETER VALUE
2 -
Partner Academy
1 -
Pending State
2 -
Performance Tuning
1 -
Permission
1 -
Photon Engine
1 -
Pickle
1 -
Pickle Files
2 -
Pip
2 -
Points
1 -
Possible
1 -
Postgres
1 -
Pricing
2 -
Primary Key
1 -
Primary Key Constraint
1 -
Progress bar
2 -
Proven Practices
2 -
Public
2 -
Pycharm IDE
1 -
Pymc3 Models
2 -
PyPI
1 -
Pyspark
6 -
Python
21 -
Python API
1 -
Python Code
1 -
Python Function
3 -
Python Libraries
1 -
Python Packages
1 -
Python Project
1 -
Python script
1 -
Pytorch
3 -
Reading-excel
2 -
Redis
2 -
Region
1 -
Remote RPC Client
1 -
Rest
1 -
RESTAPI
1 -
Result
1 -
Runtime update
1 -
Sagemaker
1 -
Salesforce
1 -
SAP
1 -
Scalability
1 -
Scalable Machine
2 -
Schema evolution
1 -
Script
1 -
Search
1 -
Security
2 -
Security Exception
1 -
Self Service Notebooks
1 -
Server
1 -
Serverless
1 -
Serving
1 -
Shap
2 -
Size
1 -
Sklearn
1 -
Slow
1 -
Small Scale Experimentation
1 -
Source Table
1 -
Spark
13 -
Spark config
1 -
Spark connector
1 -
Spark Error
1 -
Spark MLlib
2 -
Spark Pandas Api
1 -
Spark ui
1 -
Spark Version
2 -
Spark-submit
1 -
SparkML Models
2 -
Sparknlp
3 -
Spot
1 -
SQL
19 -
SQL Editor
1 -
SQL Queries
1 -
SQL Visualizations
1 -
Stage failure
2 -
Storage
3 -
Stream
2 -
Stream Data
1 -
Structtype
1 -
Structured streaming
2 -
Study Material
1 -
Summit23
2 -
Support
1 -
Support Team
1 -
Synapse
1 -
Synapse ML
1 -
Table
4 -
Table access control
1 -
Tableau
1 -
Task
1 -
Temporary View
1 -
Tensor flow
1 -
Test
1 -
Timeseries
1 -
Timestamps
1 -
TODAY
1 -
Training
6 -
Transaction Log
1 -
Trying
1 -
Tuning
2 -
UAT
1 -
Ui
1 -
Unexpected Error
1 -
Unity Catalog
12 -
Use Case
2 -
Use cases
1 -
Uuid
1 -
Validate ML Model
2 -
Values
1 -
Variable
1 -
Vector
1 -
Versioncontrol
1 -
Visualization
2 -
Web App Azure Databricks
1 -
Web ui
1 -
Weekly Release Notes
2 -
Whl
1 -
Worker Nodes
1 -
Workflow
2 -
Workflow Jobs
1 -
Workspace
2 -
Write
1 -
Writing
1 -
Z-ordering
1 -
Zorder
1
- « Previous
- Next »
User | Count |
---|---|
89 | |
39 | |
36 | |
25 | |
25 |