- 2175 Views
- 2 replies
- 0 kudos
Update model serving endpoint
Hi all,I've been able to create a model serving endpoint through the api, following the docs, but then when trying to update the version, I get the following error:'{"error_code":"RESOURCE_ALREADY_EXISTS","message":"Endpoint with name \'ml-project\' ...
- 2175 Views
- 2 replies
- 0 kudos
- 0 kudos
Folks, Alternate way you can also deploy the models in serving layer with different versions. Though I am using mLflow. You may also refer to the below link if it its helpful How to Quickly Deploy, Test & Manage ML Models as REST Endpoints with Datab...
- 0 kudos
- 6447 Views
- 6 replies
- 0 kudos
TypeError: 'JavaPackage' object is not callable
Hi Team,I am facing issue with above error while I am trying to do BERT embeddings, by specifying the model path and it is giving error while downloading the model.spark version is 3.3.0Can any one of you help me on this?
- 6447 Views
- 6 replies
- 0 kudos
- 0 kudos
please give error detail.i think you need add some Maven package, like https://community.databricks.com/t5/machine-learning/synapse-ml-typeerror-javapackage-object-is-not-callable/td-p/58897
- 0 kudos
- 2092 Views
- 1 replies
- 0 kudos
Synapse ML - TypeError: 'JavaPackage' object is not callable
from synapse.ml.lightgbm import *LightGBMRegressor()gives me TypeError: 'JavaPackage' object is not callableDBR 8.2 (includes Apache Spark 3.1.1, Scala 2.12)
- 2092 Views
- 1 replies
- 0 kudos
- 0 kudos
You need install Maven package `com.microsoft.azure:synapseml_2.12:1.0.2`https://microsoft.github.io/SynapseML/docs/Get%20Started/Install%20SynapseML/#databricks
- 0 kudos
- 4279 Views
- 3 replies
- 0 kudos
org.apache.spark.SparkException: Job aborted due to stage failure during Model Training
org.apache.spark.SparkException: Job aborted due to stage failure: Could not recover from a failed barrier ResultStage. Most recent failure reason: Stage failed because barrier task ResultTask(160, 13) finished unsuccessfully.
- 4279 Views
- 3 replies
- 0 kudos
- 0 kudos
Could you share the stage details where the issue happened?
- 0 kudos
- 7861 Views
- 4 replies
- 2 kudos
Resolved! Error when accessing rdd of DataFrame
I need to run this kind of code: from pyspark.sql import SparkSession import pandas as pd # Create a Spark session spark = SparkSession.builder.appName("example").getOrCreate() # Sample data data = [("Alice", 1), ("Bob", 2), ("Charlie", 3), ("David...
- 7861 Views
- 4 replies
- 2 kudos
- 2 kudos
I found a good solution that works both locally and in the cloud. Copy pasting the code in case it helps someone.This is the higher level function in charge of partitioning the data and sending the data and the function fn to each node. def decrypt_d...
- 2 kudos
- 1446 Views
- 0 replies
- 0 kudos
How to use mlflow to log a composite estimator (multiple pipes) and then deploy it as rest endpoint
Hello,I am trying to deploy a composite estimator as single model, by logging the run with mlflow and registering the model.Can anyone help with how this can be done? This estimator contains different chains-text: data- tfidf- svm- svm.decision_funct...
- 1446 Views
- 0 replies
- 0 kudos
- 5380 Views
- 7 replies
- 1 kudos
Getting Permission Denied on model
while creating service endpoint , getting permission denied error. Need help how to provide the permission.
- 5380 Views
- 7 replies
- 1 kudos
- 1 kudos
@Debayan I managed to solve the issue using databricks sdk library. From UI it was failing with the same error as mentioned by @BalaRamesh
- 1 kudos
- 3233 Views
- 5 replies
- 0 kudos
Cant access path for Shared Access Mode Cluster
Hi! I was able to successfully download and run selenium on both single and no isolation mode clusters. But the shared cluster do not seem to have permission to access the path of drivers needed such as chrome and chromedriver.1) what is the general ...
- 3233 Views
- 5 replies
- 0 kudos
- 0 kudos
Hi, Is the cluster UC enabled? Also, what are the types of the drivers (chrome and chromedriver) ?
- 0 kudos
- 3722 Views
- 2 replies
- 0 kudos
Serving endpoints: model server failed to load the model: the file bash was not found: uknown
While trying to create a serving endpoint with my custom model, I get a "Failed" state:Model server failed to load the model. Please see service logs for more information.The service logs show the following:Container failed with: failed to create con...
- 3722 Views
- 2 replies
- 0 kudos
- 0 kudos
I have faced the similar issue. still didn't find the right solution. In my case, the below is the error trace i found from service logs. Not sure where the issue could be"An error occurred while loading the model. You haven't configured the CLI yet!...
- 0 kudos
- 2113 Views
- 1 replies
- 0 kudos
Databricks OpenAI Integration Issue
Facing challenges in leveraging Databricks for serving, logging, and monitoring OpenAI usage within an Azure environment. Specifically, encountering issues with Inference Tables not being enabled in the UI when creating serving endpoints with externa...
- 2113 Views
- 1 replies
- 0 kudos
- 0 kudos
Hi, In addition to the previous message, you can refer to https://learn.microsoft.com/en-us/azure/databricks/large-language-models/ai-generate-text-example.
- 0 kudos
- 1986 Views
- 3 replies
- 1 kudos
Unable to distribute the workload to different worker
Hello Team, I am unable to distribute the workload to databricks different worker while using the hugging face GPT2 LLM model. Jobs always use the 1 node even though we have the min and max worker node setting with 2. Appreciate if anyone can share a...
- 1986 Views
- 3 replies
- 1 kudos
- 1 kudos
Hello, We have around between 5k to 10K transcript files available in the ADLS gen 2 and we are using hugging face gpt2 model to train and server the model and expecting to pass the workload to different cluster nodes while serving the LLM model and ...
- 1 kudos
- 2280 Views
- 0 replies
- 0 kudos
using openai Api in Databricks without iterating rows
Hi to everyone,I have a delta table with a column 'comment' I would like to add a new column 'sentiment', and I would like to calculate it using openai API.I already know how to create a databricks endpoint to an external model and how to use it (us...
- 2280 Views
- 0 replies
- 0 kudos
- 4328 Views
- 3 replies
- 6 kudos
Resolved! Vector Search Index not provisioning
I am trying to deploy a Vector Search Index using both the UI and the Python VectorSearchClient. In both cases the command is successful but in the Catalog Explorer the newly created index stalls with the status 'Provisioning Index' for hours. Previo...
- 4328 Views
- 3 replies
- 6 kudos
- 6 kudos
Thanks, I didn't change anything either and now is working.
- 6 kudos
- 10257 Views
- 13 replies
- 6 kudos
Resolved! Can't Run an AutoML Experiment Because Button is Greyed Out
I am trying to run an AutoML experiment but the button stays greyed out no matter what I do. I've tried different cluster configurations, different datasets, even blew away the instance in Azure and re-created it across two different Azure accounts s...
- 10257 Views
- 13 replies
- 6 kudos
- 3164 Views
- 4 replies
- 0 kudos
How do I distribute machine learning process in my spark data frame
Hi,I'm trying to use around 5 numerical features on 3.5 million rows to train and test my model with a spark data frame.My cluster has 60 nodes available but is only using 2. How can I distribute the process or make it for efficient and faster.My cod...
- 3164 Views
- 4 replies
- 0 kudos
- 0 kudos
@mohaimen_syed - can you please try using pyspark.ml implementation of randomForestClassifier instead of sklearn and see if it works. Below is an example - https://github.com/apache/spark/blob/master/examples/src/main/python/ml/random_forest_classif...
- 0 kudos
Join Us as a Local Community Builder!
Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!
Sign Up Now-
Access control
3 -
Access Data
2 -
AccessKeyVault
1 -
ADB
2 -
Airflow
1 -
Amazon
2 -
Apache
1 -
Apache spark
3 -
APILimit
1 -
Artifacts
1 -
Audit
1 -
Autoloader
6 -
Autologging
2 -
Automation
2 -
Automl
32 -
AWS
7 -
Aws databricks
1 -
AWSSagemaker
1 -
Azure
32 -
Azure active directory
1 -
Azure blob storage
2 -
Azure data lake
1 -
Azure Data Lake Storage
3 -
Azure data lake store
1 -
Azure databricks
32 -
Azure event hub
1 -
Azure key vault
1 -
Azure sql database
1 -
Azure Storage
2 -
Azure synapse
1 -
Azure Unity Catalog
1 -
Azure vm
1 -
AzureML
2 -
Bar
1 -
Beta
1 -
Better Way
1 -
BI Integrations
1 -
BI Tool
1 -
Billing and Cost Management
1 -
Blob
1 -
Blog
1 -
Blog Post
1 -
Broadcast variable
1 -
Business Intelligence
1 -
CatalogDDL
1 -
Centralized Model Registry
1 -
Certification
2 -
Certification Badge
1 -
Change
1 -
Change Logs
1 -
Chatgpt
2 -
Check
2 -
Classification Model
1 -
Cloud Storage
1 -
Cluster
10 -
Cluster policy
1 -
Cluster Start
1 -
Cluster Termination
2 -
Clustering
1 -
ClusterMemory
1 -
CNN HOF
1 -
Column names
1 -
Community Edition
1 -
Community Edition Password
1 -
Community Members
1 -
Company Email
1 -
Condition
1 -
Config
1 -
Configure
3 -
Confluent Cloud
1 -
Container
2 -
ContainerServices
1 -
Control Plane
1 -
ControlPlane
1 -
Copy
1 -
Copy into
2 -
CosmosDB
1 -
Cost
1 -
Courses
2 -
Csv files
1 -
Dashboards
1 -
Data
8 -
Data Engineer Associate
1 -
Data Engineer Certification
1 -
Data Explorer
1 -
Data Ingestion
2 -
Data Ingestion & connectivity
11 -
Data Quality
1 -
Data Quality Checks
1 -
Data Science & Engineering
2 -
databricks
5 -
Databricks Academy
3 -
Databricks Account
1 -
Databricks AutoML
9 -
Databricks Cluster
3 -
Databricks Community
5 -
Databricks community edition
4 -
Databricks connect
1 -
Databricks dbfs
1 -
Databricks Feature Store
1 -
Databricks Job
1 -
Databricks Lakehouse
1 -
Databricks Mlflow
4 -
Databricks Model
2 -
Databricks notebook
10 -
Databricks ODBC
1 -
Databricks Platform
1 -
Databricks Pyspark
1 -
Databricks Python Notebook
1 -
Databricks Runtime
9 -
Databricks SQL
8 -
Databricks SQL Permission Problems
1 -
Databricks Terraform
1 -
Databricks Training
2 -
Databricks Unity Catalog
1 -
Databricks V2
1 -
Databricks version
1 -
Databricks Workflow
2 -
Databricks Workflows
1 -
Databricks workspace
2 -
Databricks-connect
1 -
DatabricksContainer
1 -
DatabricksML
6 -
Dataframe
3 -
DataSharing
1 -
Datatype
1 -
DataVersioning
1 -
Date Column
1 -
Dateadd
1 -
DB Notebook
1 -
DB Runtime
1 -
DBFS
5 -
DBFS Rest Api
1 -
Dbt
1 -
Dbu
1 -
DDL
1 -
DDP
1 -
Dear Community
1 -
DecisionTree
1 -
Deep learning
4 -
Default Location
1 -
Delete
1 -
Delt Lake
4 -
Delta
24 -
Delta lake table
1 -
Delta Live
1 -
Delta Live Tables
6 -
Delta log
1 -
Delta Sharing
3 -
Delta-lake
1 -
Deploy
1 -
DESC
1 -
Details
1 -
Dev
1 -
Devops
1 -
Df
1 -
Different Notebook
1 -
Different Parameters
1 -
DimensionTables
1 -
Directory
3 -
Disable
1 -
Distribution
1 -
DLT
6 -
DLT Pipeline
3 -
Dolly
5 -
Dolly Demo
2 -
Download
2 -
EC2
1 -
Emr
2 -
Ensemble Models
1 -
Environment Variable
1 -
Epoch
1 -
Error handling
1 -
Error log
2 -
Eventhub
1 -
Example
1 -
Experiments
4 -
External Sources
1 -
Extract
1 -
Fact Tables
1 -
Failure
2 -
Feature Lookup
2 -
Feature Store
52 -
Feature Store API
2 -
Feature Store Table
1 -
Feature Table
6 -
Feature Tables
4 -
Features
2 -
FeatureStore
2 -
File Path
2 -
File Size
1 -
Fine Tune Spark Jobs
1 -
Forecasting
2 -
Forgot Password
2 -
Garbage Collection
1 -
Garbage Collection Optimization
1 -
Github
2 -
Github actions
2 -
Github Repo
2 -
Gitlab
1 -
GKE
1 -
Global Init Script
1 -
Global init scripts
4 -
Governance
1 -
Hi
1 -
Horovod
1 -
Html
1 -
Hyperopt
4 -
Hyperparameter Tuning
2 -
Iam
1 -
Image
3 -
Image Data
1 -
Inference Setup Error
1 -
INFORMATION
1 -
Input
1 -
Insert
1 -
Instance Profile
1 -
Int
2 -
Interactive cluster
1 -
Internal error
1 -
INVALID STATE
1 -
Invalid Type Code
1 -
IP
1 -
Ipython
1 -
Ipywidgets
1 -
JDBC Connections
1 -
Jira
1 -
Job
4 -
Job Parameters
1 -
Job Runs
1 -
JOBS
5 -
Jobs & Workflows
6 -
Join
1 -
Jsonfile
1 -
Kafka consumer
1 -
Key Management
1 -
Kinesis
1 -
Lakehouse
1 -
Large Datasets
1 -
Latest Version
1 -
Learning
1 -
Limit
3 -
LLM
3 -
Local computer
1 -
Local Machine
1 -
Log Model
2 -
Logging
1 -
Login
1 -
Logs
1 -
Long Time
2 -
Low Latency APIs
2 -
LTS ML
3 -
Machine
3 -
Machine Learning
24 -
Machine Learning Associate
1 -
Managed Table
1 -
Max Retries
1 -
Maximum Number
1 -
Medallion Architecture
1 -
Memory
3 -
Metadata
1 -
Metrics
3 -
Microsoft azure
1 -
ML Lifecycle
4 -
ML Model
4 -
ML Practioner
3 -
ML Runtime
1 -
MlFlow
75 -
MLflow API
5 -
MLflow Artifacts
2 -
MLflow Experiment
6 -
MLflow Experiments
3 -
Mlflow Model
10 -
Mlflow project
4 -
Mlflow registry
3 -
Mlflow Run
1 -
Mlflow Server
5 -
MLFlow Tracking Server
3 -
MLModels
2 -
Model Deployment
4 -
Model Lifecycle
6 -
Model Loading
2 -
Model Monitoring
1 -
Model registry
5 -
Model Serving
1 -
Model Serving Cluster
2 -
Model Serving REST API
6 -
Model Training
2 -
Model Tuning
1 -
Models
8 -
Module
3 -
Modulenotfounderror
1 -
MongoDB
1 -
Mount Point
1 -
Mounts
1 -
Multi
1 -
Multiline
1 -
Multiple users
1 -
Nested
1 -
New
2 -
New Feature
1 -
New Features
1 -
New Workspace
1 -
Nlp
3 -
Note
1 -
Notebook
6 -
Notebooks
5 -
Notification
2 -
Object
3 -
Onboarding
1 -
Online Feature Store Table
1 -
OOM Error
1 -
Open Source MLflow
4 -
Optimization
2 -
Optimize Command
1 -
OSS
3 -
Overwatch
1 -
Overwrite
2 -
Packages
2 -
Pandas udf
4 -
Pandas_udf
1 -
Parallel
1 -
Parallel processing
1 -
Parallel Runs
1 -
Parallelism
1 -
Parameter
2 -
PARAMETER VALUE
2 -
Partner Academy
1 -
Pending State
2 -
Performance Tuning
1 -
Permission
1 -
Photon Engine
1 -
Pickle
1 -
Pickle Files
2 -
Pip
2 -
Points
1 -
Possible
1 -
Postgres
1 -
Pricing
2 -
Primary Key
1 -
Primary Key Constraint
1 -
Progress bar
2 -
Proven Practices
2 -
Public
2 -
Pycharm IDE
1 -
Pymc3 Models
2 -
PyPI
1 -
Pyspark
6 -
Python
21 -
Python API
1 -
Python Code
1 -
Python Function
3 -
Python Libraries
1 -
Python Packages
1 -
Python Project
1 -
Python script
1 -
Pytorch
3 -
Reading-excel
2 -
Redis
2 -
Region
1 -
Remote RPC Client
1 -
Rest
1 -
RESTAPI
1 -
Result
1 -
Runtime update
1 -
Sagemaker
1 -
Salesforce
1 -
SAP
1 -
Scalability
1 -
Scalable Machine
2 -
Schema evolution
1 -
Script
1 -
Search
1 -
Security
2 -
Security Exception
1 -
Self Service Notebooks
1 -
Server
1 -
Serverless
1 -
Serving
1 -
Shap
2 -
Size
1 -
Sklearn
1 -
Slow
1 -
Small Scale Experimentation
1 -
Source Table
1 -
Spark
13 -
Spark config
1 -
Spark connector
1 -
Spark Error
1 -
Spark MLlib
2 -
Spark Pandas Api
1 -
Spark ui
1 -
Spark Version
2 -
Spark-submit
1 -
SparkML Models
2 -
Sparknlp
3 -
Spot
1 -
SQL
19 -
SQL Editor
1 -
SQL Queries
1 -
SQL Visualizations
1 -
Stage failure
2 -
Storage
3 -
Stream
2 -
Stream Data
1 -
Structtype
1 -
Structured streaming
2 -
Study Material
1 -
Summit23
2 -
Support
1 -
Support Team
1 -
Synapse
1 -
Synapse ML
1 -
Table
4 -
Table access control
1 -
Tableau
1 -
Task
1 -
Temporary View
1 -
Tensor flow
1 -
Test
1 -
Timeseries
1 -
Timestamps
1 -
TODAY
1 -
Training
6 -
Transaction Log
1 -
Trying
1 -
Tuning
2 -
UAT
1 -
Ui
1 -
Unexpected Error
1 -
Unity Catalog
12 -
Use Case
2 -
Use cases
1 -
Uuid
1 -
Validate ML Model
2 -
Values
1 -
Variable
1 -
Vector
1 -
Versioncontrol
1 -
Visualization
2 -
Web App Azure Databricks
1 -
Web ui
1 -
Weekly Release Notes
2 -
Whl
1 -
Worker Nodes
1 -
Workflow
2 -
Workflow Jobs
1 -
Workspace
2 -
Write
1 -
Writing
1 -
Z-ordering
1 -
Zorder
1
- « Previous
- Next »
User | Count |
---|---|
89 | |
39 | |
36 | |
25 | |
25 |