cancel
Showing results for 
Search instead for 
Did you mean: 
Machine Learning
Dive into the world of machine learning on the Databricks platform. Explore discussions on algorithms, model training, deployment, and more. Connect with ML enthusiasts and experts.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

BeadsPlayer
by New Contributor II
  • 566 Views
  • 0 replies
  • 0 kudos

Streaming inference with Delta Live Tables for a model registered in Unity Catalog

Hi there, I'm trying to run a streaming inference with Delta Live Tables with tables and a model registered in Unity Catalog, but it fails for unclear reasons. The DLT pipeline is based on a notebook, the channel is set to 'Preview', presumably runni...

Machine Learning
Delta Live Tables
streaming
Unity Catalog
  • 566 Views
  • 0 replies
  • 0 kudos
145093
by New Contributor II
  • 5746 Views
  • 2 replies
  • 2 kudos

MLFlow model loading taking long time and "model serving" failing during init

I am trying to load a simple Minmaxscaler model that was logged as a run through spark's ML Pipeline api for reuse. On average it takes 40+ seconds just to load the model with the following example: This is fine and the model transforms my data corre...

simple model load sometimes the model takes almost 3 min just to load
  • 5746 Views
  • 2 replies
  • 2 kudos
Latest Reply
DanSimpson
New Contributor II
  • 2 kudos

Hello,Any solutions found for this issue?I'm serving up a large number of models at a time, but since we converted to PySpark (due to our data demands), the mlflow.spark.load_model() is taking hours.Part of the reason to switch to spark was to help w...

  • 2 kudos
1 More Replies
Poised2Learn
by New Contributor III
  • 5216 Views
  • 2 replies
  • 1 kudos

Resolved! Model serving Ran out of memory

Hi fellows, I encountered memory(?) error when sending POST requests to my real-time endpoint, and I'm unable to find hardware setting to increase memory, as suggested by the Service Logs (below). Steps to Repro:(1) I registered a custom MLFlow model...

  • 5216 Views
  • 2 replies
  • 1 kudos
Latest Reply
Poised2Learn
New Contributor III
  • 1 kudos

Thank you for your responses, @Annapurna_Hiriy and @Retired_modIndeed, it appeared that my original model (~800MB) was too big for the current server. Based on your suggestion, I made a simpler/smaller model for this project, and then I was able to d...

  • 1 kudos
1 More Replies
chidifrank
by New Contributor II
  • 3876 Views
  • 1 replies
  • 0 kudos

Provisioned concurrency of serving endpoints scales to zero

Hi, We provisioned the endpoint with 4 DBUs and also disabled the scale_to_zero option. For some reason, it randomly drops to 0 provisioned concurrency. Logs available in the serving endpoint service are not insightful. Currently, we are provisioning...

chidifrank_0-1690968368091.png
Machine Learning
Model serving
serving endpoints
  • 3876 Views
  • 1 replies
  • 0 kudos
Latest Reply
chidifrank
New Contributor II
  • 0 kudos

Hi,I apologize if my question wasn't clear; let me clarify it.We are not using the scale_to_zero option and we are not doing any warmup requests so it should never scale to zero despite traffic or zero traffic right?   

  • 0 kudos
raghagra
by New Contributor III
  • 2058 Views
  • 2 replies
  • 2 kudos

Resolved! sparkxgbregressor and RandomForestRegressor not able to deploy for inferencing

I have been trying to deploy spark ML Models from the experiement page via UI, the deployment gets aborted after a long run, any particular reason for why this might be happening? I have also taken care of dependencies still it is failing.Dependency ...

  • 2058 Views
  • 2 replies
  • 2 kudos
Latest Reply
raghagra
New Contributor III
  • 2 kudos

@Kumaran Thanks for the reply kumaram The deployment was finally successful for Random Forest algorithm, failing for sparkxgbregressor.Sharing code snippet:from xgboost.spark import SparkXGBRegressor vec_assembler = VectorAssembler(inputCols=train_df...

  • 2 kudos
1 More Replies
shane
by New Contributor II
  • 2385 Views
  • 3 replies
  • 0 kudos

Not able to configure cluster settings instance type using mlflow api 2.0 to enable model serving.

I'm able to enable model serving by using the mlflow api 2.0 with the following code...instance = f'https://{workspace}.cloud.databricks.com' headers = {'Authorization': f'Bearer {api_workflow_access_token}'}   # Enable Model Serving import request...

Screen Shot 2023-02-02 at 3.53.16 PM
  • 2385 Views
  • 3 replies
  • 0 kudos
Latest Reply
Anonymous
Not applicable
  • 0 kudos

Hi @Shane Piesik​ Thank you for your question! To assist you better, please take a moment to review the answer and let me know if it best fits your needs.Please help us select the best solution by clicking on "Select As Best" if it does.Your feedback...

  • 0 kudos
2 More Replies
Rajib_Kumar_De
by New Contributor II
  • 2422 Views
  • 3 replies
  • 3 kudos

Databricks AutoML (Forecasting) Python SDK for Model Serving

I am using Databricks AutoML ( Python SDK) to forecast bed occupancy. (Actually, Databricks used MLflow experiments for AutoML run). After training with different iterations, I registered the best model in the Databricks Model registry. Now I am tryi...

  • 2422 Views
  • 3 replies
  • 3 kudos
Latest Reply
Debayan
Databricks Employee
  • 3 kudos

Hi, It can be a bug if the python version is 3.9.5 and still the error is on compatibility. Could you please raise a support case to look into it further?

  • 3 kudos
2 More Replies
anthonylavado
by New Contributor III
  • 2324 Views
  • 2 replies
  • 7 kudos

Can't Add Cluster-scoped Init Script to Model Serving Cluster

Similar to this other question: https://community.databricks.com/s/question/0D58Y00008hahwuSAA/cant-edit-the-cluster-created-by-mlflow-model-servingWe're using Azure Databricks, and have a model that requires a WHL to be downloaded from a private add...

  • 2324 Views
  • 2 replies
  • 7 kudos
Latest Reply
939772
New Contributor III
  • 7 kudos

Has anyone had success with this? Trying to solve a resolve issue.

  • 7 kudos
1 More Replies
jhonw901227
by New Contributor II
  • 997 Views
  • 0 replies
  • 0 kudos

MLflow Model Serving on Azure Databricks

I know that in the documentation about model serving says.The cluster is maintained as long as serving is enabled, even if no active model version exists. To terminate the serving cluster, disable model serving for the registered model.The cluster is...

  • 997 Views
  • 0 replies
  • 0 kudos
Mohit_m
by Valued Contributor II
  • 2369 Views
  • 2 replies
  • 3 kudos

Resolved! How to enable and disable Model Serving using Rest API

ML flow model serving in Databricks docs details the options to enable and disable from the UIhttps://docs.databricks.com/applications/mlflow/model-serving.html

  • 2369 Views
  • 2 replies
  • 3 kudos
Latest Reply
Mohit_m
Valued Contributor II
  • 3 kudos

Please find below the REST APIs to enable and disable Model-ServingBelow are the examples in PythonYou need to use the token to interact with Rest APItoken = "dxxxxxx"instance = "https://<workspacexxx>.cloud.databricks.com"headers = {'Authorization':...

  • 3 kudos
1 More Replies
Labels