cancel
Showing results forย 
Search instead forย 
Did you mean:ย 
Machine Learning
Dive into the world of machine learning on the Databricks platform. Explore discussions on algorithms, model training, deployment, and more. Connect with ML enthusiasts and experts.
cancel
Showing results forย 
Search instead forย 
Did you mean:ย 

Model Serving - Shadow Deployment - Azure

ryojikn
New Contributor III

Hey,

I'm composing an architecture within the usage of Model Serving Endpoints and one of the needs that we're aiming to resolve is Shadow Deployment.

Currently, it seems that the traffic configurations available in model serving do not allow this type of behavior, mixing a mirroring requests effect with "fire and forget" responses from the shadow application.

 

Do you have this as a feature backlog? Or do you have any already implemented architecture composed within Azure pieces that I could use for that?

 

Thanks in advance

2 REPLIES 2

irtizak
New Contributor

I have the same query.

KaushalVachhani
Databricks Employee
Databricks Employee

@ryojikn and @irtizak , youโ€™re right. Databricks Model Serving allows splitting traffic between model versions, but it doesnโ€™t have a true shadow deployment where live production traffic is mirrored to a new model for monitoring without affecting user responses.

For now, you can try a couple of custom approaches:

1) Deploy one endpoint with your production model and another with the shadow model. On the client side, duplicate each incoming request to both endpoints, but return only the production modelโ€™s response to the user. You can capture and compare both responses later using the inference table for analysis.

2) Wrap your models inside a PyFunc and handle routing within the wrapper itself. You can reference models dynamically using aliases (like champion and challenger) so that whenever a model version changes, you donโ€™t need to update the wrapper code. Itโ€™ll automatically select the correct model version based on the alias when the endpoint is updated.