โ09-26-2023 02:10 AM
I have a model registered in unity catalog which works fine and I can load / run and get results returned. I wanted to create a serving endpoint but when I try I get this error.
Served model creation failed for served model 'model_name', config version 2. Error message: Container creation failed. Please see build logs for more information.
I can't see any build logs as build logs are only available once the model container has been built.
โ10-11-2023 09:18 AM - edited โ10-11-2023 09:24 AM
@kashy Looks like the model is not correctly referenced while loading. You should reference the path of the model till โmodel-bestโ, which is the top-level directory.
loaded_model = mlflow.spacy.load_model("</path/to/your/model/>/model-best")
โ10-03-2023 07:50 AM
Hi @kashy
Thank you for your question in the Databricks community.
you can check the service logs to see if there were any other errors or failure messages that may help you resolve the issue.
โ10-07-2023 04:51 AM
thank you @Kumaran but there is no service log created either. All I see under Service Logs is:
"There are currently no replicas in a running state"
โ10-10-2023 03:03 AM
@kashy Please check the events tab too.
Also, has a model-serving endpoint ever been created in your workspace?
Is there any restricted networking setup for the model serving cluster? This is usually enabled via a global init script.
We recommend you check these points as well.
โ10-10-2023 04:09 AM
Thanks @Annapurna_Hiriy I managed to get a bit further this time. The Container Image was successfully created but then "Served model creation aborted for served model `model-version`, config version 1, since the update timed out
In the service log I see this error
[86cccn6zst] [2023-10-10 10:56:44 +0000] [7] [INFO] Starting gunicorn 20.1.0
[86cccn6zst] [2023-10-10 10:56:44 +0000] [7] [INFO] Listening at: http://0.0.0.0:8080 (7)
[86cccn6zst] [2023-10-10 10:56:44 +0000] [7] [INFO] Using worker: sync
[86cccn6zst] [2023-10-10 10:56:44 +0000] [8] [INFO] Booting worker with pid: 8
[86cccn6zst] An error occurred while loading the model. [E053] Could not read meta.json from /model/artifacts/[model_name].
The model in question was saved when the model created and first run in Ml_flow
โ10-11-2023 12:30 AM
@kashy We need more details to analyze this.
What is the model that you are using, size of the model, if it's GPU/CPU model serving, how you are loading the model, and more info on wrapper model you have logged. For models with huge size GPU serving is recommended. Even if you are using GPU there would be certain constraints that you should take care of.
Please create a support ticket with us with all the details mentioned above and we will be able to assist you better.
โ10-11-2023 02:07 AM
Thanks @Annapurna_Hiriy It is a spaCy text classification model.
I was using GPU to create the endpoint which is created successfully now. Actually failing at the next stage where I get an service log error:
[6558blggf5] An error occurred while loading the model. [E053] Could not read meta.json from /model/artifacts/spacy_model.
Model Load:
โ10-11-2023 09:18 AM - edited โ10-11-2023 09:24 AM
@kashy Looks like the model is not correctly referenced while loading. You should reference the path of the model till โmodel-bestโ, which is the top-level directory.
loaded_model = mlflow.spacy.load_model("</path/to/your/model/>/model-best")
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you wonโt want to miss the chance to attend and share knowledge.
If there isnโt a group near you, start one and help create a community that brings people together.
Request a New Group