Explore discussions on generative artificial intelligence techniques and applications within the Databricks Community. Share ideas, challenges, and breakthroughs in this cutting-edge field.
I am facing this below error while serving codellama model: Exception: Request failed with status 400, {"error_code":"INVALID_PARAMETER_VALUE","message":"The provisioned throughput model Llama 2 7B is deprecated and no longer supported in serving. See https://docs.databricks.com/aws/en/machine-learning/foundation-model-apis/supported-models for information on currently supported models.","details":[{"@type":"type.googleapis.com/google.rpc.RequestInfo","request_id":"ff0707d9-1c8d-4b82-b34f-c6d138709b70","serving_data":""}]}
I just wanted to know the method by which databricks identifies the model whether it is deprecated or not. Does this check the model name or config?
Thanks for the response, Suppose I register a model (say Llama2‑7B) in Databricks using some custom name. How does Databricks “know” that this model is actually Llama2‑7B under the hood? Is the identification based on configuration files, tokenizer metadata, or something else?
I guess they identify the model based on some kind of metadata. This is an implementation detail though, so unless someone from the Databricks team joins the discussion, we can only guess.
Join Us as a Local Community Builder!
Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!