cancel
Showing results forย 
Search instead forย 
Did you mean:ย 
Generative AI
Explore discussions on generative artificial intelligence techniques and applications within the Databricks Community. Share ideas, challenges, and breakthroughs in this cutting-edge field.
cancel
Showing results forย 
Search instead forย 
Did you mean:ย 

getting 500 on embedding model invocation call

tefrati
New Contributor III

I'm getting the following error message "{"error_code": "INTERNAL_ERROR", "message": "The server received an invalid response from an upstream server."}" when making a call to bge-large-en embedding model.

5 REPLIES 5

nayan_wylde
Honored Contributor III

Seems to me like a rate limit issue. Can you please confirm if the rate limit is not zero.

nayan_wylde_0-1760126703954.png

 

tefrati
New Contributor III

I don't see that option on my serving endpoint.
Also, if there's a rate limit, I'd expect to receive a 429 with relevant message and not 500.

tefrati_3-1760133664864.png

 

nayan_wylde
Honored Contributor III

Can you please click on the edit AI gateway it will take you show you the rate limit and share the screenshot

tefrati
New Contributor III

tefrati_0-1760135381782.png

no rate limitation is enabled

nayan_wylde
Honored Contributor III

Yeah your rate limit seems to be good. Can you also check the following points.

1. Use the Databricks-specific name (e.g., databricks-bge-large-en), not the Hugging Face model name. Check in Serving โ†’ Endpoints.
2. Validate Payload Format
{ "input": "text to embed" }
3. Test via Databricks UIUse Query endpoint in the Serving page. If that works, issue is client config.

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local communityโ€”sign up today to get started!

Sign Up Now