Change8
Error2 reports

Fix NotFoundError

in LiteLLM

Solution

"NotFoundError" in litellm usually means the model name you're using is invalid or the specified provider isn't supported for that model. Double-check your model name in the litellm.completion() call against the available models for your chosen provider (e.g., OpenAI, Azure). Ensure the provider setting in your code correctly reflects where the model is hosted; a typo can cause this error, so verify the spelling.

Timeline

First reported:Mar 7, 2026
Last reported:Mar 7, 2026

Need More Help?

View the full changelog and migration guides for LiteLLM

View LiteLLM Changelog