Codex throwing 404 NotFoundError via LiteLLM Proxy #19928
Replies: 1 comment
-
|
The literal string "model" in the error is the smoking gun. Your LiteLLM config has Fix the LiteLLM entry to use the actual served model name from vLLM (whatever you passed to model_list:
- model_name: gpt-5.4
litellm_params:
model: openai/gemma-4-26B-A4B-it
api_base: http://192.168.3.54:8000/v1
api_key: "any"Then about [model_providers.tommy]
name = "tommy-model-8889"
base_url = "http://192.168.3.54:8889/v1"
wire_api = "chat"
api_key = "any"That makes Codex hit Quick sanity check before retesting through Codex: curl http://192.168.3.54:8889/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer any" \
-d '{"model":"gpt-5.4","messages":[{"role":"user","content":"hi"}]}'If that 200s, Codex will work too. If you still get a 404 there, it's the vLLM served-model-name mismatch. Hit One thing to watch: some of the newer Codex-only models on the OpenAI side refuse chat completions and require lmk what |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Beta Was this translation helpful? Give feedback.
All reactions