feat: add gpt-5.3-codex-spark and qwen2.5-14b-local to LiteLLM init

- Add gpt-5.3-codex-spark OpenAI Codex model
- Add qwen2.5-14b-local: Qwen2.5-14B-Instruct running locally via
  llama.cpp at 192.168.153.113:18806, with model_info (chat mode,
  8192 max tokens, 32768 input, supports function calling)

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
This commit is contained in:
William Valentin
2026-03-19 15:35:37 -07:00
parent ed3273d1ed
commit 227bff9e43

View File

@@ -110,6 +110,7 @@ add_model "gpt-5.1-codex-max" "openai/gpt-5.1-codex-max" "OPENAI_API_KEY"
add_model "gpt-5.1-codex-mini" "openai/gpt-5.1-codex-mini" "OPENAI_API_KEY"
add_model "gpt-5.2-codex" "openai/gpt-5.2-codex" "OPENAI_API_KEY"
add_model "gpt-5.3-codex" "openai/gpt-5.3-codex" "OPENAI_API_KEY"
add_model "gpt-5.3-codex-spark" "openai/gpt-5.3-codex-spark" "OPENAI_API_KEY"
add_model "gpt-5.4" "openai/gpt-5.4" "OPENAI_API_KEY"
# Anthropic
@@ -169,4 +170,34 @@ add_copilot_model "copilot-gemini-3-pro" "gemini-3-pro-preview"
add_copilot_model "copilot-gemini-3.1-pro" "gemini-3.1-pro-preview"
add_copilot_model "copilot-grok-code-fast" "grok-code-fast-1"
# Local models (llama.cpp — no API key, custom model_info)
if ! echo "$EXISTING" | grep -qx "qwen2.5-14b-local"; then
status=$(curl -s -o /tmp/model_resp.json -w "%{http_code}" \
-X POST "$LITELLM_URL/model/new" \
-H "Authorization: Bearer $LITELLM_MASTER_KEY" \
-H "Content-Type: application/json" \
-d '{
"model_name": "qwen2.5-14b-local",
"litellm_params": {
"model": "openai/Qwen2.5-14B-Instruct-Q4_K_M.gguf",
"api_base": "http://192.168.153.113:18806/v1"
},
"model_info": {
"mode": "chat",
"max_tokens": 8192,
"max_input_tokens": 32768,
"description": "Qwen2.5-14B-Instruct running locally via llama.cpp. Supports tool calling. Private/offline use.",
"supports_function_calling": true,
"supports_vision": false
}
}')
if [ "$status" = "200" ] || [ "$status" = "201" ]; then
echo " [ok] qwen2.5-14b-local"
else
echo " [fail] qwen2.5-14b-local — HTTP $status: $(cat /tmp/model_resp.json)"
fi
else
echo " [skip] qwen2.5-14b-local"
fi
echo "Done."