feat: add gpt-5.3-codex-spark and qwen2.5-14b-local to LiteLLM init
- Add gpt-5.3-codex-spark OpenAI Codex model - Add qwen2.5-14b-local: Qwen2.5-14B-Instruct running locally via llama.cpp at 192.168.153.113:18806, with model_info (chat mode, 8192 max tokens, 32768 input, supports function calling) Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
This commit is contained in:
@@ -110,6 +110,7 @@ add_model "gpt-5.1-codex-max" "openai/gpt-5.1-codex-max" "OPENAI_API_KEY"
|
||||
add_model "gpt-5.1-codex-mini" "openai/gpt-5.1-codex-mini" "OPENAI_API_KEY"
|
||||
add_model "gpt-5.2-codex" "openai/gpt-5.2-codex" "OPENAI_API_KEY"
|
||||
add_model "gpt-5.3-codex" "openai/gpt-5.3-codex" "OPENAI_API_KEY"
|
||||
add_model "gpt-5.3-codex-spark" "openai/gpt-5.3-codex-spark" "OPENAI_API_KEY"
|
||||
add_model "gpt-5.4" "openai/gpt-5.4" "OPENAI_API_KEY"
|
||||
|
||||
# Anthropic
|
||||
@@ -169,4 +170,34 @@ add_copilot_model "copilot-gemini-3-pro" "gemini-3-pro-preview"
|
||||
add_copilot_model "copilot-gemini-3.1-pro" "gemini-3.1-pro-preview"
|
||||
add_copilot_model "copilot-grok-code-fast" "grok-code-fast-1"
|
||||
|
||||
# Local models (llama.cpp — no API key, custom model_info)
|
||||
if ! echo "$EXISTING" | grep -qx "qwen2.5-14b-local"; then
|
||||
status=$(curl -s -o /tmp/model_resp.json -w "%{http_code}" \
|
||||
-X POST "$LITELLM_URL/model/new" \
|
||||
-H "Authorization: Bearer $LITELLM_MASTER_KEY" \
|
||||
-H "Content-Type: application/json" \
|
||||
-d '{
|
||||
"model_name": "qwen2.5-14b-local",
|
||||
"litellm_params": {
|
||||
"model": "openai/Qwen2.5-14B-Instruct-Q4_K_M.gguf",
|
||||
"api_base": "http://192.168.153.113:18806/v1"
|
||||
},
|
||||
"model_info": {
|
||||
"mode": "chat",
|
||||
"max_tokens": 8192,
|
||||
"max_input_tokens": 32768,
|
||||
"description": "Qwen2.5-14B-Instruct running locally via llama.cpp. Supports tool calling. Private/offline use.",
|
||||
"supports_function_calling": true,
|
||||
"supports_vision": false
|
||||
}
|
||||
}')
|
||||
if [ "$status" = "200" ] || [ "$status" = "201" ]; then
|
||||
echo " [ok] qwen2.5-14b-local"
|
||||
else
|
||||
echo " [fail] qwen2.5-14b-local — HTTP $status: $(cat /tmp/model_resp.json)"
|
||||
fi
|
||||
else
|
||||
echo " [skip] qwen2.5-14b-local"
|
||||
fi
|
||||
|
||||
echo "Done."
|
||||
|
||||
Reference in New Issue
Block a user