feat: add gpt-5.3-codex-spark and qwen2.5-14b-local to LiteLLM init
- Add gpt-5.3-codex-spark OpenAI Codex model - Add qwen2.5-14b-local: Qwen2.5-14B-Instruct running locally via llama.cpp at 192.168.153.113:18806, with model_info (chat mode, 8192 max tokens, 32768 input, supports function calling) Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
This commit is contained in:
@@ -110,6 +110,7 @@ add_model "gpt-5.1-codex-max" "openai/gpt-5.1-codex-max" "OPENAI_API_KEY"
|
|||||||
add_model "gpt-5.1-codex-mini" "openai/gpt-5.1-codex-mini" "OPENAI_API_KEY"
|
add_model "gpt-5.1-codex-mini" "openai/gpt-5.1-codex-mini" "OPENAI_API_KEY"
|
||||||
add_model "gpt-5.2-codex" "openai/gpt-5.2-codex" "OPENAI_API_KEY"
|
add_model "gpt-5.2-codex" "openai/gpt-5.2-codex" "OPENAI_API_KEY"
|
||||||
add_model "gpt-5.3-codex" "openai/gpt-5.3-codex" "OPENAI_API_KEY"
|
add_model "gpt-5.3-codex" "openai/gpt-5.3-codex" "OPENAI_API_KEY"
|
||||||
|
add_model "gpt-5.3-codex-spark" "openai/gpt-5.3-codex-spark" "OPENAI_API_KEY"
|
||||||
add_model "gpt-5.4" "openai/gpt-5.4" "OPENAI_API_KEY"
|
add_model "gpt-5.4" "openai/gpt-5.4" "OPENAI_API_KEY"
|
||||||
|
|
||||||
# Anthropic
|
# Anthropic
|
||||||
@@ -169,4 +170,34 @@ add_copilot_model "copilot-gemini-3-pro" "gemini-3-pro-preview"
|
|||||||
add_copilot_model "copilot-gemini-3.1-pro" "gemini-3.1-pro-preview"
|
add_copilot_model "copilot-gemini-3.1-pro" "gemini-3.1-pro-preview"
|
||||||
add_copilot_model "copilot-grok-code-fast" "grok-code-fast-1"
|
add_copilot_model "copilot-grok-code-fast" "grok-code-fast-1"
|
||||||
|
|
||||||
|
# Local models (llama.cpp — no API key, custom model_info)
|
||||||
|
if ! echo "$EXISTING" | grep -qx "qwen2.5-14b-local"; then
|
||||||
|
status=$(curl -s -o /tmp/model_resp.json -w "%{http_code}" \
|
||||||
|
-X POST "$LITELLM_URL/model/new" \
|
||||||
|
-H "Authorization: Bearer $LITELLM_MASTER_KEY" \
|
||||||
|
-H "Content-Type: application/json" \
|
||||||
|
-d '{
|
||||||
|
"model_name": "qwen2.5-14b-local",
|
||||||
|
"litellm_params": {
|
||||||
|
"model": "openai/Qwen2.5-14B-Instruct-Q4_K_M.gguf",
|
||||||
|
"api_base": "http://192.168.153.113:18806/v1"
|
||||||
|
},
|
||||||
|
"model_info": {
|
||||||
|
"mode": "chat",
|
||||||
|
"max_tokens": 8192,
|
||||||
|
"max_input_tokens": 32768,
|
||||||
|
"description": "Qwen2.5-14B-Instruct running locally via llama.cpp. Supports tool calling. Private/offline use.",
|
||||||
|
"supports_function_calling": true,
|
||||||
|
"supports_vision": false
|
||||||
|
}
|
||||||
|
}')
|
||||||
|
if [ "$status" = "200" ] || [ "$status" = "201" ]; then
|
||||||
|
echo " [ok] qwen2.5-14b-local"
|
||||||
|
else
|
||||||
|
echo " [fail] qwen2.5-14b-local — HTTP $status: $(cat /tmp/model_resp.json)"
|
||||||
|
fi
|
||||||
|
else
|
||||||
|
echo " [skip] qwen2.5-14b-local"
|
||||||
|
fi
|
||||||
|
|
||||||
echo "Done."
|
echo "Done."
|
||||||
|
|||||||
Reference in New Issue
Block a user