From 227bff9e43d84df6beabe9df483a229a41bf8edb Mon Sep 17 00:00:00 2001 From: William Valentin Date: Thu, 19 Mar 2026 15:35:37 -0700 Subject: [PATCH] feat: add gpt-5.3-codex-spark and qwen2.5-14b-local to LiteLLM init - Add gpt-5.3-codex-spark OpenAI Codex model - Add qwen2.5-14b-local: Qwen2.5-14B-Instruct running locally via llama.cpp at 192.168.153.113:18806, with model_info (chat mode, 8192 max tokens, 32768 input, supports function calling) Co-Authored-By: Claude Opus 4.6 --- litellm-init-models.sh | 31 +++++++++++++++++++++++++++++++ 1 file changed, 31 insertions(+) diff --git a/litellm-init-models.sh b/litellm-init-models.sh index e3be3f6..3e20803 100755 --- a/litellm-init-models.sh +++ b/litellm-init-models.sh @@ -110,6 +110,7 @@ add_model "gpt-5.1-codex-max" "openai/gpt-5.1-codex-max" "OPENAI_API_KEY" add_model "gpt-5.1-codex-mini" "openai/gpt-5.1-codex-mini" "OPENAI_API_KEY" add_model "gpt-5.2-codex" "openai/gpt-5.2-codex" "OPENAI_API_KEY" add_model "gpt-5.3-codex" "openai/gpt-5.3-codex" "OPENAI_API_KEY" +add_model "gpt-5.3-codex-spark" "openai/gpt-5.3-codex-spark" "OPENAI_API_KEY" add_model "gpt-5.4" "openai/gpt-5.4" "OPENAI_API_KEY" # Anthropic @@ -169,4 +170,34 @@ add_copilot_model "copilot-gemini-3-pro" "gemini-3-pro-preview" add_copilot_model "copilot-gemini-3.1-pro" "gemini-3.1-pro-preview" add_copilot_model "copilot-grok-code-fast" "grok-code-fast-1" +# Local models (llama.cpp — no API key, custom model_info) +if ! echo "$EXISTING" | grep -qx "qwen2.5-14b-local"; then + status=$(curl -s -o /tmp/model_resp.json -w "%{http_code}" \ + -X POST "$LITELLM_URL/model/new" \ + -H "Authorization: Bearer $LITELLM_MASTER_KEY" \ + -H "Content-Type: application/json" \ + -d '{ + "model_name": "qwen2.5-14b-local", + "litellm_params": { + "model": "openai/Qwen2.5-14B-Instruct-Q4_K_M.gguf", + "api_base": "http://192.168.153.113:18806/v1" + }, + "model_info": { + "mode": "chat", + "max_tokens": 8192, + "max_input_tokens": 32768, + "description": "Qwen2.5-14B-Instruct running locally via llama.cpp. Supports tool calling. Private/offline use.", + "supports_function_calling": true, + "supports_vision": false + } + }') + if [ "$status" = "200" ] || [ "$status" = "201" ]; then + echo " [ok] qwen2.5-14b-local" + else + echo " [fail] qwen2.5-14b-local — HTTP $status: $(cat /tmp/model_resp.json)" + fi +else + echo " [skip] qwen2.5-14b-local" +fi + echo "Done."