{ "providers": { "zai": { "baseUrl": "https://api.z.ai/api/coding/paas/v4", "api": "openai-completions", "models": [ { "id": "glm-5", "name": "GLM-5", "reasoning": true, "input": [ "text" ], "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 204800, "maxTokens": 131072, "api": "openai-completions" }, { "id": "glm-4.7", "name": "GLM-4.7", "reasoning": true, "input": [ "text" ], "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 204800, "maxTokens": 131072, "api": "openai-completions" }, { "id": "glm-4.7-flash", "name": "GLM-4.7 Flash", "reasoning": true, "input": [ "text" ], "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 204800, "maxTokens": 131072, "api": "openai-completions" }, { "id": "glm-4.7-flashx", "name": "GLM-4.7 FlashX", "reasoning": true, "input": [ "text" ], "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 204800, "maxTokens": 131072, "api": "openai-completions" } ] }, "litellm": { "baseUrl": "http://192.168.153.113:18804/v1", "apiKey": "sk-3oxoovwm69RyWRREGPJfDQ", "api": "openai-responses", "authHeader": true, "models": [ { "id": "gpt-4o", "name": "gpt-4o", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gpt-4o-mini", "name": "gpt-4o-mini", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gpt-4.1", "name": "gpt-4.1", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gpt-4.1-mini", "name": "gpt-4.1-mini", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gpt-4.1-nano", "name": "gpt-4.1-nano", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gpt-5", "name": "gpt-5", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gpt-5-mini", "name": "gpt-5-mini", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gpt-5-nano", "name": "gpt-5-nano", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gpt-5-pro", "name": "gpt-5-pro", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gpt-5.1", "name": "gpt-5.1", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gpt-5.2", "name": "gpt-5.2", "reasoning": true, "input": [ "text" ], "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gpt-5.2-pro", "name": "gpt-5.2-pro", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "o1", "name": "o1", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "o1-mini", "name": "o1-mini", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "o1-pro", "name": "o1-pro", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "o3", "name": "o3", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "o3-mini", "name": "o3-mini", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "o4-mini", "name": "o4-mini", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gpt-5-codex", "name": "gpt-5-codex", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gpt-5.1-codex", "name": "gpt-5.1-codex", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gpt-5.1-codex-mini", "name": "gpt-5.1-codex-mini", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gpt-5.2-codex", "name": "gpt-5.2-codex", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gpt-5.3-codex", "name": "gpt-5.3-codex", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "claude-opus-4-1", "name": "claude-opus-4-1", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "claude-opus-4", "name": "claude-opus-4", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "claude-haiku-4-5", "name": "claude-haiku-4-5", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "claude-3-haiku", "name": "claude-3-haiku", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gemini-2.0-flash", "name": "gemini-2.0-flash", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gemini-2.0-flash-lite", "name": "gemini-2.0-flash-lite", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gemini-2.5-flash-lite", "name": "gemini-2.5-flash-lite", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gemini-2.5-pro", "name": "gemini-2.5-pro", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gemini-3-flash-preview", "name": "gemini-3-flash-preview", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gpt-5.1-codex-max", "name": "gpt-5.1-codex-max", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "claude-opus-4-6", "name": "claude-opus-4-6", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "claude-sonnet-4-6", "name": "claude-sonnet-4-6", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "claude-opus-4-5", "name": "claude-opus-4-5", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "claude-sonnet-4-5", "name": "claude-sonnet-4-5", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "claude-sonnet-4", "name": "claude-sonnet-4", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gemini-2.5-flash", "name": "gemini-2.5-flash", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gemini-3-pro-preview", "name": "gemini-3-pro-preview", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gemini-flash-latest", "name": "gemini-flash-latest", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gemini-flash-lite-latest", "name": "gemini-flash-lite-latest", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "zai-glm-4.7", "name": "zai-glm-4.7", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gemini-3.1-pro-preview", "name": "gemini-3.1-pro-preview", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "gemini-pro-latest", "name": "gemini-pro-latest", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "zai-glm-4.5", "name": "zai-glm-4.5", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "zai-glm-4.5-air", "name": "zai-glm-4.5-air", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "zai-glm-4.6", "name": "zai-glm-4.6", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "zai-glm-5", "name": "zai-glm-5", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "openai/gpt-5.3-codex", "name": "gpt-5.3 Codex", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" }, { "id": "openai/gpt-5.2", "name": "gpt-5.2", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-responses" } ] }, "liteproxy": { "baseUrl": "http://192.168.153.113:18804/v1", "apiKey": "sk-3oxoovwm69RyWRREGPJfDQ", "api": "openai-completions", "models": [ { "id": "claude-haiku-4-5", "name": "Claude Haiku 4.5", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-completions" }, { "id": "gemini-2.5-flash", "name": "Gemini 2.5 Flash", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-completions" }, { "id": "gemini-2.5-flash-lite", "name": "Gemini 2.5 Flash Lite", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-completions" }, { "id": "zai-glm-4.7", "name": "GLM 4.7", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-completions" }, { "id": "gpt-4o", "name": "gpt-4o", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-completions" }, { "id": "gpt-4o-mini", "name": "gpt-4o-mini", "input": [ "text" ], "reasoning": false, "cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 }, "contextWindow": 200000, "maxTokens": 8192, "api": "openai-completions" } ] }, "github-copilot": { "baseUrl": "https://api.githubcopilot.com", "api": "openai-completions", "models": [] } } }