UNPKG

ultimate-mcp-server

Version:

The definitive all-in-one Model Context Protocol server for AI-assisted coding across 30+ platforms

220 lines 7.96 kB
// Centralized model configuration - Updated with latest models (2025) // Models verified for OpenRouter availability export const MODELS = { // OpenAI models GPT_4O: "openai/gpt-4o", GPT_4O_MINI: "openai/gpt-4o-mini", GPT_4_TURBO: "openai/gpt-4-turbo", GPT_4: "openai/gpt-4", GPT_3_5_TURBO: "openai/gpt-3.5-turbo", // Anthropic models CLAUDE_3_OPUS: "anthropic/claude-3-opus", CLAUDE_3_SONNET: "anthropic/claude-3-sonnet", CLAUDE_3_HAIKU: "anthropic/claude-3-haiku", CLAUDE_3_5_SONNET: "anthropic/claude-3.5-sonnet", CLAUDE_3_5_SONNET_LATEST: "anthropic/claude-3.5-sonnet-20241022", // Google models GEMINI_2_PRO: "google/gemini-2.5-pro", GEMINI_2_FLASH: "google/gemini-2.5-flash", GEMINI_PRO: "google/gemini-pro", GEMINI_PRO_VISION: "google/gemini-pro-vision", GEMINI_2_FLASH_THINKING: "google/gemini-2.0-flash-thinking-exp", // xAI models GROK_2: "x-ai/grok-2", GROK_2_VISION: "x-ai/grok-2-vision", GROK_3_BETA: "x-ai/grok-3-beta", GROK_4: "x-ai/grok-4-beta", // Latest Grok model // Meta models LLAMA_3_3_70B: "meta-llama/llama-3.3-70b-instruct", LLAMA_3_1_405B: "meta-llama/llama-3.1-405b-instruct", LLAMA_3_1_70B: "meta-llama/llama-3.1-70b-instruct", LLAMA_3_1_8B: "meta-llama/llama-3.1-8b-instruct", LLAMA_3_2_90B_VISION: "meta-llama/llama-3.2-90b-vision-instruct", LLAMA_3_2_11B_VISION: "meta-llama/llama-3.2-11b-vision-instruct", LLAMA_3_2_3B: "meta-llama/llama-3.2-3b-instruct", LLAMA_3_2_1B: "meta-llama/llama-3.2-1b-instruct", // Mistral models MISTRAL_LARGE: "mistralai/mistral-large-2411", MISTRAL_MEDIUM: "mistralai/mistral-medium", MISTRAL_SMALL: "mistralai/mistral-small", MIXTRAL_8X7B: "mistralai/mixtral-8x7b-instruct", MIXTRAL_8X22B: "mistralai/mixtral-8x22b-instruct", CODESTRAL: "mistralai/codestral-2501", // Latest code model PIXTRAL_LARGE: "mistralai/pixtral-large-2411", // DeepSeek models DEEPSEEK_V3: "deepseek/deepseek-v3", DEEPSEEK_CHAT: "deepseek/deepseek-chat", DEEPSEEK_CODER_V2: "deepseek/deepseek-coder-v2-instruct", DEEPSEEK_R1: "deepseek/deepseek-r1", DEEPSEEK_R1_LITE: "deepseek/deepseek-r1-lite-preview", // Qwen models QWEN_2_5_CODER_32B: "qwen/qwen-2.5-coder-32b-instruct", QWEN_2_5_72B: "qwen/qwen-2.5-72b-instruct", QWEN_2_5_14B: "qwen/qwen-2.5-14b-instruct", QWEN_2_5_7B: "qwen/qwen-2.5-7b-instruct", QWQ_32B_PREVIEW: "qwen/qwq-32b-preview", // Moonshot models KIMI_K2: "moonshotai/kimi-k2-1t", KIMI_K1: "moonshotai/kimi-k1", // Alibaba models QWEN_VL_MAX: "qwen/qwen-vl-max", MARCO_O1: "alibaba/marco-o1", // Cohere models COMMAND_R_PLUS: "cohere/command-r-plus", COMMAND_R: "cohere/command-r", // AI21 models JAMBA_1_5_LARGE: "ai21/jamba-1-5-large", JAMBA_1_5_MINI: "ai21/jamba-1-5-mini", // Databricks models DBRX_INSTRUCT: "databricks/dbrx-instruct", // Nvidia models LLAMA_3_1_NEMOTRON_70B: "nvidia/llama-3.1-nemotron-70b-instruct", // Perplexity models PERPLEXITY_ONLINE: "perplexity/llama-3.1-sonar-large-128k-online", PERPLEXITY_CHAT: "perplexity/llama-3.1-sonar-large-128k-chat", // 01 models O1_PREVIEW: "openai/o1-preview", O1_MINI: "openai/o1-mini", // Specialized models ANALYSIS_MODEL: "google/gemini-2.5-pro", SYNTHESIS_MODEL: "google/gemini-2.5-pro", CODE_MODEL: "qwen/qwen-2.5-coder-32b-instruct", DEBATE_MODEL: "x-ai/grok-3-beta", REASONING_MODEL: "deepseek/deepseek-r1", VISION_MODEL: "x-ai/grok-2-vision", }; // Model capabilities and context limits export const MODEL_CONTEXT_LIMITS = { // Google models "google/gemini-2.5-pro": 2097152, // 2M tokens "google/gemini-2.5-flash": 1048576, // 1M tokens "google/gemini-pro": 32768, "google/gemini-2.0-flash-thinking-exp": 32768, // OpenAI models "openai/gpt-4o": 128000, "openai/gpt-4o-mini": 128000, "openai/gpt-4-turbo": 128000, "openai/gpt-4": 8192, "openai/o1-preview": 128000, "openai/o1-mini": 128000, // Anthropic models "anthropic/claude-3-opus": 200000, "anthropic/claude-3-sonnet": 200000, "anthropic/claude-3-haiku": 200000, "anthropic/claude-3.5-sonnet": 200000, "anthropic/claude-3.5-sonnet-20241022": 200000, // xAI models "x-ai/grok-2": 131072, "x-ai/grok-2-vision": 131072, "x-ai/grok-3-beta": 131072, "x-ai/grok-4-beta": 131072, // Meta models "meta-llama/llama-3.3-70b-instruct": 128000, "meta-llama/llama-3.1-405b-instruct": 128000, "meta-llama/llama-3.1-70b-instruct": 128000, "meta-llama/llama-3.2-90b-vision-instruct": 128000, // Moonshot models "moonshotai/kimi-k2-1t": 1000000, // 1M context window "moonshotai/kimi-k1": 128000, // DeepSeek models "deepseek/deepseek-v3": 128000, "deepseek/deepseek-r1": 128000, "deepseek/deepseek-coder-v2-instruct": 128000, // Qwen models "qwen/qwen-2.5-coder-32b-instruct": 128000, "qwen/qwen-2.5-72b-instruct": 128000, "qwen/qwq-32b-preview": 32768, // Mistral models "mistralai/mistral-large-2411": 128000, "mistralai/codestral-2501": 256000, "mistralai/mixtral-8x22b-instruct": 65536, }; // Model specializations export const MODEL_SPECIALIZATIONS = { coding: [ "qwen/qwen-2.5-coder-32b-instruct", "mistralai/codestral-2501", "deepseek/deepseek-coder-v2-instruct", "openai/gpt-4o", "anthropic/claude-3.5-sonnet", ], reasoning: [ "deepseek/deepseek-r1", "openai/o1-preview", "qwen/qwq-32b-preview", "alibaba/marco-o1", "x-ai/grok-4-beta", ], vision: [ "x-ai/grok-2-vision", "google/gemini-pro-vision", "meta-llama/llama-3.2-90b-vision-instruct", "qwen/qwen-vl-max", "mistralai/pixtral-large-2411", ], longContext: [ "google/gemini-2.5-pro", "google/gemini-2.5-flash", "moonshotai/kimi-k2-1t", "mistralai/codestral-2501", "anthropic/claude-3-opus", ], efficiency: [ "meta-llama/llama-3.2-1b-instruct", "meta-llama/llama-3.2-3b-instruct", "deepseek/deepseek-v3", "google/gemini-2.5-flash", ], online: [ "perplexity/llama-3.1-sonar-large-128k-online", "perplexity/llama-3.1-sonar-large-128k-chat", ], }; // Fallback models when primary ones fail export const FALLBACK_MODELS = { // Google fallbacks "google/gemini-pro": "google/gemini-2.5-pro", "google/gemini-pro-vision": "google/gemini-2.5-pro", "google/gemini-2.0-flash-exp": "google/gemini-2.5-flash", "google/gemini-2.5-flash-exp": "google/gemini-2.5-flash", // xAI fallbacks "x-ai/grok-4-beta": "x-ai/grok-3-beta", "x-ai/grok-3-beta": "x-ai/grok-2", // DeepSeek fallbacks "deepseek/deepseek-r1": "deepseek/deepseek-v3", "deepseek/deepseek-r1-lite-preview": "deepseek/deepseek-chat", // Anthropic fallbacks "anthropic/claude-3.5-sonnet-20241022": "anthropic/claude-3.5-sonnet", // Moonshot fallbacks "moonshotai/kimi-k2-1t": "moonshotai/kimi-k1", }; // Model categories for classification export const MODEL_CATEGORIES = { GENERAL: 'general', CODING: 'coding', VISION: 'vision', REASONING: 'reasoning', CREATIVE: 'creative', ANALYSIS: 'analysis' }; // Model pricing tiers (for cost optimization) export const MODEL_PRICING_TIERS = { premium: [ "openai/o1-preview", "anthropic/claude-3-opus", "moonshotai/kimi-k2-1t", ], standard: [ "openai/gpt-4o", "anthropic/claude-3.5-sonnet", "x-ai/grok-3-beta", "google/gemini-2.5-pro", ], economy: [ "openai/gpt-4o-mini", "google/gemini-2.5-flash", "meta-llama/llama-3.2-3b-instruct", "deepseek/deepseek-v3", ], }; //# sourceMappingURL=models.js.map