UNPKG

@lobehub/chat

Version:

Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.

75 lines (66 loc) 2.36 kB
import type { ChatModelCard } from '@lobechat/types'; import { ModelProvider } from 'model-bank'; import { OpenAICompatibleFactoryOptions, createOpenAICompatibleRuntime, } from '../../core/openaiCompatibleFactory'; import { resolveParameters } from '../../core/parameterResolver'; export interface MistralModelCard { capabilities: { function_calling: boolean; vision: boolean; }; description: string; id: string; max_context_length: number; } export const params = { baseURL: 'https://api.mistral.ai/v1', chatCompletion: { // Mistral API does not support stream_options: { include_usage: true } // refs: https://github.com/lobehub/lobe-chat/issues/6825 excludeUsage: true, handlePayload: (payload) => { // Resolve parameters with normalization const resolvedParams = resolveParameters( { max_tokens: payload.max_tokens, temperature: payload.temperature, top_p: payload.top_p }, { normalizeTemperature: true }, ); return { ...resolvedParams, messages: payload.messages as any, model: payload.model, stream: true, ...(payload.tools && { tools: payload.tools }), }; }, noUserId: true, }, debug: { chatCompletion: () => process.env.DEBUG_MISTRAL_CHAT_COMPLETION === '1', }, models: async ({ client }) => { const { LOBE_DEFAULT_MODEL_LIST } = await import('model-bank'); const modelsPage = (await client.models.list()) as any; const modelList: MistralModelCard[] = modelsPage.data; return modelList .map((model) => { const knownModel = LOBE_DEFAULT_MODEL_LIST.find( (m) => model.id.toLowerCase() === m.id.toLowerCase(), ); return { contextWindowTokens: model.max_context_length, description: model.description, displayName: knownModel?.displayName ?? undefined, enabled: knownModel?.enabled || false, functionCall: model.capabilities.function_calling, id: model.id, reasoning: knownModel?.abilities?.reasoning || false, vision: model.capabilities.vision, }; }) .filter(Boolean) as ChatModelCard[]; }, provider: ModelProvider.Mistral, } satisfies OpenAICompatibleFactoryOptions; export const LobeMistralAI = createOpenAICompatibleRuntime(params);