UNPKG

@lobehub/chat

Version:

Lobe Chat - an open-source, high-performance chatbot framework that supports speech synthesis, multimodal, and extensible Function Call plugin system. Supports one-click free deployment of your private ChatGPT/LLM web application.

190 lines (187 loc) 6.03 kB
import { ModelProviderCard } from '@/types/llm'; // ref https://console.groq.com/docs/tool-use const Groq: ModelProviderCard = { chatModels: [ // TODO: During preview launch, Groq is limiting 3.2 models to max_tokens of 8k. { contextWindowTokens: 131_072, description: 'Meta Llama 3.3 多语言大语言模型 ( LLM ) 是 70B(文本输入/文本输出)中的预训练和指令调整生成模型。 Llama 3.3 指令调整的纯文本模型针对多语言对话用例进行了优化,并且在常见行业基准上优于许多可用的开源和封闭式聊天模型。', displayName: 'Llama 3.3 70B', enabled: true, functionCall: true, id: 'llama-3.3-70b-versatile', maxOutput: 8192, pricing: { input: 0.05, output: 0.08, }, }, { contextWindowTokens: 8192, description: 'Llama 3.2 旨在处理结合视觉和文本数据的任务。它在图像描述和视觉问答等任务中表现出色,跨越了语言生成和视觉推理之间的鸿沟。', displayName: 'Llama 3.2 11B Vision (Preview)', enabled: true, id: 'llama-3.2-11b-vision-preview', maxOutput: 8192, pricing: { input: 0.05, output: 0.08, }, vision: true, }, { contextWindowTokens: 8192, description: 'Llama 3.2 旨在处理结合视觉和文本数据的任务。它在图像描述和视觉问答等任务中表现出色,跨越了语言生成和视觉推理之间的鸿沟。', displayName: 'Llama 3.2 90B Vision (Preview)', enabled: true, id: 'llama-3.2-90b-vision-preview', maxOutput: 8192, pricing: { input: 0.59, output: 0.79, }, vision: true, }, { contextWindowTokens: 131_072, description: 'Llama 3.1 8B 是一款高效能模型,提供了快速的文本生成能力,非常适合需要大规模效率和成本效益的应用场景。', displayName: 'Llama 3.1 8B', enabled: true, functionCall: true, id: 'llama-3.1-8b-instant', maxOutput: 8192, pricing: { input: 0.05, output: 0.08, }, }, { contextWindowTokens: 131_072, description: 'Llama 3.1 70B 提供更强大的AI推理能力,适合复杂应用,支持超多的计算处理并保证高效和准确率。', displayName: 'Llama 3.1 70B', enabled: true, functionCall: true, id: 'llama-3.1-70b-versatile', maxOutput: 8192, pricing: { input: 0.59, output: 0.79, }, }, /* // Offline due to overwhelming demand! Stay tuned for updates. { displayName: 'Llama 3.1 405B', functionCall: true, id: 'llama-3.1-405b-reasoning', tokens: 8_192, }, */ { contextWindowTokens: 8192, description: 'Llama 3 Groq 8B Tool Use 是针对高效工具使用优化的模型,支持快速并行计算。', displayName: 'Llama 3 Groq 8B Tool Use (Preview)', functionCall: true, id: 'llama3-groq-8b-8192-tool-use-preview', pricing: { input: 0.19, output: 0.19, }, }, { contextWindowTokens: 8192, description: 'Llama 3 Groq 70B Tool Use 提供强大的工具调用能力,支持复杂任务的高效处理。', displayName: 'Llama 3 Groq 70B Tool Use (Preview)', functionCall: true, id: 'llama3-groq-70b-8192-tool-use-preview', pricing: { input: 0.89, output: 0.89, }, }, { contextWindowTokens: 8192, description: 'Meta Llama 3 8B 带来优质的推理效能,适合多场景应用需求。', displayName: 'Meta Llama 3 8B', functionCall: true, id: 'llama3-8b-8192', pricing: { input: 0.05, output: 0.08, }, }, { contextWindowTokens: 8192, description: 'Meta Llama 3 70B 提供无与伦比的复杂性处理能力,为高要求项目量身定制。', displayName: 'Meta Llama 3 70B', functionCall: true, id: 'llama3-70b-8192', pricing: { input: 0.59, output: 0.79, }, }, { contextWindowTokens: 8192, description: 'Gemma 2 9B 是一款优化用于特定任务和工具整合的模型。', displayName: 'Gemma 2 9B', enabled: true, functionCall: true, id: 'gemma2-9b-it', pricing: { input: 0.2, output: 0.2, }, }, { contextWindowTokens: 8192, description: 'Gemma 7B 适合中小规模任务处理,兼具成本效益。', displayName: 'Gemma 7B', functionCall: true, id: 'gemma-7b-it', pricing: { input: 0.07, output: 0.07, }, }, { contextWindowTokens: 32_768, description: 'Mixtral 8x7B 提供高容错的并行计算能力,适合复杂任务。', displayName: 'Mixtral 8x7B', functionCall: true, id: 'mixtral-8x7b-32768', pricing: { input: 0.24, output: 0.24, }, }, { contextWindowTokens: 4096, description: 'LLaVA 1.5 7B 提供视觉处理能力融合,通过视觉信息输入生成复杂输出。', displayName: 'LLaVA 1.5 7B', id: 'llava-v1.5-7b-4096-preview', vision: true, }, ], checkModel: 'llama-3.1-8b-instant', description: 'Groq 的 LPU 推理引擎在最新的独立大语言模型(LLM)基准测试中表现卓越,以其惊人的速度和效率重新定义了 AI 解决方案的标准。Groq 是一种即时推理速度的代表,在基于云的部署中展现了良好的性能。', id: 'groq', modelList: { showModelFetcher: true }, modelsUrl: 'https://console.groq.com/docs/models', name: 'Groq', settings: { proxyUrl: { placeholder: 'https://api.groq.com/openai/v1', }, sdkType: 'openai', showModelFetcher: true, }, url: 'https://groq.com', }; export default Groq;