UNPKG

@n8n/n8n-nodes-langchain

Version:

![Banner image](https://user-images.githubusercontent.com/10284570/173569848-c624317f-42b1-45a6-ab09-f0ea3c247648.png)

269 lines 12.6 kB
"use strict"; Object.defineProperty(exports, "__esModule", { value: true }); exports.LmChatOpenRouter = void 0; const openai_1 = require("@langchain/openai"); const ai_utilities_1 = require("@n8n/ai-utilities"); const n8n_workflow_1 = require("n8n-workflow"); const error_handling_1 = require("../../vendors/OpenAi/helpers/error-handling"); function isOpenAIResponseWithChoices(json) { return (typeof json === 'object' && json !== null && 'choices' in json && Array.isArray(json.choices)); } function createOpenRouterFetch(baseFetch) { return async (input, init) => { const response = await baseFetch(input, init); const contentType = response.headers.get('content-type') ?? ''; if (!contentType.includes('json')) return response; const clone = response.clone(); const json = await response.json(); if (!isOpenAIResponseWithChoices(json)) return clone; const isInvalidArgs = (args) => typeof args !== 'string' || !args.trim(); const toolCallsToFix = json.choices .flatMap((choice) => choice.message?.tool_calls ?? []) .filter((tc) => tc.function && isInvalidArgs(tc.function.arguments)); if (toolCallsToFix.length === 0) return clone; for (const tc of toolCallsToFix) { if (!tc.function) continue; const { arguments: args } = tc.function; const isPlainObject = typeof args === 'object' && args !== null && !Array.isArray(args); tc.function.arguments = isPlainObject ? JSON.stringify(args) : '{}'; } const body = JSON.stringify(json); return new Response(body, { status: response.status, statusText: response.statusText, headers: { 'content-type': contentType }, }); }; } class LmChatOpenRouter { constructor() { this.description = { displayName: 'OpenRouter Chat Model', name: 'lmChatOpenRouter', icon: { light: 'file:openrouter.svg', dark: 'file:openrouter.dark.svg' }, group: ['transform'], version: [1], description: 'For advanced usage with an AI chain', defaults: { name: 'OpenRouter Chat Model', }, codex: { categories: ['AI'], subcategories: { AI: ['Language Models', 'Root Nodes'], 'Language Models': ['Chat Models (Recommended)'], }, resources: { primaryDocumentation: [ { url: 'https://docs.n8n.io/integrations/builtin/cluster-nodes/sub-nodes/n8n-nodes-langchain.lmchatopenrouter/', }, ], }, }, inputs: [], outputs: [n8n_workflow_1.NodeConnectionTypes.AiLanguageModel], outputNames: ['Model'], credentials: [ { name: 'openRouterApi', required: true, }, ], requestDefaults: { ignoreHttpStatusErrors: true, baseURL: '={{ $credentials?.url }}', }, properties: [ (0, ai_utilities_1.getConnectionHintNoticeField)([n8n_workflow_1.NodeConnectionTypes.AiChain, n8n_workflow_1.NodeConnectionTypes.AiAgent]), { displayName: 'If using JSON response format, you must include word "json" in the prompt in your chain or agent. Also, make sure to select latest models released post November 2023.', name: 'notice', type: 'notice', default: '', displayOptions: { show: { '/options.responseFormat': ['json_object'], }, }, }, { displayName: 'Model', name: 'model', type: 'options', description: 'The model which will generate the completion. <a href="https://openrouter.ai/docs/models">Learn more</a>.', typeOptions: { loadOptions: { routing: { request: { method: 'GET', url: '/models', }, output: { postReceive: [ { type: 'rootProperty', properties: { property: 'data', }, }, { type: 'setKeyValue', properties: { name: '={{$responseItem.id}}', value: '={{$responseItem.id}}', }, }, { type: 'sort', properties: { key: 'name', }, }, ], }, }, }, }, routing: { send: { type: 'body', property: 'model', }, }, default: 'openai/gpt-4.1-mini', builderHint: { message: 'Default to a current flagship (e.g. openai/gpt-5.4, anthropic/claude-sonnet-4.6, google/gemini-3.1-pro-preview). Avoid openai/gpt-4o, anthropic/claude-3.x, and other pre-2026 models.', }, }, { displayName: 'Options', name: 'options', placeholder: 'Add Option', description: 'Additional options to add', type: 'collection', default: {}, options: [ { displayName: 'Frequency Penalty', name: 'frequencyPenalty', default: 0, typeOptions: { maxValue: 2, minValue: -2, numberPrecision: 1 }, description: "Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim", type: 'number', }, { displayName: 'Maximum Number of Tokens', name: 'maxTokens', default: -1, description: 'The maximum number of tokens to generate in the completion. Most models have a context length of 2048 tokens (except for the newest models, which support 32,768).', type: 'number', typeOptions: { maxValue: 32768, }, }, { displayName: 'Response Format', name: 'responseFormat', default: 'text', type: 'options', options: [ { name: 'Text', value: 'text', description: 'Regular text response', }, { name: 'JSON', value: 'json_object', description: 'Enables JSON mode, which should guarantee the message the model generates is valid JSON', }, ], }, { displayName: 'Presence Penalty', name: 'presencePenalty', default: 0, typeOptions: { maxValue: 2, minValue: -2, numberPrecision: 1 }, description: "Positive values penalize new tokens based on whether they appear in the text so far, increasing the model's likelihood to talk about new topics", type: 'number', }, { displayName: 'Sampling Temperature', name: 'temperature', default: 0.7, typeOptions: { maxValue: 2, minValue: 0, numberPrecision: 1 }, description: 'Controls randomness: Lowering results in less random completions. As the temperature approaches zero, the model will become deterministic and repetitive.', type: 'number', }, { displayName: 'Timeout', name: 'timeout', default: 360000, description: 'Maximum amount of time a request is allowed to take in milliseconds', type: 'number', }, { displayName: 'Max Retries', name: 'maxRetries', default: 2, description: 'Maximum number of retries to attempt', type: 'number', }, { displayName: 'Top P', name: 'topP', default: 1, typeOptions: { maxValue: 1, minValue: 0, numberPrecision: 1 }, description: 'Controls diversity via nucleus sampling: 0.5 means half of all likelihood-weighted options are considered. We generally recommend altering this or temperature but not both.', type: 'number', }, ], }, ], }; } async supplyData(itemIndex) { const credentials = await this.getCredentials('openRouterApi'); const modelName = this.getNodeParameter('model', itemIndex); const options = this.getNodeParameter('options', itemIndex, {}); const timeout = options.timeout; const configuration = { baseURL: credentials.url, fetch: createOpenRouterFetch(globalThis.fetch), fetchOptions: { dispatcher: (0, ai_utilities_1.getProxyAgent)(credentials.url, { headersTimeout: timeout, bodyTimeout: timeout, }), }, }; const model = new openai_1.ChatOpenAI({ apiKey: credentials.apiKey, model: modelName, ...options, timeout, maxRetries: options.maxRetries ?? 2, configuration, callbacks: [new ai_utilities_1.N8nLlmTracing(this)], modelKwargs: options.responseFormat ? { response_format: { type: options.responseFormat }, } : undefined, onFailedAttempt: (0, ai_utilities_1.makeN8nLlmFailedAttemptHandler)(this, error_handling_1.openAiFailedAttemptHandler), }); return { response: model, }; } } exports.LmChatOpenRouter = LmChatOpenRouter; //# sourceMappingURL=LmChatOpenRouter.node.js.map