UNPKG

node-llama-cpp

Version:

Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level

7 lines (6 loc) 411 B
import { ChatWrapper } from "../ChatWrapper.js"; import { ChatWrapperGenerateContextStateOptions, ChatWrapperGeneratedContextState } from "../types.js"; export declare class ChatMLChatWrapper extends ChatWrapper { readonly wrapperName: string; generateContextState({ chatHistory, availableFunctions, documentFunctionParams }: ChatWrapperGenerateContextStateOptions): ChatWrapperGeneratedContextState; }