inference-server
Version:
Libraries and server to build AI applications. Adapters to various native bindings allowing local inference. Integrate it with your application, or use as a microservice.
14 lines (13 loc) • 520 B
TypeScript
import { ChatHistoryItem, ChatModelFunctions, LlamaChatResponse, LlamaChatResponseFunctionCall } from 'node-llama-cpp';
export interface LlamaChatResult<T extends ChatModelFunctions = any> {
responseText: string | null;
functionCalls?: LlamaChatResponseFunctionCall<T>[];
stopReason: LlamaChatResponse['metadata']['stopReason'];
}
export type ContextShiftStrategy = ((options: {
chatHistory: ChatHistoryItem[];
metadata: any;
}) => {
chatHistory: ChatHistoryItem[];
metadata: any;
}) | null;