UNPKG

node-llama-cpp

Version:

Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level

23 lines (22 loc) 1.01 kB
export declare function withProgressLog<T>({ loadingText, successText, failText, liveUpdates, statusIcons, initialPercentage, initialProgressBarText, progressBarLength, minPercentageChangeForNonLiveUpdates, eta, etaUpdateInterval, noProgress, progressFractionDigits, noSuccessLiveStatus, liveCtrlCSendsAbortSignal }: { loadingText: string; successText: string; failText: string; liveUpdates?: boolean; statusIcons?: boolean; initialPercentage?: number; initialProgressBarText?: string; progressBarLength?: number; minPercentageChangeForNonLiveUpdates?: number; eta?: boolean; etaUpdateInterval?: number; noProgress?: boolean; progressFractionDigits?: boolean; noSuccessLiveStatus?: boolean; liveCtrlCSendsAbortSignal?: boolean; }, callback: (progressUpdater: ProgressUpdater) => Promise<T>): Promise<T>; type ProgressUpdater = { setProgress(percentage: number, progressText?: string): ProgressUpdater; abortSignal?: AbortSignal; }; export {};