node-llama-cpp
Version:
Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level
7 lines (6 loc) • 365 B
TypeScript
import { Token, Tokenizer } from "../types.js";
import { StopGenerationDetector } from "./StopGenerationDetector.js";
export declare function getQueuedTokensBeforeStopTrigger(triggeredStops: ReturnType<typeof StopGenerationDetector["prototype"]["getTriggeredStops"]>, partiallyFreeTokens: {
tokens: Token[];
text: string;
}, tokenizer: Tokenizer): Token[];