node-llama-cpp
Version:
Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level
29 lines (28 loc) • 1.13 kB
TypeScript
import { GbnfTerminal } from "../GbnfTerminal.js";
import { GbnfGrammarGenerator } from "../GbnfGrammarGenerator.js";
import { GbnfJsonScopeState } from "../utils/GbnfJsonScopeState.js";
import { GbnfString } from "./GbnfString.js";
import { GbnfStringValue } from "./GbnfStringValue.js";
export declare class GbnfObjectMap extends GbnfTerminal {
readonly fields: Array<Readonly<{
key: GbnfString | GbnfStringValue;
value: GbnfTerminal;
required: true;
}>>;
readonly additionalProperties?: GbnfTerminal;
readonly minProperties: number;
readonly maxProperties?: number;
readonly scopeState: GbnfJsonScopeState;
constructor({ fields, additionalProperties, minProperties, maxProperties, scopeState }: {
fields: Array<Readonly<{
key: GbnfString | GbnfStringValue;
value: GbnfTerminal;
required: true;
}>>;
additionalProperties?: GbnfTerminal;
minProperties?: number;
maxProperties?: number;
scopeState?: GbnfJsonScopeState;
});
getGrammar(grammarGenerator: GbnfGrammarGenerator): string;
}