UNPKG

node-llama-cpp

Version:

Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level

38 lines (37 loc) 1.49 kB
import { ModelFileAccessTokens } from "./modelFileAccesTokens.js"; export declare const genericFilePartNumber: "{:\n{number}\n:}"; export type ParsedModelUri = ResolvedParsedModelUri | UnresolvedParsedModelUri; export type UnresolvedParsedModelUri = { type: "unresolved"; uri: string; filePrefix: string; baseFilename: string; possibleFullFilenames: (string | `${string}${typeof genericFilePartNumber}${string}`)[]; resolveDetails: { type: "hf"; user: string; model: string; tag: string; }; }; export type ResolvedParsedModelUri = { type: "resolved"; uri: string; resolvedUrl: string; filePrefix: string; filename: string; fullFilename: string; }; export declare function parseModelUri(urlOrUri: string, convertUrlToSupportedUri?: boolean): ParsedModelUri | null; export declare function isModelUri(modelUri: string): boolean; export declare function resolveParsedModelUri(modelUri: ParsedModelUri, options?: { tokens?: ModelFileAccessTokens; signal?: AbortSignal; authorizationHeader?: string; }): Promise<ResolvedParsedModelUri>; export declare function resolveParsedModelUri(modelUri: ParsedModelUri | undefined | null, options?: { tokens?: ModelFileAccessTokens; signal?: AbortSignal; authorizationHeader?: string; }): Promise<ResolvedParsedModelUri | undefined | null>; export declare function getAuthorizationHeader(headers?: Record<string, string>): string | undefined;