UNPKG

node-llama-cpp

Version:

Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level

26 lines (25 loc) 1.11 kB
import retry from "async-retry"; import { GgufReadOffset } from "../utils/GgufReadOffset.js"; import { ModelFileAccessTokens } from "../../utils/modelFileAccesTokens.js"; import { GgufFileReader } from "./GgufFileReader.js"; type GgufFetchFileReaderOptions = { url: string; retryOptions?: retry.Options; headers?: Record<string, string>; signal?: AbortSignal; tokens?: ModelFileAccessTokens; }; export declare class GgufNetworkFetchFileReader extends GgufFileReader { readonly url: string; readonly retryOptions: retry.Options; readonly headers: Record<string, string>; readonly tokens?: ModelFileAccessTokens; private readonly _signal?; private _tryHeaders; constructor({ url, retryOptions, headers, tokens, signal }: GgufFetchFileReaderOptions); readByteRange(offset: number | GgufReadOffset, length: number): Buffer<ArrayBuffer> | Promise<Buffer<ArrayBuffer>>; protected ensureHasByteRange(offset: number | GgufReadOffset, length: number): Promise<void> | undefined; private _fetchToExpandBufferUpToOffset; private _fetchByteRange; } export {};