node-llama-cpp
Version:
Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level
26 lines (25 loc) • 1.11 kB
TypeScript
import retry from "async-retry";
import { GgufReadOffset } from "../utils/GgufReadOffset.js";
import { ModelFileAccessTokens } from "../../utils/modelFileAccesTokens.js";
import { GgufFileReader } from "./GgufFileReader.js";
type GgufFetchFileReaderOptions = {
url: string;
retryOptions?: retry.Options;
headers?: Record<string, string>;
signal?: AbortSignal;
tokens?: ModelFileAccessTokens;
};
export declare class GgufNetworkFetchFileReader extends GgufFileReader {
readonly url: string;
readonly retryOptions: retry.Options;
readonly headers: Record<string, string>;
readonly tokens?: ModelFileAccessTokens;
private readonly _signal?;
private _tryHeaders;
constructor({ url, retryOptions, headers, tokens, signal }: GgufFetchFileReaderOptions);
readByteRange(offset: number | GgufReadOffset, length: number): Buffer<ArrayBuffer> | Promise<Buffer<ArrayBuffer>>;
protected ensureHasByteRange(offset: number | GgufReadOffset, length: number): Promise<void> | undefined;
private _fetchToExpandBufferUpToOffset;
private _fetchByteRange;
}
export {};