UNPKG

node-llama-cpp

Version:

Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level

37 lines (36 loc) 1.79 kB
import { GgufReadOffset } from "../utils/GgufReadOffset.js"; import { Promisable } from "../../utils/transformPromisable.js"; export declare const valueTypeToBytesToRead: { readonly uint8: 1; readonly uint16: 2; readonly uint32: 4; readonly uint64: 8; readonly int8: 1; readonly int16: 2; readonly int32: 4; readonly int64: 8; readonly float32: 4; readonly float64: 8; readonly bool: 1; }; export declare abstract class GgufFileReader { protected _buffer: Buffer<ArrayBuffer>; abstract readByteRange(offset: number | GgufReadOffset, length: number): Promisable<Buffer>; protected abstract ensureHasByteRange(offset: number | GgufReadOffset, length: number): Promisable<void>; readUint8(offset: number | GgufReadOffset): Promisable<number>; readUint16(offset: number | GgufReadOffset): Promisable<number>; readUint32(offset: number | GgufReadOffset): Promisable<number>; readUint64(offset: number | GgufReadOffset): Promisable<bigint>; readInt8(offset: number | GgufReadOffset): Promisable<number>; readInt16(offset: number | GgufReadOffset): Promisable<number>; readInt32(offset: number | GgufReadOffset): Promisable<number>; readInt64(offset: number | GgufReadOffset): Promisable<bigint>; readFloat32(offset: number | GgufReadOffset): Promisable<number>; readFloat64(offset: number | GgufReadOffset): Promisable<number>; readBool(offset: number | GgufReadOffset): Promisable<boolean>; readString(offset: number | GgufReadOffset): Promisable<string>; readStringWithLength(offset: number | GgufReadOffset, length: number): Promisable<string>; protected _addToBuffer(buffer: Buffer): void; private _withBufferRead; static castNumberIfSafe(value: bigint): number | bigint; }