UNPKG

node-llama-cpp

Version:

Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level

51 lines (50 loc) 1.07 kB
export type GgufTensorInfo = { readonly name: string; readonly dimensions: readonly (number | bigint)[]; readonly ggmlType: GgmlType; readonly offset: number | bigint; /** * Adjusted offset relative to the file. * * Added by the GGUF parser - not part of the file's metadata. */ readonly fileOffset: number | bigint; /** * For spliced metadata of multiple file parts, this will be the file part number. * Starts from `1`. * * Added by the GGUF parser - not part of the file's metadata. */ readonly filePart: number; }; export declare const enum GgmlType { F32 = 0, F16 = 1, Q4_0 = 2, Q4_1 = 3, Q4_2 = 4, Q4_3 = 5, Q5_0 = 6, Q5_1 = 7, Q8_0 = 8, Q8_1 = 9, Q2_K = 10, Q3_K = 11, Q4_K = 12, Q5_K = 13, Q6_K = 14, Q8_K = 15, IQ2_XXS = 16, IQ2_XS = 17, IQ3_XXS = 18, IQ1_S = 19, IQ4_NL = 20, IQ3_S = 21, IQ2_S = 22, IQ4_XS = 23, I8 = 24, I16 = 25, I32 = 26, I64 = 27, F64 = 28 }