node-llama-cpp
Version:
Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level
14 lines (13 loc) • 802 B
TypeScript
import { LlamaGpuType } from "../types.js";
/**
* Get the list of GPU types that can be used with `getLlama` on the current machine.
*
* When passing `"supported"`, only the GPU types that have the
* necessary libraries and drivers installed on the current machine will be returned.
* All of these GPU types have prebuilt binaries for the current platform and architecture.
*
* When passing `"allValid"`, all GPU types that are compatible with the current OS and architecture will be returned.
* Some of these GPU types may not have prebuilt binaries for the current platform and architecture,
* as some of them are inadvisable for the current machine (like CUDA on an x64 Mac machine).
*/
export declare function getLlamaGpuTypes(include: "supported" | "allValid"): Promise<LlamaGpuType[]>;