UNPKG

node-llama-cpp

Version:

Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level

21 lines 1.02 kB
import { withCliCommandDescriptionDocsUrl } from "../../utils/withCliCommandDescriptionDocsUrl.js"; import { documentationPageUrls } from "../../../config.js"; import { InspectGgufCommand } from "./commands/InspectGgufCommand.js"; import { InspectGpuCommand } from "./commands/InspectGpuCommand.js"; import { InspectMeasureCommand } from "./commands/InspectMeasureCommand.js"; import { InspectEstimateCommand } from "./commands/InspectEstimateCommand.js"; export const InspectCommand = { command: "inspect <command>", describe: withCliCommandDescriptionDocsUrl("Inspect the inner workings of `node-llama-cpp`", documentationPageUrls.CLI.Inspect.index), builder(yargs) { return yargs .command(InspectGpuCommand) .command(InspectGgufCommand) .command(InspectMeasureCommand) .command(InspectEstimateCommand); }, async handler() { // this function must exist, even though we do nothing here } }; //# sourceMappingURL=InspectCommand.js.map