UNPKG

node-llama-cpp

Version:

Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level

9 lines 544 B
import { includesText } from "../../utils/includesText.js"; import { getModelLinageNames } from "./getModelLinageNames.js"; export function isLlama3_2LightweightModel(options) { const isLlama3_2 = getModelLinageNames(options.fileInfo?.metadata) .some((modelNames) => includesText(modelNames, ["llama 3.2", "llama-3.2", "llama3.2"])); const isSmallModel = ["1B", "3B"].includes(options.fileInfo?.metadata?.general?.size_label ?? ""); return isLlama3_2 && isSmallModel; } //# sourceMappingURL=isLlama3_2LightweightModel.js.map