UNPKG

node-llama-cpp

Version:

Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level

12 lines 669 B
import { resolveModelDestination } from "../../utils/resolveModelDestination.js"; export function resolveModelRecommendationFileOptions(modelRecommendation) { return modelRecommendation.fileOptions.map((fileOption) => { const resolvedModelDestination = resolveModelDestination(fileOption, true); if (resolvedModelDestination.type === "file") throw new Error(`File option "${fileOption}" is not a valid model URI`); if (resolvedModelDestination.type === "uri") return resolvedModelDestination.uri; return resolvedModelDestination.url; }); } //# sourceMappingURL=resolveModelRecommendationFileOptions.js.map