UNPKG

@huggingface/inference

Version:

Typescript client for the Hugging Face Inference Providers and Inference Endpoints

36 lines (35 loc) 946 B
"use strict"; Object.defineProperty(exports, "__esModule", { value: true }); exports.HARDCODED_MODEL_INFERENCE_MAPPING = void 0; /** * If you want to try to run inference for a new model locally before it's registered on huggingface.co * for a given Inference Provider, * you can add it to the following dictionary, for dev purposes. * * We also inject into this dictionary from tests. */ exports.HARDCODED_MODEL_INFERENCE_MAPPING = { /** * "HF model ID" => "Model ID on Inference Provider's side" * * Example: * "Qwen/Qwen2.5-Coder-32B-Instruct": "Qwen2.5-Coder-32B-Instruct", */ "black-forest-labs": {}, cerebras: {}, cohere: {}, "fal-ai": {}, "featherless-ai": {}, "fireworks-ai": {}, groq: {}, "hf-inference": {}, hyperbolic: {}, nebius: {}, novita: {}, nscale: {}, openai: {}, ovhcloud: {}, replicate: {}, sambanova: {}, together: {}, };