UNPKG

@huggingface/inference

Version:

Typescript client for the Hugging Face Inference Providers and Inference Endpoints

40 lines (39 loc) 1.5 kB
"use strict"; Object.defineProperty(exports, "__esModule", { value: true }); exports.GroqConversationalTask = exports.GroqTextGenerationTask = void 0; const providerHelper_js_1 = require("./providerHelper.js"); /** * See the registered mapping of HF model ID => Groq model ID here: * * https://huggingface.co/api/partners/groq/models * * This is a publicly available mapping. * * If you want to try to run inference for a new model locally before it's registered on huggingface.co, * you can add it to the dictionary "HARDCODED_MODEL_ID_MAPPING" in consts.ts, for dev purposes. * * - If you work at Groq and want to update this mapping, please use the model mapping API we provide on huggingface.co * - If you're a community member and want to add a new supported HF model to Groq, please open an issue on the present repo * and we will tag Groq team members. * * Thanks! */ const GROQ_API_BASE_URL = "https://api.groq.com"; class GroqTextGenerationTask extends providerHelper_js_1.BaseTextGenerationTask { constructor() { super("groq", GROQ_API_BASE_URL); } makeRoute() { return "/openai/v1/chat/completions"; } } exports.GroqTextGenerationTask = GroqTextGenerationTask; class GroqConversationalTask extends providerHelper_js_1.BaseConversationalTask { constructor() { super("groq", GROQ_API_BASE_URL); } makeRoute() { return "/openai/v1/chat/completions"; } } exports.GroqConversationalTask = GroqConversationalTask;