@huggingface/inference
Version:
Typescript client for the Hugging Face Inference Providers and Inference Endpoints
40 lines (39 loc) • 1.52 kB
JavaScript
import { BaseConversationalTask, BaseTextGenerationTask } from "./providerHelper.js";
import { omit } from "../utils/omit.js";
import { InferenceClientProviderOutputError } from "../errors.js";
const FEATHERLESS_API_BASE_URL = "https://api.featherless.ai";
export class FeatherlessAIConversationalTask extends BaseConversationalTask {
constructor() {
super("featherless-ai", FEATHERLESS_API_BASE_URL);
}
}
export class FeatherlessAITextGenerationTask extends BaseTextGenerationTask {
constructor() {
super("featherless-ai", FEATHERLESS_API_BASE_URL);
}
preparePayload(params) {
return {
model: params.model,
...omit(params.args, ["inputs", "parameters"]),
...(params.args.parameters
? {
max_tokens: params.args.parameters.max_new_tokens,
...omit(params.args.parameters, "max_new_tokens"),
}
: undefined),
prompt: params.args.inputs,
};
}
async getResponse(response) {
if (typeof response === "object" &&
"choices" in response &&
Array.isArray(response?.choices) &&
typeof response?.model === "string") {
const completion = response.choices[0];
return {
generated_text: completion.text,
};
}
throw new InferenceClientProviderOutputError("Received malformed response from Featherless AI text generation API");
}
}