UNPKG

inference-server

Version:

Libraries and server to build AI applications. Adapters to various native bindings allowing local inference. Integrate it with your application, or use as a microservice.

26 lines (22 loc) 747 B
import type { IncomingMessage, ServerResponse } from 'node:http' import type { OpenAI } from 'openai' import type { InferenceServer } from '#package/server' // handler for v1/models // https://platform.openai.com/docs/api-reference/models/list export function createModelsHandler(inferenceServer: InferenceServer) { return async (req: IncomingMessage, res: ServerResponse) => { const models = inferenceServer.store.getStatus() const data: OpenAI.Model[] = Object.entries(models).map( ([id, info]) => { return { object: 'model', id, created: 0, owned_by: info.engine, } }, ) res.writeHead(200, { 'Content-Type': 'application/json' }) res.end(JSON.stringify({ object: 'list', data }, null, 2)) } }