inference-server
Version:
Libraries and server to build AI applications. Adapters to various native bindings allowing local inference. Integrate it with your application, or use as a microservice.
26 lines (22 loc) • 747 B
text/typescript
import type { IncomingMessage, ServerResponse } from 'node:http'
import type { OpenAI } from 'openai'
import type { InferenceServer } from '#package/server'
// handler for v1/models
// https://platform.openai.com/docs/api-reference/models/list
export function createModelsHandler(inferenceServer: InferenceServer) {
return async (req: IncomingMessage, res: ServerResponse) => {
const models = inferenceServer.store.getStatus()
const data: OpenAI.Model[] = Object.entries(models).map(
([id, info]) => {
return {
object: 'model',
id,
created: 0,
owned_by: info.engine,
}
},
)
res.writeHead(200, { 'Content-Type': 'application/json' })
res.end(JSON.stringify({ object: 'list', data }, null, 2))
}
}