@mlc-ai/web-llm
Version:
Hardware accelerated language model chats on browsers
10 lines • 1.15 kB
TypeScript
export { ModelRecord, AppConfig, ChatOptions, MLCEngineConfig, GenerationConfig, ModelType, prebuiltAppConfig, modelVersion, modelLibURLPrefix, functionCallingModelIds, } from "./config";
export { InitProgressCallback, InitProgressReport, MLCEngineInterface, LogitProcessor, LogLevel, } from "./types";
export { MLCEngine, CreateMLCEngine } from "./engine";
export { hasModelInCache, deleteChatConfigInCache, deleteModelAllInfoInCache, deleteModelWasmInCache, deleteModelInCache, } from "./cache_util";
export { WebWorkerMLCEngineHandler, WebWorkerMLCEngine, CreateWebWorkerMLCEngine, } from "./web_worker";
export { WorkerRequest, WorkerResponse, CustomRequestParams } from "./message";
export { ServiceWorkerMLCEngineHandler, ServiceWorkerMLCEngine, CreateServiceWorkerMLCEngine, } from "./service_worker";
export { ServiceWorkerMLCEngineHandler as ExtensionServiceWorkerMLCEngineHandler, ServiceWorkerMLCEngine as ExtensionServiceWorkerMLCEngine, CreateServiceWorkerMLCEngine as CreateExtensionServiceWorkerMLCEngine, } from "./extension_service_worker";
export * from "./openai_api_protocols/index";
//# sourceMappingURL=index.d.ts.map