UNPKG

@assistant-ui/react

Version:

Typescript/React library for AI Chat

45 lines 2.14 kB
import type { AppendMessage } from "../../types"; import type { ChatModelAdapter } from "./ChatModelAdapter"; import { LocalRuntimeOptionsBase } from "./LocalRuntimeOptions"; import { AddToolResultOptions, ThreadSuggestion, ThreadRuntimeCore, StartRunConfig, ResumeRunConfig } from "../core/ThreadRuntimeCore"; import { BaseThreadRuntimeCore } from "../core/BaseThreadRuntimeCore"; import { ModelContextProvider } from "../../model-context"; export declare class LocalThreadRuntimeCore extends BaseThreadRuntimeCore implements ThreadRuntimeCore { readonly capabilities: { switchToBranch: boolean; edit: boolean; reload: boolean; cancel: boolean; unstable_copy: boolean; speech: boolean; attachments: boolean; feedback: boolean; }; private abortController; readonly isDisabled = false; private _suggestions; private _suggestionsController; get suggestions(): readonly ThreadSuggestion[]; get adapters(): { chatModel: ChatModelAdapter; history?: import("../adapters/thread-history/ThreadHistoryAdapter").ThreadHistoryAdapter | undefined; attachments?: import("..").AttachmentAdapter | undefined; speech?: import("..").SpeechSynthesisAdapter | undefined; feedback?: import("..").FeedbackAdapter | undefined; suggestion?: import("..").SuggestionAdapter | undefined; }; constructor(contextProvider: ModelContextProvider, options: LocalRuntimeOptionsBase); private _options; private _lastRunConfig; get extras(): undefined; __internal_setOptions(options: LocalRuntimeOptionsBase): void; private _loadPromise; __internal_load(): Promise<void>; append(message: AppendMessage): Promise<void>; resumeRun({ stream, ...startConfig }: ResumeRunConfig): Promise<void>; startRun({ parentId, runConfig }: StartRunConfig, runCallback?: ChatModelAdapter["run"]): Promise<void>; private performRoundtrip; cancelRun(): void; addToolResult({ messageId, toolCallId, result, }: AddToolResultOptions): void; } //# sourceMappingURL=LocalThreadRuntimeCore.d.ts.map