@azure-rest/ai-inference
Version:
Inference API for Azure-supported AI models
149 lines • 6.28 kB
TypeScript
/**
* THIS IS AN AUTO-GENERATED FILE - DO NOT EDIT!
*
* Any changes you make here may be lost.
*
* If you need to make changes, please do so in the original source file, \{project-root\}/sources/custom
*/
/** A function tool call requested by the AI model. */
export interface ChatCompletionsToolCallOutput {
/** The ID of the tool call. */
id: string;
/** The type of tool call. Currently, only `function` is supported. */
type: "function";
/** The details of the function call requested by the AI model. */
function: FunctionCallOutput;
}
/** The name and arguments of a function that should be called, as generated by the model. */
export interface FunctionCallOutput {
/** The name of the function to call. */
name: string;
/**
* The arguments to call the function with, as generated by the model in JSON format.
* Note that the model does not always generate valid JSON, and may hallucinate parameters
* not defined by your function schema. Validate the arguments in your code before calling
* your function.
*/
arguments: string;
}
/**
* Representation of the response data from a chat completions request.
* Completions support a wide variety of tasks and generate text that continues from or "completes"
* provided prompt data.
*/
export interface ChatCompletionsOutput {
/** A unique identifier associated with this chat completions response. */
id: string;
/**
* The first timestamp associated with generation activity for this completions response,
* represented as seconds since the beginning of the Unix epoch of 00:00 on 1 Jan 1970.
*/
created: number;
/** The model used for the chat completion. */
model: string;
/**
* The collection of completions choices associated with this completions response.
* Generally, `n` choices are generated per provided prompt with a default value of 1.
* Token limits and other settings may limit the number of choices generated.
*/
choices: Array<ChatChoiceOutput>;
/** Usage information for tokens processed and generated as part of this completions operation. */
usage: CompletionsUsageOutput;
}
/**
* The representation of a single prompt completion as part of an overall chat completions request.
* Generally, `n` choices are generated per provided prompt with a default value of 1.
* Token limits and other settings may limit the number of choices generated.
*/
export interface ChatChoiceOutput {
/** The ordered index associated with this chat completions choice. */
index: number;
/** The reason that this chat completions choice completed its generated. */
finish_reason: CompletionsFinishReasonOutput | null;
/** The chat message for a given chat completions prompt. */
message: ChatResponseMessageOutput;
}
/** A representation of a chat message as received in a response. */
export interface ChatResponseMessageOutput {
/**
* The chat role associated with the message.
*
* Possible values: "system", "user", "assistant", "tool", "developer"
*/
role: ChatRoleOutput;
/** The content of the message. */
content: string | null;
/**
* The tool calls that must be resolved and have their outputs appended to subsequent input messages for the chat
* completions request to resolve as configured.
*/
tool_calls?: Array<ChatCompletionsToolCallOutput>;
}
/**
* Representation of the token counts processed for a completions request.
* Counts consider all tokens across prompts, choices, choice alternates, best_of generations, and
* other consumers.
*/
export interface CompletionsUsageOutput {
/** The number of tokens generated across all completions emissions. */
completion_tokens: number;
/** The number of tokens in the provided prompts for the completions request. */
prompt_tokens: number;
/** The total number of tokens processed for the completions request and response. */
total_tokens: number;
}
/** Represents some basic information about the AI model. */
export interface ModelInfoOutput {
/** The name of the AI model. For example: `Phi21` */
model_name: string;
/**
* The type of the AI model. A Unique identifier for the profile.
*
* Possible values: "embeddings", "image_generation", "text_generation", "image_embeddings", "audio_generation", "chat_completion"
*/
model_type: ModelTypeOutput;
/** The model provider name. For example: `Microsoft Research` */
model_provider_name: string;
}
/**
* Representation of the response data from an embeddings request.
* Embeddings measure the relatedness of text strings and are commonly used for search, clustering,
* recommendations, and other similar scenarios.
*/
export interface EmbeddingsResultOutput {
/** Unique identifier for the embeddings result. */
id: string;
/** Embedding values for the prompts submitted in the request. */
data: Array<EmbeddingItemOutput>;
/** Usage counts for tokens input using the embeddings API. */
usage: EmbeddingsUsageOutput;
/** The model ID used to generate this result. */
model: string;
}
/** Representation of a single embeddings relatedness comparison. */
export interface EmbeddingItemOutput {
/**
* List of embedding values for the input prompt. These represent a measurement of the
* vector-based relatedness of the provided input. Or a base64 encoded string of the embedding vector.
*/
embedding: string | number[];
/** Index of the prompt to which the EmbeddingItem corresponds. */
index: number;
}
/** Measurement of the amount of tokens used in this request and response. */
export interface EmbeddingsUsageOutput {
/** Number of tokens in the request. */
prompt_tokens: number;
/**
* Total number of tokens transacted in this request/response. Should equal the
* number of tokens in the request.
*/
total_tokens: number;
}
/** Alias for ChatRoleOutput */
export type ChatRoleOutput = string;
/** Alias for CompletionsFinishReasonOutput */
export type CompletionsFinishReasonOutput = string;
/** Alias for ModelTypeOutput */
export type ModelTypeOutput = string;
//# sourceMappingURL=outputModels.d.ts.map