@braintrust/proxy
Version:
A proxy server that load balances across AI providers.
1,744 lines (1,741 loc) • 150 kB
JavaScript
"use strict";
var __create = Object.create;
var __defProp = Object.defineProperty;
var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
var __getOwnPropNames = Object.getOwnPropertyNames;
var __getProtoOf = Object.getPrototypeOf;
var __hasOwnProp = Object.prototype.hasOwnProperty;
var __export = (target, all) => {
for (var name in all)
__defProp(target, name, { get: all[name], enumerable: true });
};
var __copyProps = (to, from, except, desc) => {
if (from && typeof from === "object" || typeof from === "function") {
for (let key of __getOwnPropNames(from))
if (!__hasOwnProp.call(to, key) && key !== except)
__defProp(to, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable });
}
return to;
};
var __toESM = (mod, isNodeMode, target) => (target = mod != null ? __create(__getProtoOf(mod)) : {}, __copyProps(
// If the importer is in node compatibility mode or this is not an ESM
// file that has been converted to a CommonJS file using a Babel-
// compatible transform (i.e. "__esModule" has not been set), then set
// "default" to the CommonJS "module.exports" for node compatibility.
isNodeMode || !mod || !mod.__esModule ? __defProp(target, "default", { value: mod, enumerable: true }) : target,
mod
));
var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
// schema/index.ts
var index_exports = {};
__export(index_exports, {
AISecretTypes: () => AISecretTypes,
APISecretSchema: () => APISecretSchema,
AvailableEndpointTypes: () => AvailableEndpointTypes,
AvailableModels: () => AvailableModels,
AzureEntraSecretSchema: () => AzureEntraSecretSchema,
AzureMetadataSchema: () => AzureMetadataSchema,
BaseMetadataSchema: () => BaseMetadataSchema,
BedrockMetadataSchema: () => BedrockMetadataSchema,
CloudSecretTypes: () => CloudSecretTypes,
DatabricksMetadataSchema: () => DatabricksMetadataSchema,
DatabricksOAuthSecretSchema: () => DatabricksOAuthSecretSchema,
DefaultEndpointTypes: () => DefaultEndpointTypes,
EndpointProviderToBaseURL: () => EndpointProviderToBaseURL,
MessageTypeToMessageType: () => MessageTypeToMessageType,
MistralMetadataSchema: () => MistralMetadataSchema,
ModelEndpointType: () => ModelEndpointType,
ModelFormats: () => ModelFormats,
ModelSchema: () => ModelSchema,
OpenAIMetadataSchema: () => OpenAIMetadataSchema,
PromptInputs: () => PromptInputs,
VertexMetadataSchema: () => VertexMetadataSchema,
anthropicImageBlockSchema: () => anthropicImageBlockSchema,
anthropicSupportedMediaTypes: () => anthropicSupportedMediaTypes,
anthropicTextBlockSchema: () => anthropicTextBlockSchema,
audioBaseMessageSchema: () => audioBaseMessageSchema,
audioContentSchema: () => audioContentSchema,
audioDeltaMessageSchema: () => audioDeltaMessageSchema,
audioDoneMessageSchema: () => audioDoneMessageSchema,
audioFormatTypeSchema: () => audioFormatTypeSchema,
audioInputTranscriptDoneMessageSchema: () => audioInputTranscriptDoneMessageSchema,
audioResponseTranscriptDoneMessageSchema: () => audioResponseTranscriptDoneMessageSchema,
baseMessageSchema: () => baseMessageSchema,
baseResponseSchema: () => baseResponseSchema,
buildClassicChatPrompt: () => buildClassicChatPrompt,
cancelResponseMessageSchema: () => cancelResponseMessageSchema,
clientAudioAppendMessageSchema: () => clientAudioAppendMessageSchema,
clientAudioCommitMessageSchema: () => clientAudioCommitMessageSchema,
conversationItemCreateMessageSchema: () => conversationItemCreateMessageSchema,
credentialsRequestSchema: () => credentialsRequestSchema,
defaultModelParamSettings: () => defaultModelParamSettings,
errorMessageSchema: () => errorMessageSchema,
functionCallBaseMessageSchema: () => functionCallBaseMessageSchema,
functionCallDeltaMessageSchema: () => functionCallDeltaMessageSchema,
functionCallDoneMessageSchema: () => functionCallDoneMessageSchema,
getModelEndpointTypes: () => getModelEndpointTypes,
inputAudioContentSchema: () => inputAudioContentSchema,
inputItemSchema: () => inputItemSchema,
inputTextContentSchema: () => inputTextContentSchema,
messageContentSchema: () => messageContentSchema,
modelParamToModelParam: () => modelParamToModelParam,
modelProviderHasReasoning: () => modelProviderHasReasoning,
modelProviderHasTools: () => modelProviderHasTools,
mp3BitrateSchema: () => mp3BitrateSchema,
openAiRealtimeMessageSchema: () => openAiRealtimeMessageSchema,
outputItemSchema: () => outputItemSchema,
pcmAudioFormatSchema: () => pcmAudioFormatSchema,
proxyLoggingParamSchema: () => proxyLoggingParamSchema,
responseCreatedMessageSchema: () => responseCreatedMessageSchema,
responseDoneMssageSchema: () => responseDoneMssageSchema,
responseOutputItemAddedSchema: () => responseOutputItemAddedSchema,
responseStatusSchema: () => responseStatusSchema,
sessionMessageSchema: () => sessionMessageSchema,
sessionResourceTypeSchema: () => sessionResourceTypeSchema,
sliderSpecs: () => sliderSpecs,
speechEndedMessageSchema: () => speechEndedMessageSchema,
speechStartedMessageSchema: () => speechStartedMessageSchema,
tempCredentialJwtPayloadSchema: () => tempCredentialJwtPayloadSchema,
tempCredentialsCacheValueSchema: () => tempCredentialsCacheValueSchema,
textContentSchema: () => textContentSchema,
toolDefinitionTypeSchema: () => toolDefinitionTypeSchema,
translateParams: () => translateParams,
turnDetectionServerVadTypeSchema: () => turnDetectionServerVadTypeSchema,
unhandledMessageSchema: () => unhandledMessageSchema,
usageTypeSchema: () => usageTypeSchema
});
module.exports = __toCommonJS(index_exports);
var import_zod7 = require("zod");
// schema/models.ts
var import_zod = require("zod");
// schema/model_list.json
var model_list_default = {
"gpt-4o": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 2.5,
output_cost_per_mil_tokens: 10,
input_cache_read_cost_per_mil_tokens: 1.25,
displayName: "GPT-4o",
max_input_tokens: 128e3,
max_output_tokens: 16384
},
"gpt-4o-2024-11-20": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 2.5,
output_cost_per_mil_tokens: 10,
input_cache_read_cost_per_mil_tokens: 1.25,
parent: "gpt-4o",
max_input_tokens: 128e3,
max_output_tokens: 16384
},
"gpt-4o-2024-08-06": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 2.5,
output_cost_per_mil_tokens: 10,
input_cache_read_cost_per_mil_tokens: 1.25,
parent: "gpt-4o",
max_input_tokens: 128e3,
max_output_tokens: 16384
},
"gpt-4o-2024-05-13": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 5,
output_cost_per_mil_tokens: 15,
parent: "gpt-4o",
max_input_tokens: 128e3,
max_output_tokens: 4096
},
"gpt-4.1": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 2,
output_cost_per_mil_tokens: 8,
input_cache_read_cost_per_mil_tokens: 0.5,
displayName: "GPT-4.1",
max_input_tokens: 1047576,
max_output_tokens: 32768
},
"gpt-4.1-2025-04-14": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 2,
output_cost_per_mil_tokens: 8,
input_cache_read_cost_per_mil_tokens: 0.5,
parent: "gpt-4.1",
max_input_tokens: 1047576,
max_output_tokens: 32768
},
"gpt-4o-mini": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 0.15,
output_cost_per_mil_tokens: 0.6,
input_cache_read_cost_per_mil_tokens: 0.075,
displayName: "GPT-4o mini",
max_input_tokens: 128e3,
max_output_tokens: 16384
},
"gpt-4o-mini-2024-07-18": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 0.15,
output_cost_per_mil_tokens: 0.6,
input_cache_read_cost_per_mil_tokens: 0.075,
parent: "gpt-4o-mini",
max_input_tokens: 128e3,
max_output_tokens: 16384
},
"gpt-4.1-mini": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 0.4,
output_cost_per_mil_tokens: 1.6,
input_cache_read_cost_per_mil_tokens: 0.1,
displayName: "GPT-4.1 mini",
max_input_tokens: 1047576,
max_output_tokens: 32768
},
"gpt-4.1-mini-2025-04-14": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 0.4,
output_cost_per_mil_tokens: 1.6,
input_cache_read_cost_per_mil_tokens: 0.1,
parent: "gpt-4.1-mini",
max_input_tokens: 1047576,
max_output_tokens: 32768
},
"gpt-4.1-nano": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 0.1,
output_cost_per_mil_tokens: 0.4,
input_cache_read_cost_per_mil_tokens: 0.025,
displayName: "GPT-4.1 nano",
max_input_tokens: 1047576,
max_output_tokens: 32768
},
"gpt-4.1-nano-2025-04-14": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 0.1,
output_cost_per_mil_tokens: 0.4,
input_cache_read_cost_per_mil_tokens: 0.025,
parent: "gpt-4.1-nano",
max_input_tokens: 1047576,
max_output_tokens: 32768
},
"o4-mini": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 1.1,
output_cost_per_mil_tokens: 4.4,
input_cache_read_cost_per_mil_tokens: 0.275,
reasoning: true,
max_input_tokens: 2e5,
max_output_tokens: 1e5
},
"o4-mini-2025-04-16": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 1.1,
output_cost_per_mil_tokens: 4.4,
input_cache_read_cost_per_mil_tokens: 0.275,
reasoning: true,
parent: "o4-mini",
max_input_tokens: 2e5,
max_output_tokens: 1e5
},
"o3-mini": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 1.1,
output_cost_per_mil_tokens: 4.4,
input_cache_read_cost_per_mil_tokens: 0.55,
reasoning: true,
max_input_tokens: 2e5,
max_output_tokens: 1e5
},
"o3-mini-2025-01-31": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 1.1,
output_cost_per_mil_tokens: 4.4,
input_cache_read_cost_per_mil_tokens: 0.55,
reasoning: true,
parent: "o3-mini",
max_input_tokens: 2e5,
max_output_tokens: 1e5
},
"o3-pro": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 20,
output_cost_per_mil_tokens: 40,
reasoning: true,
displayName: "o3 Pro",
max_input_tokens: 2e5,
max_output_tokens: 1e5
},
"o3-pro-2025-06-10": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 20,
output_cost_per_mil_tokens: 40,
reasoning: true,
parent: "o3-pro",
max_input_tokens: 2e5,
max_output_tokens: 1e5
},
o3: {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 2,
output_cost_per_mil_tokens: 8,
input_cache_read_cost_per_mil_tokens: 0.5,
reasoning: true,
max_input_tokens: 2e5,
max_output_tokens: 1e5
},
"o3-2025-04-16": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 2,
output_cost_per_mil_tokens: 8,
input_cache_read_cost_per_mil_tokens: 0.5,
reasoning: true,
parent: "o3",
max_input_tokens: 2e5,
max_output_tokens: 1e5
},
o1: {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 15,
output_cost_per_mil_tokens: 60,
input_cache_read_cost_per_mil_tokens: 7.5,
reasoning: true,
max_input_tokens: 2e5,
max_output_tokens: 1e5
},
"o1-2024-12-17": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 15,
output_cost_per_mil_tokens: 60,
input_cache_read_cost_per_mil_tokens: 7.5,
reasoning: true,
parent: "o1",
max_input_tokens: 2e5,
max_output_tokens: 1e5
},
"o1-mini": {
format: "openai",
flavor: "chat",
multimodal: false,
input_cost_per_mil_tokens: 1.1,
output_cost_per_mil_tokens: 4.4,
input_cache_read_cost_per_mil_tokens: 0.55,
reasoning: true,
max_input_tokens: 128e3,
max_output_tokens: 65536
},
"o1-mini-2024-09-12": {
format: "openai",
flavor: "chat",
multimodal: false,
input_cost_per_mil_tokens: 3,
output_cost_per_mil_tokens: 12,
input_cache_read_cost_per_mil_tokens: 1.5,
reasoning: true,
parent: "o1-mini",
max_input_tokens: 128e3,
max_output_tokens: 65536
},
"o1-pro": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 150,
output_cost_per_mil_tokens: 600,
reasoning: true,
max_input_tokens: 2e5,
max_output_tokens: 1e5
},
"o1-pro-2025-03-19": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 150,
output_cost_per_mil_tokens: 600,
reasoning: true,
parent: "o1-pro",
max_input_tokens: 2e5,
max_output_tokens: 1e5
},
"chatgpt-4o-latest": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 5,
output_cost_per_mil_tokens: 15,
displayName: "ChatGPT-4o",
max_input_tokens: 128e3,
max_output_tokens: 4096
},
"gpt-4-turbo": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 10,
output_cost_per_mil_tokens: 30,
displayName: "GPT-4 Turbo",
max_input_tokens: 128e3,
max_output_tokens: 4096
},
"gpt-4-turbo-2024-04-09": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 10,
output_cost_per_mil_tokens: 30,
parent: "gpt-4-turbo",
max_input_tokens: 128e3,
max_output_tokens: 4096
},
"gpt-4-turbo-preview": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 10,
output_cost_per_mil_tokens: 30,
parent: "gpt-4-turbo",
max_input_tokens: 128e3,
max_output_tokens: 4096
},
"gpt-4": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 30,
output_cost_per_mil_tokens: 60,
displayName: "GPT-4",
max_input_tokens: 8192,
max_output_tokens: 4096
},
"gpt-4-0125-preview": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 10,
output_cost_per_mil_tokens: 30,
experimental: true,
parent: "gpt-4",
max_input_tokens: 128e3,
max_output_tokens: 4096
},
"gpt-4-1106-preview": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 10,
output_cost_per_mil_tokens: 30,
experimental: true,
parent: "gpt-4",
max_input_tokens: 128e3,
max_output_tokens: 4096
},
"gpt-4-0613": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 30,
output_cost_per_mil_tokens: 60,
parent: "gpt-4",
max_input_tokens: 8192,
max_output_tokens: 4096
},
"gpt-4-0314": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 30,
output_cost_per_mil_tokens: 60,
parent: "gpt-4",
max_input_tokens: 8192,
max_output_tokens: 4096
},
"gpt-4.5-preview": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 75,
output_cost_per_mil_tokens: 150,
input_cache_read_cost_per_mil_tokens: 37.5,
displayName: "GPT-4.5",
experimental: true,
max_input_tokens: 128e3,
max_output_tokens: 16384
},
"gpt-4.5-preview-2025-02-27": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 75,
output_cost_per_mil_tokens: 150,
input_cache_read_cost_per_mil_tokens: 37.5,
experimental: true,
parent: "gpt-4.5-preview",
max_input_tokens: 128e3,
max_output_tokens: 16384
},
"o1-preview": {
format: "openai",
flavor: "chat",
multimodal: false,
input_cost_per_mil_tokens: 15,
output_cost_per_mil_tokens: 60,
input_cache_read_cost_per_mil_tokens: 7.5,
reasoning: true,
experimental: true,
parent: "o1",
max_input_tokens: 128e3,
max_output_tokens: 32768
},
"o1-preview-2024-09-12": {
format: "openai",
flavor: "chat",
multimodal: false,
input_cost_per_mil_tokens: 15,
output_cost_per_mil_tokens: 60,
input_cache_read_cost_per_mil_tokens: 7.5,
reasoning: true,
experimental: true,
parent: "o1",
max_input_tokens: 128e3,
max_output_tokens: 32768
},
"gpt-4o-search-preview": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 2.5,
output_cost_per_mil_tokens: 10,
input_cache_read_cost_per_mil_tokens: 1.25,
displayName: "GPT-4o Search Preview",
experimental: true,
max_input_tokens: 128e3,
max_output_tokens: 16384
},
"gpt-4o-search-preview-2025-03-11": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 2.5,
output_cost_per_mil_tokens: 10,
input_cache_read_cost_per_mil_tokens: 1.25,
experimental: true,
parent: "gpt-4o-search-preview",
max_input_tokens: 128e3,
max_output_tokens: 16384
},
"gpt-4o-mini-search-preview": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 0.15,
output_cost_per_mil_tokens: 0.6,
input_cache_read_cost_per_mil_tokens: 0.075,
displayName: "GPT-4o mini Search Preview",
experimental: true,
max_input_tokens: 128e3,
max_output_tokens: 16384
},
"gpt-4o-mini-search-preview-2025-03-11": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 0.15,
output_cost_per_mil_tokens: 0.6,
input_cache_read_cost_per_mil_tokens: 0.075,
experimental: true,
parent: "gpt-4o-mini-search-preview",
max_input_tokens: 128e3,
max_output_tokens: 16384
},
"gpt-3.5-turbo-0125": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.5,
output_cost_per_mil_tokens: 1.5,
displayName: "GPT 3.5T 0125",
deprecated: true,
max_input_tokens: 16385,
max_output_tokens: 4096
},
"gpt-3.5-turbo": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 1.5,
output_cost_per_mil_tokens: 2,
displayName: "GPT 3.5T",
deprecated: true,
max_input_tokens: 16385,
max_output_tokens: 4096
},
"gpt-35-turbo": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.5,
output_cost_per_mil_tokens: 1.5,
displayName: "GPT 3.5T",
deprecated: true
},
"gpt-3.5-turbo-1106": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 1,
output_cost_per_mil_tokens: 2,
displayName: "GPT 3.5T 1106",
deprecated: true,
max_input_tokens: 16385,
max_output_tokens: 4096
},
"gpt-3.5-turbo-instruct": {
format: "openai",
flavor: "completion",
input_cost_per_mil_tokens: 1.5,
output_cost_per_mil_tokens: 2,
displayName: "GPT 3.5T Instruct",
deprecated: true,
max_input_tokens: 8192,
max_output_tokens: 4096
},
"gpt-3.5-turbo-instruct-0914": {
format: "openai",
flavor: "completion",
input_cost_per_mil_tokens: 1.5,
output_cost_per_mil_tokens: 2,
displayName: "GPT 3.5T Instruct 0914",
deprecated: true,
max_input_tokens: 8192,
max_output_tokens: 4097
},
"gpt-4-32k": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 60,
output_cost_per_mil_tokens: 120,
displayName: "GPT 4 32k",
deprecated: true,
max_input_tokens: 32768,
max_output_tokens: 4096
},
"gpt-4-32k-0613": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 60,
output_cost_per_mil_tokens: 120,
displayName: "GPT 4 32k 0613",
deprecated: true,
max_input_tokens: 32768,
max_output_tokens: 4096
},
"gpt-4-32k-0314": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 60,
output_cost_per_mil_tokens: 120,
displayName: "GPT 4 32k 0314",
deprecated: true,
max_input_tokens: 32768,
max_output_tokens: 4096
},
"gpt-4-vision-preview": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 10,
output_cost_per_mil_tokens: 30,
displayName: "GPT 4 Vision-Preview",
deprecated: true,
max_input_tokens: 128e3,
max_output_tokens: 4096
},
"gpt-4-1106-vision-preview": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 10,
output_cost_per_mil_tokens: 30,
displayName: "GPT 4 1106 Vision-Preview",
deprecated: true,
max_input_tokens: 128e3,
max_output_tokens: 4096
},
"gpt-3.5-turbo-16k": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 3,
output_cost_per_mil_tokens: 4,
displayName: "GPT 3.5T 16k",
deprecated: true,
max_input_tokens: 16385,
max_output_tokens: 4096
},
"gpt-35-turbo-16k": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 3,
output_cost_per_mil_tokens: 4,
displayName: "GPT 3.5T 16k",
deprecated: true
},
"gpt-3.5-turbo-16k-0613": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 3,
output_cost_per_mil_tokens: 4,
displayName: "GPT 3.5T 16k 0613",
deprecated: true,
max_input_tokens: 16385,
max_output_tokens: 4096
},
"gpt-3.5-turbo-0613": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 1.5,
output_cost_per_mil_tokens: 2,
displayName: "GPT 3.5T 0613",
deprecated: true,
max_input_tokens: 4097,
max_output_tokens: 4096
},
"gpt-3.5-turbo-0301": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 1.5,
output_cost_per_mil_tokens: 2,
displayName: "GPT 3.5T 0301",
deprecated: true,
max_input_tokens: 4097,
max_output_tokens: 4096
},
"text-davinci-003": {
format: "openai",
flavor: "completion",
input_cost_per_mil_tokens: 2,
output_cost_per_mil_tokens: 2,
displayName: "Text Davinci 003",
deprecated: true
},
"claude-sonnet-4-20250514": {
format: "anthropic",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 3,
output_cost_per_mil_tokens: 15,
input_cache_read_cost_per_mil_tokens: 0.3,
input_cache_write_cost_per_mil_tokens: 3.75,
displayName: "Claude 4 Sonnet",
reasoning: true,
reasoning_budget: true,
max_input_tokens: 2e5,
max_output_tokens: 64e3
},
"claude-4-sonnet-20250514": {
format: "anthropic",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 3,
output_cost_per_mil_tokens: 15,
input_cache_read_cost_per_mil_tokens: 0.3,
input_cache_write_cost_per_mil_tokens: 3.75,
displayName: "Claude 4 Sonnet (old naming format)",
deprecated: true,
max_input_tokens: 2e5,
max_output_tokens: 64e3
},
"claude-3-7-sonnet-latest": {
format: "anthropic",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 3,
output_cost_per_mil_tokens: 15,
input_cache_read_cost_per_mil_tokens: 0.3,
input_cache_write_cost_per_mil_tokens: 3.75,
displayName: "Claude 3.7 Sonnet",
reasoning: true,
reasoning_budget: true,
max_input_tokens: 2e5,
max_output_tokens: 128e3
},
"claude-3-7-sonnet-20250219": {
format: "anthropic",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 3,
output_cost_per_mil_tokens: 15,
input_cache_read_cost_per_mil_tokens: 0.3,
input_cache_write_cost_per_mil_tokens: 3.75,
reasoning: true,
reasoning_budget: true,
parent: "claude-3-7-sonnet-latest",
max_input_tokens: 2e5,
max_output_tokens: 128e3
},
"claude-3-5-haiku-latest": {
format: "anthropic",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 1,
output_cost_per_mil_tokens: 5,
input_cache_read_cost_per_mil_tokens: 0.1,
input_cache_write_cost_per_mil_tokens: 1.25,
displayName: "Claude 3.5 Haiku",
max_input_tokens: 2e5,
max_output_tokens: 8192
},
"claude-3-5-haiku-20241022": {
format: "anthropic",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 0.8,
output_cost_per_mil_tokens: 4,
input_cache_read_cost_per_mil_tokens: 0.08,
input_cache_write_cost_per_mil_tokens: 1,
parent: "claude-3-5-haiku-latest",
max_input_tokens: 2e5,
max_output_tokens: 8192
},
"claude-3-5-sonnet-latest": {
format: "anthropic",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 3,
output_cost_per_mil_tokens: 15,
input_cache_read_cost_per_mil_tokens: 0.3,
input_cache_write_cost_per_mil_tokens: 3.75,
displayName: "Claude 3.5 Sonnet",
max_input_tokens: 2e5,
max_output_tokens: 8192
},
"claude-3-5-sonnet-20241022": {
format: "anthropic",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 3,
output_cost_per_mil_tokens: 15,
input_cache_read_cost_per_mil_tokens: 0.3,
input_cache_write_cost_per_mil_tokens: 3.75,
parent: "claude-3-5-sonnet-latest",
max_input_tokens: 2e5,
max_output_tokens: 8192
},
"claude-3-5-sonnet-20240620": {
format: "anthropic",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 3,
output_cost_per_mil_tokens: 15,
input_cache_read_cost_per_mil_tokens: 0.3,
input_cache_write_cost_per_mil_tokens: 3.75,
parent: "claude-3-5-sonnet-latest",
max_input_tokens: 2e5,
max_output_tokens: 8192
},
"claude-opus-4-20250514": {
format: "anthropic",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 15,
output_cost_per_mil_tokens: 75,
input_cache_read_cost_per_mil_tokens: 1.5,
input_cache_write_cost_per_mil_tokens: 18.75,
displayName: "Claude 4 Opus",
reasoning: true,
reasoning_budget: true,
max_input_tokens: 2e5,
max_output_tokens: 32e3
},
"claude-4-opus-20250514": {
format: "anthropic",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 15,
output_cost_per_mil_tokens: 75,
input_cache_read_cost_per_mil_tokens: 1.5,
input_cache_write_cost_per_mil_tokens: 18.75,
displayName: "Claude 4 Opus (old naming format)",
deprecated: true,
max_input_tokens: 2e5,
max_output_tokens: 32e3
},
"claude-3-opus-latest": {
format: "anthropic",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 15,
output_cost_per_mil_tokens: 75,
input_cache_read_cost_per_mil_tokens: 1.5,
input_cache_write_cost_per_mil_tokens: 18.75,
displayName: "Claude 3 Opus",
max_input_tokens: 2e5,
max_output_tokens: 4096
},
"claude-3-opus-20240229": {
format: "anthropic",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 15,
output_cost_per_mil_tokens: 75,
input_cache_read_cost_per_mil_tokens: 1.5,
input_cache_write_cost_per_mil_tokens: 18.75,
parent: "claude-3-opus-latest",
max_input_tokens: 2e5,
max_output_tokens: 4096
},
"claude-3-sonnet-20240229": {
format: "anthropic",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 3,
output_cost_per_mil_tokens: 15,
displayName: "Claude 3 Sonnet",
max_input_tokens: 2e5,
max_output_tokens: 4096
},
"claude-3-haiku-20240307": {
format: "anthropic",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 0.25,
output_cost_per_mil_tokens: 1.25,
input_cache_read_cost_per_mil_tokens: 0.03,
input_cache_write_cost_per_mil_tokens: 0.3,
displayName: "Claude 3 Haiku",
max_input_tokens: 2e5,
max_output_tokens: 4096
},
"claude-instant-1.2": {
format: "anthropic",
flavor: "chat",
input_cost_per_mil_tokens: 0.163,
output_cost_per_mil_tokens: 0.551,
displayName: "Claude Instant 1.2",
deprecated: true,
max_input_tokens: 1e5,
max_output_tokens: 8191
},
"claude-instant-1": {
format: "anthropic",
flavor: "chat",
input_cost_per_mil_tokens: 1.63,
output_cost_per_mil_tokens: 5.51,
displayName: "Claude Instant 1",
deprecated: true,
max_input_tokens: 1e5,
max_output_tokens: 8191
},
"claude-2.1": {
format: "anthropic",
flavor: "chat",
input_cost_per_mil_tokens: 8,
output_cost_per_mil_tokens: 24,
displayName: "Claude 2.1",
deprecated: true,
max_input_tokens: 2e5,
max_output_tokens: 8191
},
"claude-2.0": {
format: "anthropic",
flavor: "chat",
input_cost_per_mil_tokens: 8,
output_cost_per_mil_tokens: 24,
displayName: "Claude 2.0",
deprecated: true
},
"claude-2": {
format: "anthropic",
flavor: "chat",
input_cost_per_mil_tokens: 8,
output_cost_per_mil_tokens: 24,
displayName: "Claude 2",
deprecated: true,
max_input_tokens: 1e5,
max_output_tokens: 8191
},
"meta/llama-2-70b-chat": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.65,
output_cost_per_mil_tokens: 2.75,
displayName: "LLaMA 2 70b Chat"
},
mistral: {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0,
output_cost_per_mil_tokens: 0
},
phi: {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0,
output_cost_per_mil_tokens: 0,
deprecated: true
},
sonar: {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 1,
output_cost_per_mil_tokens: 1,
displayName: "Sonar"
},
"sonar-pro": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 3,
output_cost_per_mil_tokens: 15,
displayName: "Sonar Pro"
},
"sonar-reasoning": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 1,
output_cost_per_mil_tokens: 5,
displayName: "Sonar Reasoning"
},
"sonar-reasoning-pro": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 2,
output_cost_per_mil_tokens: 8,
displayName: "Sonar Reasoning Pro"
},
"r1-1776": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 2,
output_cost_per_mil_tokens: 8,
displayName: "R1 1776"
},
"meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8": {
format: "openai",
flavor: "chat",
multimodal: true,
displayName: "Llama 4 Maverick Instruct (17Bx128E)"
},
"meta-llama/Llama-4-Scout-17B-16E-Instruct": {
format: "openai",
flavor: "chat",
displayName: "Llama 4 Scout Instruct (17Bx16E)"
},
"meta-llama/Llama-3.3-70B-Instruct-Turbo": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.88,
output_cost_per_mil_tokens: 0.88,
displayName: "Llama 3.3 70B Instruct Turbo"
},
"meta-llama/Llama-3.3-70B-Instruct-Turbo-Free": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0,
output_cost_per_mil_tokens: 0,
displayName: "Llama 3.3 70B Instruct Turbo Free"
},
"meta-llama/Llama-3.2-90B-Vision-Instruct-Turbo": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 1.2,
output_cost_per_mil_tokens: 1.2,
displayName: "Llama 3.2 90B Vision Instruct Turbo"
},
"meta-llama/Llama-3.2-11B-Vision-Instruct-Turbo": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 0.18,
output_cost_per_mil_tokens: 0.18,
displayName: "Llama 3.2 11B Vision Instruct Turbo"
},
"meta-llama/Llama-Vision-Free": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 0,
output_cost_per_mil_tokens: 0,
displayName: "Llama Vision Free"
},
"meta-llama/Llama-3.2-3B-Instruct-Turbo": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.06,
output_cost_per_mil_tokens: 0.06,
displayName: "Llama 3.2 3B Instruct Turbo"
},
"meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 3.5,
output_cost_per_mil_tokens: 3.5,
displayName: "Llama 3.1 405B Instruct Turbo"
},
"meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.88,
output_cost_per_mil_tokens: 0.88,
displayName: "Llama 3.1 70B Instruct Turbo"
},
"meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.18,
output_cost_per_mil_tokens: 0.18,
displayName: "Llama 3.1 8B Instruct Turbo"
},
"meta-llama/Llama-3-70b-chat-hf": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.9,
output_cost_per_mil_tokens: 0.9,
displayName: "Llama 3 70B Instruct Reference"
},
"meta-llama/Meta-Llama-3-70B-Instruct-Turbo": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.88,
output_cost_per_mil_tokens: 0.88,
displayName: "Llama 3 70B Instruct Turbo"
},
"meta-llama/Meta-Llama-3-70B-Instruct-Lite": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.54,
output_cost_per_mil_tokens: 0.54,
displayName: "Llama 3 70B Instruct Lite"
},
"meta-llama/Llama-3-8b-chat-hf": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.2,
output_cost_per_mil_tokens: 0.2,
displayName: "Llama 3 8B Instruct Reference"
},
"meta-llama/Meta-Llama-3-8B-Instruct-Turbo": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.18,
output_cost_per_mil_tokens: 0.18,
displayName: "Llama 3 8B Instruct Turbo"
},
"meta-llama/Meta-Llama-3-8B-Instruct-Lite": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.1,
output_cost_per_mil_tokens: 0.1,
displayName: "Llama 3 8B Instruct Lite"
},
"google/gemma-2-27b-it": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.35,
output_cost_per_mil_tokens: 1.05,
displayName: "Gemma-2 Instruct (27B)",
max_output_tokens: 8192
},
"google/gemma-2-9b-it": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.35,
output_cost_per_mil_tokens: 1.05,
displayName: "Gemma-2 Instruct (9B)",
max_output_tokens: 8192
},
"google/gemma-2b-it": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.1,
output_cost_per_mil_tokens: 0.1,
displayName: "Gemma Instruct (2B)"
},
"mistralai/Mistral-Small-24B-Instruct-2501": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.8,
output_cost_per_mil_tokens: 0.8,
displayName: "Mistral Small (24B) Instruct 25.01"
},
"mistralai/Mistral-7B-Instruct-v0.3": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.2,
output_cost_per_mil_tokens: 0.2,
displayName: "Mistral (7B) Instruct v0.3"
},
"mistralai/Mistral-7B-Instruct-v0.2": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.2,
output_cost_per_mil_tokens: 0.2,
displayName: "Mistral (7B) Instruct v0.2"
},
"mistralai/Mistral-7B-Instruct-v0.1": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.2,
output_cost_per_mil_tokens: 0.2,
displayName: "Mistral (7B) Instruct"
},
"mistralai/Mixtral-8x22B-Instruct-v0.1": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 1.2,
output_cost_per_mil_tokens: 1.2,
displayName: "Mixtral 8x22B Instruct v0.1"
},
"mistralai/Mixtral-8x7B-Instruct-v0.1": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.6,
output_cost_per_mil_tokens: 0.6,
displayName: "Mixtral 8x7B Instruct v0.1"
},
"deepseek-ai/DeepSeek-V3": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 1.25,
output_cost_per_mil_tokens: 1.25,
displayName: "DeepSeek V3"
},
"deepseek-ai/DeepSeek-R1": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 7,
output_cost_per_mil_tokens: 7,
displayName: "DeepSeek R1"
},
"deepseek-ai/DeepSeek-R1-Distill-Llama-70B": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 2,
output_cost_per_mil_tokens: 2,
displayName: "DeepSeek R1 Distill Llama 70B"
},
"deepseek-ai/DeepSeek-R1-Distill-Llama-70B-Free": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0,
output_cost_per_mil_tokens: 0,
displayName: "DeepSeek R1 Distill Llama 70B Free"
},
"deepseek-ai/DeepSeek-R1-Distill-Qwen-14B": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 1.6,
output_cost_per_mil_tokens: 1.6,
displayName: "DeepSeek R1 Distill Qwen 14B"
},
"deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.18,
output_cost_per_mil_tokens: 0.18,
displayName: "DeepSeek R1 Distill Qwen 1.5B"
},
"deepseek-ai/deepseek-llm-67b-chat": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.9,
output_cost_per_mil_tokens: 0.9,
displayName: "DeepSeek LLM Chat (67B)"
},
"Qwen/Qwen2.5-72B-Instruct-Turbo": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 1.2,
output_cost_per_mil_tokens: 1.2,
displayName: "Qwen 2.5 72B Instruct Turbo"
},
"Qwen/Qwen2.5-7B-Instruct-Turbo": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.3,
output_cost_per_mil_tokens: 0.3,
displayName: "Qwen 2.5 7B Instruct Turbo"
},
"Qwen/Qwen2.5-Coder-32B-Instruct": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.8,
output_cost_per_mil_tokens: 0.8,
displayName: "Qwen 2.5 Coder 32B Instruct"
},
"Qwen/QwQ-32B": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.8,
output_cost_per_mil_tokens: 0.8,
displayName: "Qwen QwQ 32B"
},
"Qwen/Qwen2-VL-72B-Instruct": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 1.2,
output_cost_per_mil_tokens: 1.2,
displayName: "Qwen-2VL (72B) Instruct"
},
"Qwen/Qwen2-72B-Instruct": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.9,
output_cost_per_mil_tokens: 0.9,
displayName: "Qwen 2 Instruct (72B)"
},
"nvidia/Llama-3.1-Nemotron-70B-Instruct-HF": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.88,
output_cost_per_mil_tokens: 0.88,
displayName: "Llama 3.1 Nemotron 70B Instruct HF"
},
"microsoft/WizardLM-2-8x22B": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 1.2,
output_cost_per_mil_tokens: 1.2,
displayName: "WizardLM-2 (8x22B)"
},
"databricks/dbrx-instruct": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 1.2,
output_cost_per_mil_tokens: 1.2,
displayName: "DBRX Instruct"
},
"NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.6,
output_cost_per_mil_tokens: 0.6,
displayName: "Nous Hermes 2 - Mixtral 8x7B-DPO"
},
"Gryphe/MythoMax-L2-13b": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.3,
output_cost_per_mil_tokens: 0.3,
displayName: "MythoMax-L2 (13B)"
},
"Gryphe/MythoMax-L2-13b-Lite": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.1,
output_cost_per_mil_tokens: 0.1,
displayName: "Gryphe MythoMax L2 Lite (13B)"
},
"meta-llama/Meta-Llama-3-70B": {
format: "openai",
flavor: "completion",
input_cost_per_mil_tokens: 0.9,
output_cost_per_mil_tokens: 0.9,
displayName: "Llama 3 70b",
deprecated: true
},
"meta-llama/Llama-3-8b-hf": {
format: "openai",
flavor: "completion",
input_cost_per_mil_tokens: 0.2,
output_cost_per_mil_tokens: 0.2,
displayName: "Llama 3 8b HF",
deprecated: true
},
"meta-llama/Llama-2-70b-chat-hf": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.9,
output_cost_per_mil_tokens: 0.9,
displayName: "Llama 2 70b Chat HF",
deprecated: true
},
"deepseek-ai/deepseek-coder-33b-instruct": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.8,
output_cost_per_mil_tokens: 0.8,
displayName: "Deepseek Coder 33b Instruct",
deprecated: true
},
"Qwen/QwQ-32B-Preview": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.8,
output_cost_per_mil_tokens: 0.8,
displayName: "Qwen QwQ 32B Preview",
deprecated: true
},
"NousResearch/Nous-Hermes-2-Yi-34B": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.8,
output_cost_per_mil_tokens: 0.8,
displayName: "Nous Hermes 2 Yi 34B",
deprecated: true
},
"magistral-medium-latest": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 2,
output_cost_per_mil_tokens: 5,
displayName: "Magistral Medium Latest",
max_input_tokens: 40960,
max_output_tokens: 4e4
},
"magistral-medium-2506": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 2,
output_cost_per_mil_tokens: 5,
parent: "magistral-medium-latest",
max_input_tokens: 40960,
max_output_tokens: 4e4
},
"magistral-small-latest": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.5,
output_cost_per_mil_tokens: 1.5,
displayName: "Magistral Small Latest",
max_input_tokens: 4e4,
max_output_tokens: 4e4
},
"magistral-small-2506": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.5,
output_cost_per_mil_tokens: 1.5,
parent: "magistral-small-latest",
max_input_tokens: 4e4,
max_output_tokens: 4e4
},
"mistralai/mixtral-8x7b-32kseqlen": {
format: "openai",
flavor: "completion",
input_cost_per_mil_tokens: 0.06,
output_cost_per_mil_tokens: 0.06,
displayName: "Mixtral 8x7B 32k",
deprecated: true
},
"mistralai/Mixtral-8x7B-Instruct-v0.1-json": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.6,
output_cost_per_mil_tokens: 0.6,
displayName: "Mixtral 8x7B Instruct v0.1 JSON",
deprecated: true
},
"mistralai/Mixtral-8x22B": {
format: "openai",
flavor: "completion",
input_cost_per_mil_tokens: 1.08,
output_cost_per_mil_tokens: 1.08,
displayName: "Mixtral 8x22B",
deprecated: true
},
"mistral-large-latest": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 2,
output_cost_per_mil_tokens: 6,
displayName: "Mistral Large"
},
"mistral-large-2411": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 2,
output_cost_per_mil_tokens: 6,
parent: "mistral-large-latest"
},
"pixtral-large-latest": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 2,
output_cost_per_mil_tokens: 6,
displayName: "Pixtral Large"
},
"pixtral-large-2411": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 2,
output_cost_per_mil_tokens: 6,
parent: "pixtral-large-latest"
},
"mistral-medium-latest": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 0.4,
output_cost_per_mil_tokens: 2,
displayName: "Mistral Medium 3"
},
"mistral-medium-2505": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 0.4,
output_cost_per_mil_tokens: 2,
parent: "mistral-medium-latest"
},
"mistral-small-latest": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.1,
output_cost_per_mil_tokens: 0.3,
displayName: "Mistral Small"
},
"mistral-small-2501": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.1,
output_cost_per_mil_tokens: 0.3,
parent: "mistral-small-latest"
},
"codestral-latest": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.3,
output_cost_per_mil_tokens: 0.9,
displayName: "Codestral"
},
"codestral-2501": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.3,
output_cost_per_mil_tokens: 0.9,
parent: "codestral-latest"
},
"ministral-8b-latest": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.1,
output_cost_per_mil_tokens: 0.1,
displayName: "Ministral 8B"
},
"ministral-8b-2410": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.1,
output_cost_per_mil_tokens: 0.1,
parent: "ministral-8b-latest"
},
"ministral-3b-latest": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.04,
output_cost_per_mil_tokens: 0.04,
displayName: "Ministral 3B"
},
"ministral-3b-2410": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.04,
output_cost_per_mil_tokens: 0.04,
parent: "ministral-3b-latest"
},
"mistral-saba-latest": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.2,
output_cost_per_mil_tokens: 0.6,
displayName: "Mistral Saba"
},
"mistral-saba-2502": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.2,
output_cost_per_mil_tokens: 0.6,
parent: "mistral-saba-latest"
},
"pixtral-12b-2409": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 0.15,
output_cost_per_mil_tokens: 0.15,
displayName: "Pixtral 12B"
},
"open-mistral-nemo": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.15,
output_cost_per_mil_tokens: 0.15,
displayName: "Mistral NeMo"
},
"open-mistral-nemo-2407": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.15,
output_cost_per_mil_tokens: 0.15
},
"open-codestral-mamba": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.6,
output_cost_per_mil_tokens: 0.6,
displayName: "Codestral Mamba"
},
"open-mixtral-8x22b": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 2,
output_cost_per_mil_tokens: 6,
displayName: "Mixtral 8x22B",
deprecated: true
},
"mistral-tiny": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.15,
output_cost_per_mil_tokens: 0.46,
displayName: "Mistral Tiny",
deprecated: true
},
"mistral-small": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 1,
output_cost_per_mil_tokens: 3,
displayName: "Mistral Small",
deprecated: true
},
"mistral-medium": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 2.75,
output_cost_per_mil_tokens: 8.1,
displayName: "Mistral Medium",
deprecated: true
},
"llama-3.3-70b-versatile": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.59,
output_cost_per_mil_tokens: 0.79,
displayName: "Llama 3.3 70B Versatile 128k"
},
"llama-3.1-8b-instant": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.05,
output_cost_per_mil_tokens: 0.08,
displayName: "Llama 3.1 8B Instant 128k"
},
"llama3-70b-8192": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.64,
output_cost_per_mil_tokens: 0.8,
displayName: "Llama 3 70B 8k"
},
"llama3-8b-8192": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.1,
output_cost_per_mil_tokens: 0.1,
displayName: "Llama 3 8B 8k"
},
"llama-guard-3-8b": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.2,
output_cost_per_mil_tokens: 0.2,
displayName: "Llama Guard 3 8B 8k"
},
"gemma2-9b-it": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.2,
output_cost_per_mil_tokens: 0.2,
displayName: "Gemma 2 9B"
},
"meta-llama/llama-4-maverick-17b-128e-instruct": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 0.5,
output_cost_per_mil_tokens: 0.77,
displayName: "Llama 4 Maverick (17Bx128E)",
experimental: true
},
"meta-llama/llama-4-scout-17b-16e-instruct": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.11,
output_cost_per_mil_tokens: 0.34,
displayName: "Llama 4 Scout (17Bx16E)",
experimental: true
},
"llama-3.3-70b-specdec": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.59,
output_cost_per_mil_tokens: 0.99,
displayName: "Llama 3.3 70B SpecDec 8k",
experimental: true
},
"llama-3.2-90b-vision-preview": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 0.9,
output_cost_per_mil_tokens: 0.9,
displayName: "Llama 3.2 90B Vision 8k (Preview)",
experimental: true
},
"llama-3.2-11b-vision-preview": {
format: "openai",
flavor: "chat",
multimodal: true,
input_cost_per_mil_tokens: 0.18,
output_cost_per_mil_tokens: 0.18,
displayName: "Llama 3.2 11B Vision 8k (Preview)",
experimental: true
},
"llama-3.2-3b-preview": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.06,
output_cost_per_mil_tokens: 0.06,
displayName: "Llama 3.2 3B (Preview) 8k",
experimental: true
},
"llama-3.2-1b-preview": {
format: "openai",
flavor: "chat",
input_cost_per_mil_tokens: 0.04,
out