@sentry/core
Version:
Base implementation for all Sentry JavaScript SDKs
335 lines (305 loc) • 10.8 kB
JavaScript
import { OPENAI_OPERATIONS, GEN_AI_REQUEST_MODEL_ATTRIBUTE, GEN_AI_REQUEST_TEMPERATURE_ATTRIBUTE, GEN_AI_REQUEST_TOP_P_ATTRIBUTE, GEN_AI_REQUEST_FREQUENCY_PENALTY_ATTRIBUTE, GEN_AI_REQUEST_PRESENCE_PENALTY_ATTRIBUTE, GEN_AI_REQUEST_STREAM_ATTRIBUTE, GEN_AI_REQUEST_ENCODING_FORMAT_ATTRIBUTE, GEN_AI_REQUEST_DIMENSIONS_ATTRIBUTE, GEN_AI_CONVERSATION_ID_ATTRIBUTE, GEN_AI_RESPONSE_ID_ATTRIBUTE, OPENAI_RESPONSE_ID_ATTRIBUTE, GEN_AI_RESPONSE_MODEL_ATTRIBUTE, OPENAI_RESPONSE_MODEL_ATTRIBUTE, OPENAI_RESPONSE_TIMESTAMP_ATTRIBUTE, GEN_AI_RESPONSE_FINISH_REASONS_ATTRIBUTE, GEN_AI_RESPONSE_TOOL_CALLS_ATTRIBUTE, GEN_AI_USAGE_INPUT_TOKENS_ATTRIBUTE, OPENAI_USAGE_PROMPT_TOKENS_ATTRIBUTE, GEN_AI_USAGE_OUTPUT_TOKENS_ATTRIBUTE, OPENAI_USAGE_COMPLETION_TOKENS_ATTRIBUTE, GEN_AI_USAGE_TOTAL_TOKENS_ATTRIBUTE } from '../ai/gen-ai-attributes.js';
import { INSTRUMENTED_METHODS } from './constants.js';
/**
* Maps OpenAI method paths to Sentry operation names
*/
function getOperationName(methodPath) {
if (methodPath.includes('chat.completions')) {
return OPENAI_OPERATIONS.CHAT;
}
if (methodPath.includes('responses')) {
return OPENAI_OPERATIONS.RESPONSES;
}
if (methodPath.includes('embeddings')) {
return OPENAI_OPERATIONS.EMBEDDINGS;
}
if (methodPath.includes('conversations')) {
return OPENAI_OPERATIONS.CONVERSATIONS;
}
return methodPath.split('.').pop() || 'unknown';
}
/**
* Get the span operation for OpenAI methods
* Following Sentry's convention: "gen_ai.{operation_name}"
*/
function getSpanOperation(methodPath) {
return `gen_ai.${getOperationName(methodPath)}`;
}
/**
* Check if a method path should be instrumented
*/
function shouldInstrument(methodPath) {
return INSTRUMENTED_METHODS.includes(methodPath );
}
/**
* Build method path from current traversal
*/
function buildMethodPath(currentPath, prop) {
return currentPath ? `${currentPath}.${prop}` : prop;
}
/**
* Check if response is a Chat Completion object
*/
function isChatCompletionResponse(response) {
return (
response !== null &&
typeof response === 'object' &&
'object' in response &&
(response ).object === 'chat.completion'
);
}
/**
* Check if response is a Responses API object
*/
function isResponsesApiResponse(response) {
return (
response !== null &&
typeof response === 'object' &&
'object' in response &&
(response ).object === 'response'
);
}
/**
* Check if response is an Embeddings API object
*/
function isEmbeddingsResponse(response) {
if (response === null || typeof response !== 'object' || !('object' in response)) {
return false;
}
const responseObject = response ;
return (
responseObject.object === 'list' &&
typeof responseObject.model === 'string' &&
responseObject.model.toLowerCase().includes('embedding')
);
}
/**
* Check if response is a Conversations API object
* @see https://platform.openai.com/docs/api-reference/conversations
*/
function isConversationResponse(response) {
return (
response !== null &&
typeof response === 'object' &&
'object' in response &&
(response ).object === 'conversation'
);
}
/**
* Check if streaming event is from the Responses API
*/
function isResponsesApiStreamEvent(event) {
return (
event !== null &&
typeof event === 'object' &&
'type' in event &&
typeof (event ).type === 'string' &&
((event ).type ).startsWith('response.')
);
}
/**
* Check if streaming event is a chat completion chunk
*/
function isChatCompletionChunk(event) {
return (
event !== null &&
typeof event === 'object' &&
'object' in event &&
(event ).object === 'chat.completion.chunk'
);
}
/**
* Add attributes for Chat Completion responses
*/
function addChatCompletionAttributes(
span,
response,
recordOutputs,
) {
setCommonResponseAttributes(span, response.id, response.model, response.created);
if (response.usage) {
setTokenUsageAttributes(
span,
response.usage.prompt_tokens,
response.usage.completion_tokens,
response.usage.total_tokens,
);
}
if (Array.isArray(response.choices)) {
const finishReasons = response.choices
.map(choice => choice.finish_reason)
.filter((reason) => reason !== null);
if (finishReasons.length > 0) {
span.setAttributes({
[]: JSON.stringify(finishReasons),
});
}
// Extract tool calls from all choices (only if recordOutputs is true)
if (recordOutputs) {
const toolCalls = response.choices
.map(choice => choice.message?.tool_calls)
.filter(calls => Array.isArray(calls) && calls.length > 0)
.flat();
if (toolCalls.length > 0) {
span.setAttributes({
[]: JSON.stringify(toolCalls),
});
}
}
}
}
/**
* Add attributes for Responses API responses
*/
function addResponsesApiAttributes(span, response, recordOutputs) {
setCommonResponseAttributes(span, response.id, response.model, response.created_at);
if (response.status) {
span.setAttributes({
[]: JSON.stringify([response.status]),
});
}
if (response.usage) {
setTokenUsageAttributes(
span,
response.usage.input_tokens,
response.usage.output_tokens,
response.usage.total_tokens,
);
}
// Extract function calls from output (only if recordOutputs is true)
if (recordOutputs) {
const responseWithOutput = response ;
if (Array.isArray(responseWithOutput.output) && responseWithOutput.output.length > 0) {
// Filter for function_call type objects in the output array
const functionCalls = responseWithOutput.output.filter(
(item) =>
typeof item === 'object' && item !== null && (item ).type === 'function_call',
);
if (functionCalls.length > 0) {
span.setAttributes({
[]: JSON.stringify(functionCalls),
});
}
}
}
}
/**
* Add attributes for Embeddings API responses
*/
function addEmbeddingsAttributes(span, response) {
span.setAttributes({
[]: response.model,
[]: response.model,
});
if (response.usage) {
setTokenUsageAttributes(span, response.usage.prompt_tokens, undefined, response.usage.total_tokens);
}
}
/**
* Add attributes for Conversations API responses
* @see https://platform.openai.com/docs/api-reference/conversations
*/
function addConversationAttributes(span, response) {
const { id, created_at } = response;
span.setAttributes({
[]: id,
[]: id,
// The conversation id is used to link messages across API calls
[]: id,
});
if (created_at) {
span.setAttributes({
[]: new Date(created_at * 1000).toISOString(),
});
}
}
/**
* Set token usage attributes
* @param span - The span to add attributes to
* @param promptTokens - The number of prompt tokens
* @param completionTokens - The number of completion tokens
* @param totalTokens - The number of total tokens
*/
function setTokenUsageAttributes(
span,
promptTokens,
completionTokens,
totalTokens,
) {
if (promptTokens !== undefined) {
span.setAttributes({
[]: promptTokens,
[]: promptTokens,
});
}
if (completionTokens !== undefined) {
span.setAttributes({
[]: completionTokens,
[]: completionTokens,
});
}
if (totalTokens !== undefined) {
span.setAttributes({
[]: totalTokens,
});
}
}
/**
* Set common response attributes
* @param span - The span to add attributes to
* @param id - The response id
* @param model - The response model
* @param timestamp - The response timestamp
*/
function setCommonResponseAttributes(span, id, model, timestamp) {
span.setAttributes({
[]: id,
[]: id,
});
span.setAttributes({
[]: model,
[]: model,
});
span.setAttributes({
[]: new Date(timestamp * 1000).toISOString(),
});
}
/**
* Extract conversation ID from request parameters
* Supports both Conversations API and previous_response_id chaining
* @see https://platform.openai.com/docs/guides/conversation-state
*/
function extractConversationId(params) {
// Conversations API: conversation parameter (e.g., "conv_...")
if ('conversation' in params && typeof params.conversation === 'string') {
return params.conversation;
}
// Responses chaining: previous_response_id links to parent response
if ('previous_response_id' in params && typeof params.previous_response_id === 'string') {
return params.previous_response_id;
}
return undefined;
}
/**
* Extract request parameters including model settings and conversation context
*/
function extractRequestParameters(params) {
const attributes = {
[]: params.model ?? 'unknown',
};
if ('temperature' in params) attributes[GEN_AI_REQUEST_TEMPERATURE_ATTRIBUTE] = params.temperature;
if ('top_p' in params) attributes[GEN_AI_REQUEST_TOP_P_ATTRIBUTE] = params.top_p;
if ('frequency_penalty' in params) attributes[GEN_AI_REQUEST_FREQUENCY_PENALTY_ATTRIBUTE] = params.frequency_penalty;
if ('presence_penalty' in params) attributes[GEN_AI_REQUEST_PRESENCE_PENALTY_ATTRIBUTE] = params.presence_penalty;
if ('stream' in params) attributes[GEN_AI_REQUEST_STREAM_ATTRIBUTE] = params.stream;
if ('encoding_format' in params) attributes[GEN_AI_REQUEST_ENCODING_FORMAT_ATTRIBUTE] = params.encoding_format;
if ('dimensions' in params) attributes[GEN_AI_REQUEST_DIMENSIONS_ATTRIBUTE] = params.dimensions;
// Capture conversation ID for linking messages across API calls
const conversationId = extractConversationId(params);
if (conversationId) {
attributes[GEN_AI_CONVERSATION_ID_ATTRIBUTE] = conversationId;
}
return attributes;
}
export { addChatCompletionAttributes, addConversationAttributes, addEmbeddingsAttributes, addResponsesApiAttributes, buildMethodPath, extractRequestParameters, getOperationName, getSpanOperation, isChatCompletionChunk, isChatCompletionResponse, isConversationResponse, isEmbeddingsResponse, isResponsesApiResponse, isResponsesApiStreamEvent, setCommonResponseAttributes, setTokenUsageAttributes, shouldInstrument };
//# sourceMappingURL=utils.js.map