Files
gemini-cli/packages/core/src/telemetry/metrics.ts
T

685 lines
19 KiB
TypeScript
Raw Normal View History

/**
* @license
* Copyright 2025 Google LLC
* SPDX-License-Identifier: Apache-2.0
*/
import type { Attributes, Meter, Counter, Histogram } from '@opentelemetry/api';
import { diag, metrics, ValueType } from '@opentelemetry/api';
import {
SERVICE_NAME,
METRIC_TOOL_CALL_COUNT,
METRIC_TOOL_CALL_LATENCY,
METRIC_API_REQUEST_COUNT,
METRIC_API_REQUEST_LATENCY,
2025-06-11 06:56:53 +00:00
METRIC_TOKEN_USAGE,
METRIC_SESSION_COUNT,
2025-06-15 16:24:53 -04:00
METRIC_FILE_OPERATION_COUNT,
EVENT_CHAT_COMPRESSION,
METRIC_INVALID_CHUNK_COUNT,
METRIC_CONTENT_RETRY_COUNT,
METRIC_CONTENT_RETRY_FAILURE_COUNT,
METRIC_MODEL_ROUTING_LATENCY,
METRIC_MODEL_ROUTING_FAILURE_COUNT,
2025-09-23 18:06:03 -04:00
METRIC_MODEL_SLASH_COMMAND_CALL_COUNT,
// Performance Monitoring Metrics
METRIC_STARTUP_TIME,
METRIC_MEMORY_USAGE,
METRIC_CPU_USAGE,
METRIC_TOOL_QUEUE_DEPTH,
METRIC_TOOL_EXECUTION_BREAKDOWN,
METRIC_TOKEN_EFFICIENCY,
METRIC_API_REQUEST_BREAKDOWN,
METRIC_PERFORMANCE_SCORE,
METRIC_REGRESSION_DETECTION,
METRIC_REGRESSION_PERCENTAGE_CHANGE,
METRIC_BASELINE_COMPARISON,
} from './constants.js';
import type { Config } from '../config/config.js';
2025-09-23 18:06:03 -04:00
import type { ModelRoutingEvent, ModelSlashCommandEvent } from './types.js';
2025-06-15 16:24:53 -04:00
export enum FileOperation {
CREATE = 'create',
READ = 'read',
UPDATE = 'update',
}
export enum PerformanceMetricType {
STARTUP = 'startup',
MEMORY = 'memory',
CPU = 'cpu',
TOOL_EXECUTION = 'tool_execution',
API_REQUEST = 'api_request',
TOKEN_EFFICIENCY = 'token_efficiency',
}
export enum MemoryMetricType {
HEAP_USED = 'heap_used',
HEAP_TOTAL = 'heap_total',
EXTERNAL = 'external',
RSS = 'rss',
}
export enum ToolExecutionPhase {
VALIDATION = 'validation',
PREPARATION = 'preparation',
EXECUTION = 'execution',
RESULT_PROCESSING = 'result_processing',
}
export enum ApiRequestPhase {
REQUEST_PREPARATION = 'request_preparation',
NETWORK_LATENCY = 'network_latency',
RESPONSE_PROCESSING = 'response_processing',
TOKEN_PROCESSING = 'token_processing',
}
let cliMeter: Meter | undefined;
let toolCallCounter: Counter | undefined;
let toolCallLatencyHistogram: Histogram | undefined;
let apiRequestCounter: Counter | undefined;
let apiRequestLatencyHistogram: Histogram | undefined;
2025-06-11 06:56:53 +00:00
let tokenUsageCounter: Counter | undefined;
2025-06-15 16:24:53 -04:00
let fileOperationCounter: Counter | undefined;
let chatCompressionCounter: Counter | undefined;
let invalidChunkCounter: Counter | undefined;
let contentRetryCounter: Counter | undefined;
let contentRetryFailureCounter: Counter | undefined;
let modelRoutingLatencyHistogram: Histogram | undefined;
let modelRoutingFailureCounter: Counter | undefined;
2025-09-23 18:06:03 -04:00
let modelSlashCommandCallCounter: Counter | undefined;
// Performance Monitoring Metrics
let startupTimeHistogram: Histogram | undefined;
let memoryUsageGauge: Histogram | undefined; // Using Histogram until ObservableGauge is available
let cpuUsageGauge: Histogram | undefined;
let toolQueueDepthGauge: Histogram | undefined;
let toolExecutionBreakdownHistogram: Histogram | undefined;
let tokenEfficiencyHistogram: Histogram | undefined;
let apiRequestBreakdownHistogram: Histogram | undefined;
let performanceScoreGauge: Histogram | undefined;
let regressionDetectionCounter: Counter | undefined;
let regressionPercentageChangeHistogram: Histogram | undefined;
let baselineComparisonHistogram: Histogram | undefined;
let isMetricsInitialized = false;
let isPerformanceMonitoringEnabled = false;
function getCommonAttributes(config: Config): Attributes {
return {
'session.id': config.getSessionId(),
};
}
export function getMeter(): Meter | undefined {
if (!cliMeter) {
cliMeter = metrics.getMeter(SERVICE_NAME);
}
return cliMeter;
}
export function initializeMetrics(config: Config): void {
if (isMetricsInitialized) return;
const meter = getMeter();
if (!meter) return;
// Initialize core metrics
toolCallCounter = meter.createCounter(METRIC_TOOL_CALL_COUNT, {
description: 'Counts tool calls, tagged by function name and success.',
valueType: ValueType.INT,
});
toolCallLatencyHistogram = meter.createHistogram(METRIC_TOOL_CALL_LATENCY, {
description: 'Latency of tool calls in milliseconds.',
unit: 'ms',
valueType: ValueType.INT,
});
apiRequestCounter = meter.createCounter(METRIC_API_REQUEST_COUNT, {
description: 'Counts API requests, tagged by model and status.',
valueType: ValueType.INT,
});
apiRequestLatencyHistogram = meter.createHistogram(
METRIC_API_REQUEST_LATENCY,
{
description: 'Latency of API requests in milliseconds.',
unit: 'ms',
valueType: ValueType.INT,
},
);
2025-06-11 06:56:53 +00:00
tokenUsageCounter = meter.createCounter(METRIC_TOKEN_USAGE, {
description: 'Counts the total number of tokens used.',
valueType: ValueType.INT,
});
2025-06-15 16:24:53 -04:00
fileOperationCounter = meter.createCounter(METRIC_FILE_OPERATION_COUNT, {
description: 'Counts file operations (create, read, update).',
valueType: ValueType.INT,
});
chatCompressionCounter = meter.createCounter(EVENT_CHAT_COMPRESSION, {
description: 'Counts chat compression events.',
valueType: ValueType.INT,
});
// New counters for content errors
invalidChunkCounter = meter.createCounter(METRIC_INVALID_CHUNK_COUNT, {
description: 'Counts invalid chunks received from a stream.',
valueType: ValueType.INT,
});
contentRetryCounter = meter.createCounter(METRIC_CONTENT_RETRY_COUNT, {
description: 'Counts retries due to content errors (e.g., empty stream).',
valueType: ValueType.INT,
});
contentRetryFailureCounter = meter.createCounter(
METRIC_CONTENT_RETRY_FAILURE_COUNT,
{
description: 'Counts occurrences of all content retries failing.',
valueType: ValueType.INT,
},
);
modelRoutingLatencyHistogram = meter.createHistogram(
METRIC_MODEL_ROUTING_LATENCY,
{
description: 'Latency of model routing decisions in milliseconds.',
unit: 'ms',
valueType: ValueType.INT,
},
);
modelRoutingFailureCounter = meter.createCounter(
METRIC_MODEL_ROUTING_FAILURE_COUNT,
{
description: 'Counts model routing failures.',
valueType: ValueType.INT,
},
);
2025-09-23 18:06:03 -04:00
modelSlashCommandCallCounter = meter.createCounter(
METRIC_MODEL_SLASH_COMMAND_CALL_COUNT,
{
description: 'Counts model slash command calls.',
valueType: ValueType.INT,
},
);
const sessionCounter = meter.createCounter(METRIC_SESSION_COUNT, {
description: 'Count of CLI sessions started.',
valueType: ValueType.INT,
});
sessionCounter.add(1, getCommonAttributes(config));
// Initialize performance monitoring metrics if enabled
initializePerformanceMonitoring(config);
isMetricsInitialized = true;
}
export function recordChatCompressionMetrics(
config: Config,
args: { tokens_before: number; tokens_after: number },
) {
if (!chatCompressionCounter || !isMetricsInitialized) return;
chatCompressionCounter.add(1, {
...getCommonAttributes(config),
...args,
});
}
export function recordToolCallMetrics(
config: Config,
functionName: string,
durationMs: number,
success: boolean,
decision?: 'accept' | 'reject' | 'modify' | 'auto_accept',
tool_type?: 'native' | 'mcp',
): void {
if (!toolCallCounter || !toolCallLatencyHistogram || !isMetricsInitialized)
return;
const metricAttributes: Attributes = {
...getCommonAttributes(config),
function_name: functionName,
success,
decision,
tool_type,
};
toolCallCounter.add(1, metricAttributes);
toolCallLatencyHistogram.record(durationMs, {
...getCommonAttributes(config),
function_name: functionName,
});
}
2025-06-11 06:56:53 +00:00
export function recordTokenUsageMetrics(
config: Config,
model: string,
2025-06-11 06:56:53 +00:00
tokenCount: number,
type: 'input' | 'output' | 'thought' | 'cache' | 'tool',
): void {
2025-06-11 06:56:53 +00:00
if (!tokenUsageCounter || !isMetricsInitialized) return;
tokenUsageCounter.add(tokenCount, {
...getCommonAttributes(config),
model,
type,
});
}
export function recordApiResponseMetrics(
config: Config,
model: string,
durationMs: number,
statusCode?: number | string,
): void {
if (
!apiRequestCounter ||
!apiRequestLatencyHistogram ||
!isMetricsInitialized
)
return;
const metricAttributes: Attributes = {
...getCommonAttributes(config),
model,
status_code: statusCode ?? 'ok',
};
apiRequestCounter.add(1, metricAttributes);
apiRequestLatencyHistogram.record(durationMs, {
...getCommonAttributes(config),
model,
});
}
export function recordApiErrorMetrics(
config: Config,
model: string,
durationMs: number,
statusCode?: number | string,
errorType?: string,
): void {
if (
!apiRequestCounter ||
!apiRequestLatencyHistogram ||
!isMetricsInitialized
)
return;
const metricAttributes: Attributes = {
...getCommonAttributes(config),
model,
status_code: statusCode ?? 'error',
error_type: errorType ?? 'unknown',
};
apiRequestCounter.add(1, metricAttributes);
apiRequestLatencyHistogram.record(durationMs, {
...getCommonAttributes(config),
model,
});
}
2025-06-15 16:24:53 -04:00
export function recordFileOperationMetric(
config: Config,
operation: FileOperation,
lines?: number,
mimetype?: string,
extension?: string,
programming_language?: string,
2025-06-15 16:24:53 -04:00
): void {
if (!fileOperationCounter || !isMetricsInitialized) return;
const attributes: Attributes = {
...getCommonAttributes(config),
operation,
};
if (lines !== undefined) attributes['lines'] = lines;
if (mimetype !== undefined) attributes['mimetype'] = mimetype;
if (extension !== undefined) attributes['extension'] = extension;
if (programming_language !== undefined) {
attributes['programming_language'] = programming_language;
}
2025-06-15 16:24:53 -04:00
fileOperationCounter.add(1, attributes);
}
// --- New Metric Recording Functions ---
/**
* Records a metric for when an invalid chunk is received from a stream.
*/
export function recordInvalidChunk(config: Config): void {
if (!invalidChunkCounter || !isMetricsInitialized) return;
invalidChunkCounter.add(1, getCommonAttributes(config));
}
/**
* Records a metric for when a retry is triggered due to a content error.
*/
export function recordContentRetry(config: Config): void {
if (!contentRetryCounter || !isMetricsInitialized) return;
contentRetryCounter.add(1, getCommonAttributes(config));
}
/**
* Records a metric for when all content error retries have failed for a request.
*/
export function recordContentRetryFailure(config: Config): void {
if (!contentRetryFailureCounter || !isMetricsInitialized) return;
contentRetryFailureCounter.add(1, getCommonAttributes(config));
}
2025-09-23 18:06:03 -04:00
export function recordModelSlashCommand(
config: Config,
event: ModelSlashCommandEvent,
): void {
if (!modelSlashCommandCallCounter || !isMetricsInitialized) return;
modelSlashCommandCallCounter.add(1, {
...getCommonAttributes(config),
'slash_command.model.model_name': event.model_name,
});
}
export function recordModelRoutingMetrics(
config: Config,
event: ModelRoutingEvent,
): void {
if (
!modelRoutingLatencyHistogram ||
!modelRoutingFailureCounter ||
!isMetricsInitialized
)
return;
modelRoutingLatencyHistogram.record(event.routing_latency_ms, {
...getCommonAttributes(config),
'routing.decision_model': event.decision_model,
'routing.decision_source': event.decision_source,
});
if (event.failed) {
modelRoutingFailureCounter.add(1, {
...getCommonAttributes(config),
'routing.decision_source': event.decision_source,
'routing.error_message': event.error_message,
});
}
}
// Performance Monitoring Functions
export function initializePerformanceMonitoring(config: Config): void {
const meter = getMeter();
if (!meter) return;
// Check if performance monitoring is enabled in config
// For now, enable performance monitoring when telemetry is enabled
// TODO: Add specific performance monitoring settings to config
isPerformanceMonitoringEnabled = config.getTelemetryEnabled();
if (!isPerformanceMonitoringEnabled) return;
// Initialize startup time histogram
startupTimeHistogram = meter.createHistogram(METRIC_STARTUP_TIME, {
description:
'CLI startup time in milliseconds, broken down by initialization phase.',
unit: 'ms',
valueType: ValueType.DOUBLE,
});
// Initialize memory usage histogram (using histogram until ObservableGauge is available)
memoryUsageGauge = meter.createHistogram(METRIC_MEMORY_USAGE, {
description: 'Memory usage in bytes.',
unit: 'bytes',
valueType: ValueType.INT,
});
// Initialize CPU usage histogram
cpuUsageGauge = meter.createHistogram(METRIC_CPU_USAGE, {
description: 'CPU usage percentage.',
unit: 'percent',
valueType: ValueType.DOUBLE,
});
// Initialize tool queue depth histogram
toolQueueDepthGauge = meter.createHistogram(METRIC_TOOL_QUEUE_DEPTH, {
description: 'Number of tools in execution queue.',
valueType: ValueType.INT,
});
// Initialize performance breakdowns
toolExecutionBreakdownHistogram = meter.createHistogram(
METRIC_TOOL_EXECUTION_BREAKDOWN,
{
description: 'Tool execution time breakdown by phase in milliseconds.',
unit: 'ms',
valueType: ValueType.INT,
},
);
tokenEfficiencyHistogram = meter.createHistogram(METRIC_TOKEN_EFFICIENCY, {
description:
'Token efficiency metrics (tokens per operation, cache hit rate, etc.).',
valueType: ValueType.DOUBLE,
});
apiRequestBreakdownHistogram = meter.createHistogram(
METRIC_API_REQUEST_BREAKDOWN,
{
description: 'API request time breakdown by phase in milliseconds.',
unit: 'ms',
valueType: ValueType.INT,
},
);
// Initialize performance score and regression detection
performanceScoreGauge = meter.createHistogram(METRIC_PERFORMANCE_SCORE, {
description: 'Composite performance score (0-100).',
unit: 'score',
valueType: ValueType.DOUBLE,
});
regressionDetectionCounter = meter.createCounter(
METRIC_REGRESSION_DETECTION,
{
description: 'Performance regression detection events.',
valueType: ValueType.INT,
},
);
regressionPercentageChangeHistogram = meter.createHistogram(
METRIC_REGRESSION_PERCENTAGE_CHANGE,
{
description:
'Percentage change compared to baseline for detected regressions.',
unit: 'percent',
valueType: ValueType.DOUBLE,
},
);
baselineComparisonHistogram = meter.createHistogram(
METRIC_BASELINE_COMPARISON,
{
description:
'Performance comparison to established baseline (percentage change).',
unit: 'percent',
valueType: ValueType.DOUBLE,
},
);
}
export function recordStartupPerformance(
config: Config,
phase: string,
durationMs: number,
details?: Record<string, string | number | boolean>,
): void {
if (!startupTimeHistogram || !isPerformanceMonitoringEnabled) return;
const attributes: Attributes = {
...getCommonAttributes(config),
phase,
...details,
};
startupTimeHistogram.record(durationMs, attributes);
}
export function recordMemoryUsage(
config: Config,
memoryType: MemoryMetricType,
bytes: number,
component?: string,
): void {
if (!memoryUsageGauge || !isPerformanceMonitoringEnabled) return;
const attributes: Attributes = {
...getCommonAttributes(config),
memory_type: memoryType,
component,
};
memoryUsageGauge.record(bytes, attributes);
}
export function recordCpuUsage(
config: Config,
percentage: number,
component?: string,
): void {
if (!cpuUsageGauge || !isPerformanceMonitoringEnabled) return;
const attributes: Attributes = {
...getCommonAttributes(config),
component,
};
cpuUsageGauge.record(percentage, attributes);
}
export function recordToolQueueDepth(config: Config, queueDepth: number): void {
if (!toolQueueDepthGauge || !isPerformanceMonitoringEnabled) return;
const attributes: Attributes = {
...getCommonAttributes(config),
};
toolQueueDepthGauge.record(queueDepth, attributes);
}
export function recordToolExecutionBreakdown(
config: Config,
functionName: string,
phase: ToolExecutionPhase,
durationMs: number,
): void {
if (!toolExecutionBreakdownHistogram || !isPerformanceMonitoringEnabled)
return;
const attributes: Attributes = {
...getCommonAttributes(config),
function_name: functionName,
phase,
};
toolExecutionBreakdownHistogram.record(durationMs, attributes);
}
export function recordTokenEfficiency(
config: Config,
model: string,
metric: string,
value: number,
context?: string,
): void {
if (!tokenEfficiencyHistogram || !isPerformanceMonitoringEnabled) return;
const attributes: Attributes = {
...getCommonAttributes(config),
model,
metric,
context,
};
tokenEfficiencyHistogram.record(value, attributes);
}
export function recordApiRequestBreakdown(
config: Config,
model: string,
phase: ApiRequestPhase,
durationMs: number,
): void {
if (!apiRequestBreakdownHistogram || !isPerformanceMonitoringEnabled) return;
const attributes: Attributes = {
...getCommonAttributes(config),
model,
phase,
};
apiRequestBreakdownHistogram.record(durationMs, attributes);
}
export function recordPerformanceScore(
config: Config,
score: number,
category: string,
baseline?: number,
): void {
if (!performanceScoreGauge || !isPerformanceMonitoringEnabled) return;
const attributes: Attributes = {
...getCommonAttributes(config),
category,
baseline,
};
performanceScoreGauge.record(score, attributes);
}
export function recordPerformanceRegression(
config: Config,
metric: string,
currentValue: number,
baselineValue: number,
severity: 'low' | 'medium' | 'high',
): void {
if (!regressionDetectionCounter || !isPerformanceMonitoringEnabled) return;
const attributes: Attributes = {
...getCommonAttributes(config),
metric,
severity,
current_value: currentValue,
baseline_value: baselineValue,
};
regressionDetectionCounter.add(1, attributes);
if (baselineValue !== 0 && regressionPercentageChangeHistogram) {
const percentageChange =
((currentValue - baselineValue) / baselineValue) * 100;
regressionPercentageChangeHistogram.record(percentageChange, attributes);
}
}
export function recordBaselineComparison(
config: Config,
metric: string,
currentValue: number,
baselineValue: number,
category: string,
): void {
if (!baselineComparisonHistogram || !isPerformanceMonitoringEnabled) return;
if (baselineValue === 0) {
diag.warn('Baseline value is zero, skipping comparison.');
return;
}
const percentageChange =
((currentValue - baselineValue) / baselineValue) * 100;
const attributes: Attributes = {
...getCommonAttributes(config),
metric,
category,
current_value: currentValue,
baseline_value: baselineValue,
};
baselineComparisonHistogram.record(percentageChange, attributes);
}
// Utility function to check if performance monitoring is enabled
export function isPerformanceMonitoringActive(): boolean {
return isPerformanceMonitoringEnabled && isMetricsInitialized;
}