diff --git a/internal/runtime/executor/kiro_executor.go b/internal/runtime/executor/kiro_executor.go index ec376fe1..e346b744 100644 --- a/internal/runtime/executor/kiro_executor.go +++ b/internal/runtime/executor/kiro_executor.go @@ -2894,7 +2894,7 @@ func (e *KiroExecutor) streamToChannel(ctx context.Context, body io.Reader, out if calculatedInputTokens > 0 { localEstimate := totalUsage.InputTokens totalUsage.InputTokens = calculatedInputTokens - log.Infof("kiro: using contextUsagePercentage (%.2f%%) to calculate input tokens: %d (local estimate was: %d)", + log.Debugf("kiro: using contextUsagePercentage (%.2f%%) to calculate input tokens: %d (local estimate was: %d)", upstreamContextPercentage, calculatedInputTokens, localEstimate) } } @@ -2903,7 +2903,7 @@ func (e *KiroExecutor) streamToChannel(ctx context.Context, body io.Reader, out // Log upstream usage information if received if hasUpstreamUsage { - log.Infof("kiro: upstream usage - credits: %.4f, context: %.2f%%, final tokens - input: %d, output: %d, total: %d", + log.Debugf("kiro: upstream usage - credits: %.4f, context: %.2f%%, final tokens - input: %d, output: %d, total: %d", upstreamCreditUsage, upstreamContextPercentage, totalUsage.InputTokens, totalUsage.OutputTokens, totalUsage.TotalTokens) } diff --git a/internal/translator/kiro/openai/kiro_openai_request.go b/internal/translator/kiro/openai/kiro_openai_request.go index 00a05854..e4f3e767 100644 --- a/internal/translator/kiro/openai/kiro_openai_request.go +++ b/internal/translator/kiro/openai/kiro_openai_request.go @@ -244,7 +244,7 @@ IMPORTANT: You MUST use ... tags to show your reasoning pro } else { systemPrompt = thinkingHint } - log.Infof("kiro-openai: injected thinking prompt") + log.Debugf("kiro-openai: injected thinking prompt") } // Process messages and build history