mirror of
https://github.com/router-for-me/CLIProxyAPIPlus.git
synced 2026-03-29 16:54:41 +00:00
Compare commits
30 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
4eaf769894 | ||
|
|
ebec293497 | ||
|
|
e02ceecd35 | ||
|
|
9116392a45 | ||
|
|
c8b33a8cc3 | ||
|
|
b9d1e70ac2 | ||
|
|
fdf5720217 | ||
|
|
f40bd0cd51 | ||
|
|
e33676bb87 | ||
|
|
b1f1cee1e5 | ||
|
|
a1ecc9ab00 | ||
|
|
2a663d5cba | ||
|
|
ba486ca6b7 | ||
|
|
750b930679 | ||
|
|
3902fd7501 | ||
|
|
4fc3d5e935 | ||
|
|
2d2f4572a7 | ||
|
|
8f4c46f38d | ||
|
|
b6ba51bc2a | ||
|
|
6a66d32d37 | ||
|
|
8d15723195 | ||
|
|
736e0aae86 | ||
|
|
8bf3305b2b | ||
|
|
d00e3ea973 | ||
|
|
89db4e9481 | ||
|
|
e332419081 | ||
|
|
47b9503112 | ||
|
|
3b9253c2be | ||
|
|
d241359153 | ||
|
|
f4d4249ba5 |
128
docker-build.sh
128
docker-build.sh
@@ -5,9 +5,115 @@
|
|||||||
# This script automates the process of building and running the Docker container
|
# This script automates the process of building and running the Docker container
|
||||||
# with version information dynamically injected at build time.
|
# with version information dynamically injected at build time.
|
||||||
|
|
||||||
# Exit immediately if a command exits with a non-zero status.
|
# Hidden feature: Preserve usage statistics across rebuilds
|
||||||
|
# Usage: ./docker-build.sh --with-usage
|
||||||
|
# First run prompts for management API key, saved to temp/stats/.api_secret
|
||||||
|
|
||||||
set -euo pipefail
|
set -euo pipefail
|
||||||
|
|
||||||
|
STATS_DIR="temp/stats"
|
||||||
|
STATS_FILE="${STATS_DIR}/.usage_backup.json"
|
||||||
|
SECRET_FILE="${STATS_DIR}/.api_secret"
|
||||||
|
WITH_USAGE=false
|
||||||
|
|
||||||
|
get_port() {
|
||||||
|
if [[ -f "config.yaml" ]]; then
|
||||||
|
grep -E "^port:" config.yaml | sed -E 's/^port: *["'"'"']?([0-9]+)["'"'"']?.*$/\1/'
|
||||||
|
else
|
||||||
|
echo "8317"
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
|
export_stats_api_secret() {
|
||||||
|
if [[ -f "${SECRET_FILE}" ]]; then
|
||||||
|
API_SECRET=$(cat "${SECRET_FILE}")
|
||||||
|
else
|
||||||
|
if [[ ! -d "${STATS_DIR}" ]]; then
|
||||||
|
mkdir -p "${STATS_DIR}"
|
||||||
|
fi
|
||||||
|
echo "First time using --with-usage. Management API key required."
|
||||||
|
read -r -p "Enter management key: " -s API_SECRET
|
||||||
|
echo
|
||||||
|
echo "${API_SECRET}" > "${SECRET_FILE}"
|
||||||
|
chmod 600 "${SECRET_FILE}"
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
|
check_container_running() {
|
||||||
|
local port
|
||||||
|
port=$(get_port)
|
||||||
|
|
||||||
|
if ! curl -s -o /dev/null -w "%{http_code}" "http://localhost:${port}/" | grep -q "200"; then
|
||||||
|
echo "Error: cli-proxy-api service is not responding at localhost:${port}"
|
||||||
|
echo "Please start the container first or use without --with-usage flag."
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
|
export_stats() {
|
||||||
|
local port
|
||||||
|
port=$(get_port)
|
||||||
|
|
||||||
|
if [[ ! -d "${STATS_DIR}" ]]; then
|
||||||
|
mkdir -p "${STATS_DIR}"
|
||||||
|
fi
|
||||||
|
check_container_running
|
||||||
|
echo "Exporting usage statistics..."
|
||||||
|
EXPORT_RESPONSE=$(curl -s -w "\n%{http_code}" -H "X-Management-Key: ${API_SECRET}" \
|
||||||
|
"http://localhost:${port}/v0/management/usage/export")
|
||||||
|
HTTP_CODE=$(echo "${EXPORT_RESPONSE}" | tail -n1)
|
||||||
|
RESPONSE_BODY=$(echo "${EXPORT_RESPONSE}" | sed '$d')
|
||||||
|
|
||||||
|
if [[ "${HTTP_CODE}" != "200" ]]; then
|
||||||
|
echo "Export failed (HTTP ${HTTP_CODE}): ${RESPONSE_BODY}"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
echo "${RESPONSE_BODY}" > "${STATS_FILE}"
|
||||||
|
echo "Statistics exported to ${STATS_FILE}"
|
||||||
|
}
|
||||||
|
|
||||||
|
import_stats() {
|
||||||
|
local port
|
||||||
|
port=$(get_port)
|
||||||
|
|
||||||
|
echo "Importing usage statistics..."
|
||||||
|
IMPORT_RESPONSE=$(curl -s -w "\n%{http_code}" -X POST \
|
||||||
|
-H "X-Management-Key: ${API_SECRET}" \
|
||||||
|
-H "Content-Type: application/json" \
|
||||||
|
-d @"${STATS_FILE}" \
|
||||||
|
"http://localhost:${port}/v0/management/usage/import")
|
||||||
|
IMPORT_CODE=$(echo "${IMPORT_RESPONSE}" | tail -n1)
|
||||||
|
IMPORT_BODY=$(echo "${IMPORT_RESPONSE}" | sed '$d')
|
||||||
|
|
||||||
|
if [[ "${IMPORT_CODE}" == "200" ]]; then
|
||||||
|
echo "Statistics imported successfully"
|
||||||
|
else
|
||||||
|
echo "Import failed (HTTP ${IMPORT_CODE}): ${IMPORT_BODY}"
|
||||||
|
fi
|
||||||
|
|
||||||
|
rm -f "${STATS_FILE}"
|
||||||
|
}
|
||||||
|
|
||||||
|
wait_for_service() {
|
||||||
|
local port
|
||||||
|
port=$(get_port)
|
||||||
|
|
||||||
|
echo "Waiting for service to be ready..."
|
||||||
|
for i in {1..30}; do
|
||||||
|
if curl -s -o /dev/null -w "%{http_code}" "http://localhost:${port}/" | grep -q "200"; then
|
||||||
|
break
|
||||||
|
fi
|
||||||
|
sleep 1
|
||||||
|
done
|
||||||
|
sleep 2
|
||||||
|
}
|
||||||
|
|
||||||
|
if [[ "${1:-}" == "--with-usage" ]]; then
|
||||||
|
WITH_USAGE=true
|
||||||
|
export_stats_api_secret
|
||||||
|
fi
|
||||||
|
|
||||||
# --- Step 1: Choose Environment ---
|
# --- Step 1: Choose Environment ---
|
||||||
echo "Please select an option:"
|
echo "Please select an option:"
|
||||||
echo "1) Run using Pre-built Image (Recommended)"
|
echo "1) Run using Pre-built Image (Recommended)"
|
||||||
@@ -18,7 +124,14 @@ read -r -p "Enter choice [1-2]: " choice
|
|||||||
case "$choice" in
|
case "$choice" in
|
||||||
1)
|
1)
|
||||||
echo "--- Running with Pre-built Image ---"
|
echo "--- Running with Pre-built Image ---"
|
||||||
|
if [[ "${WITH_USAGE}" == "true" ]]; then
|
||||||
|
export_stats
|
||||||
|
fi
|
||||||
docker compose up -d --remove-orphans --no-build
|
docker compose up -d --remove-orphans --no-build
|
||||||
|
if [[ "${WITH_USAGE}" == "true" ]]; then
|
||||||
|
wait_for_service
|
||||||
|
import_stats
|
||||||
|
fi
|
||||||
echo "Services are starting from remote image."
|
echo "Services are starting from remote image."
|
||||||
echo "Run 'docker compose logs -f' to see the logs."
|
echo "Run 'docker compose logs -f' to see the logs."
|
||||||
;;
|
;;
|
||||||
@@ -38,7 +151,11 @@ case "$choice" in
|
|||||||
|
|
||||||
# Build and start the services with a local-only image tag
|
# Build and start the services with a local-only image tag
|
||||||
export CLI_PROXY_IMAGE="cli-proxy-api:local"
|
export CLI_PROXY_IMAGE="cli-proxy-api:local"
|
||||||
|
|
||||||
|
if [[ "${WITH_USAGE}" == "true" ]]; then
|
||||||
|
export_stats
|
||||||
|
fi
|
||||||
|
|
||||||
echo "Building the Docker image..."
|
echo "Building the Docker image..."
|
||||||
docker compose build \
|
docker compose build \
|
||||||
--build-arg VERSION="${VERSION}" \
|
--build-arg VERSION="${VERSION}" \
|
||||||
@@ -48,6 +165,11 @@ case "$choice" in
|
|||||||
echo "Starting the services..."
|
echo "Starting the services..."
|
||||||
docker compose up -d --remove-orphans --pull never
|
docker compose up -d --remove-orphans --pull never
|
||||||
|
|
||||||
|
if [[ "${WITH_USAGE}" == "true" ]]; then
|
||||||
|
wait_for_service
|
||||||
|
import_stats
|
||||||
|
fi
|
||||||
|
|
||||||
echo "Build complete. Services are starting."
|
echo "Build complete. Services are starting."
|
||||||
echo "Run 'docker compose logs -f' to see the logs."
|
echo "Run 'docker compose logs -f' to see the logs."
|
||||||
;;
|
;;
|
||||||
@@ -55,4 +177,4 @@ case "$choice" in
|
|||||||
echo "Invalid choice. Please enter 1 or 2."
|
echo "Invalid choice. Please enter 1 or 2."
|
||||||
exit 1
|
exit 1
|
||||||
;;
|
;;
|
||||||
esac
|
esac
|
||||||
|
|||||||
@@ -33,6 +33,7 @@ import (
|
|||||||
"github.com/router-for-me/CLIProxyAPI/v6/sdk/api/handlers/claude"
|
"github.com/router-for-me/CLIProxyAPI/v6/sdk/api/handlers/claude"
|
||||||
"github.com/router-for-me/CLIProxyAPI/v6/sdk/api/handlers/gemini"
|
"github.com/router-for-me/CLIProxyAPI/v6/sdk/api/handlers/gemini"
|
||||||
"github.com/router-for-me/CLIProxyAPI/v6/sdk/api/handlers/openai"
|
"github.com/router-for-me/CLIProxyAPI/v6/sdk/api/handlers/openai"
|
||||||
|
sdkAuth "github.com/router-for-me/CLIProxyAPI/v6/sdk/auth"
|
||||||
"github.com/router-for-me/CLIProxyAPI/v6/sdk/cliproxy/auth"
|
"github.com/router-for-me/CLIProxyAPI/v6/sdk/cliproxy/auth"
|
||||||
log "github.com/sirupsen/logrus"
|
log "github.com/sirupsen/logrus"
|
||||||
"gopkg.in/yaml.v3"
|
"gopkg.in/yaml.v3"
|
||||||
@@ -987,8 +988,12 @@ func (s *Server) UpdateClients(cfg *config.Config) {
|
|||||||
log.Warnf("amp module is nil, skipping config update")
|
log.Warnf("amp module is nil, skipping config update")
|
||||||
}
|
}
|
||||||
|
|
||||||
// Count client sources from configuration and auth directory
|
// Count client sources from configuration and auth store.
|
||||||
authFiles := util.CountAuthFiles(cfg.AuthDir)
|
tokenStore := sdkAuth.GetTokenStore()
|
||||||
|
if dirSetter, ok := tokenStore.(interface{ SetBaseDir(string) }); ok {
|
||||||
|
dirSetter.SetBaseDir(cfg.AuthDir)
|
||||||
|
}
|
||||||
|
authEntries := util.CountAuthFiles(context.Background(), tokenStore)
|
||||||
geminiAPIKeyCount := len(cfg.GeminiKey)
|
geminiAPIKeyCount := len(cfg.GeminiKey)
|
||||||
claudeAPIKeyCount := len(cfg.ClaudeKey)
|
claudeAPIKeyCount := len(cfg.ClaudeKey)
|
||||||
codexAPIKeyCount := len(cfg.CodexKey)
|
codexAPIKeyCount := len(cfg.CodexKey)
|
||||||
@@ -999,10 +1004,10 @@ func (s *Server) UpdateClients(cfg *config.Config) {
|
|||||||
openAICompatCount += len(entry.APIKeyEntries)
|
openAICompatCount += len(entry.APIKeyEntries)
|
||||||
}
|
}
|
||||||
|
|
||||||
total := authFiles + geminiAPIKeyCount + claudeAPIKeyCount + codexAPIKeyCount + vertexAICompatCount + openAICompatCount
|
total := authEntries + geminiAPIKeyCount + claudeAPIKeyCount + codexAPIKeyCount + vertexAICompatCount + openAICompatCount
|
||||||
fmt.Printf("server clients and configuration updated: %d clients (%d auth files + %d Gemini API keys + %d Claude API keys + %d Codex keys + %d Vertex-compat + %d OpenAI-compat)\n",
|
fmt.Printf("server clients and configuration updated: %d clients (%d auth entries + %d Gemini API keys + %d Claude API keys + %d Codex keys + %d Vertex-compat + %d OpenAI-compat)\n",
|
||||||
total,
|
total,
|
||||||
authFiles,
|
authEntries,
|
||||||
geminiAPIKeyCount,
|
geminiAPIKeyCount,
|
||||||
claudeAPIKeyCount,
|
claudeAPIKeyCount,
|
||||||
codexAPIKeyCount,
|
codexAPIKeyCount,
|
||||||
|
|||||||
@@ -740,7 +740,7 @@ func GetIFlowModels() []*ModelInfo {
|
|||||||
{ID: "qwen3-235b-a22b-thinking-2507", DisplayName: "Qwen3-235B-A22B-Thinking", Description: "Qwen3 235B A22B Thinking (2507)", Created: 1753401600},
|
{ID: "qwen3-235b-a22b-thinking-2507", DisplayName: "Qwen3-235B-A22B-Thinking", Description: "Qwen3 235B A22B Thinking (2507)", Created: 1753401600},
|
||||||
{ID: "qwen3-235b-a22b-instruct", DisplayName: "Qwen3-235B-A22B-Instruct", Description: "Qwen3 235B A22B Instruct", Created: 1753401600},
|
{ID: "qwen3-235b-a22b-instruct", DisplayName: "Qwen3-235B-A22B-Instruct", Description: "Qwen3 235B A22B Instruct", Created: 1753401600},
|
||||||
{ID: "qwen3-235b", DisplayName: "Qwen3-235B-A22B", Description: "Qwen3 235B A22B", Created: 1753401600},
|
{ID: "qwen3-235b", DisplayName: "Qwen3-235B-A22B", Description: "Qwen3 235B A22B", Created: 1753401600},
|
||||||
{ID: "minimax-m2", DisplayName: "MiniMax-M2", Description: "MiniMax M2", Created: 1758672000},
|
{ID: "minimax-m2", DisplayName: "MiniMax-M2", Description: "MiniMax M2", Created: 1758672000, Thinking: iFlowThinkingSupport},
|
||||||
{ID: "minimax-m2.1", DisplayName: "MiniMax-M2.1", Description: "MiniMax M2.1", Created: 1766448000, Thinking: iFlowThinkingSupport},
|
{ID: "minimax-m2.1", DisplayName: "MiniMax-M2.1", Description: "MiniMax M2.1", Created: 1766448000, Thinking: iFlowThinkingSupport},
|
||||||
}
|
}
|
||||||
models := make([]*ModelInfo, 0, len(entries))
|
models := make([]*ModelInfo, 0, len(entries))
|
||||||
@@ -773,7 +773,7 @@ func GetAntigravityModelConfig() map[string]*AntigravityModelConfig {
|
|||||||
return map[string]*AntigravityModelConfig{
|
return map[string]*AntigravityModelConfig{
|
||||||
"gemini-2.5-flash": {Thinking: &ThinkingSupport{Min: 0, Max: 24576, ZeroAllowed: true, DynamicAllowed: true}, Name: "models/gemini-2.5-flash"},
|
"gemini-2.5-flash": {Thinking: &ThinkingSupport{Min: 0, Max: 24576, ZeroAllowed: true, DynamicAllowed: true}, Name: "models/gemini-2.5-flash"},
|
||||||
"gemini-2.5-flash-lite": {Thinking: &ThinkingSupport{Min: 0, Max: 24576, ZeroAllowed: true, DynamicAllowed: true}, Name: "models/gemini-2.5-flash-lite"},
|
"gemini-2.5-flash-lite": {Thinking: &ThinkingSupport{Min: 0, Max: 24576, ZeroAllowed: true, DynamicAllowed: true}, Name: "models/gemini-2.5-flash-lite"},
|
||||||
"gemini-2.5-computer-use-preview-10-2025": {Name: "models/gemini-2.5-computer-use-preview-10-2025"},
|
"gemini-2.5-computer-use-preview-10-2025": {Thinking: &ThinkingSupport{Min: 128, Max: 32768, ZeroAllowed: false, DynamicAllowed: true}, Name: "models/gemini-2.5-computer-use-preview-10-2025"},
|
||||||
"gemini-3-pro-preview": {Thinking: &ThinkingSupport{Min: 128, Max: 32768, ZeroAllowed: false, DynamicAllowed: true, Levels: []string{"low", "high"}}, Name: "models/gemini-3-pro-preview"},
|
"gemini-3-pro-preview": {Thinking: &ThinkingSupport{Min: 128, Max: 32768, ZeroAllowed: false, DynamicAllowed: true, Levels: []string{"low", "high"}}, Name: "models/gemini-3-pro-preview"},
|
||||||
"gemini-3-pro-image-preview": {Thinking: &ThinkingSupport{Min: 128, Max: 32768, ZeroAllowed: false, DynamicAllowed: true, Levels: []string{"low", "high"}}, Name: "models/gemini-3-pro-image-preview"},
|
"gemini-3-pro-image-preview": {Thinking: &ThinkingSupport{Min: 128, Max: 32768, ZeroAllowed: false, DynamicAllowed: true, Levels: []string{"low", "high"}}, Name: "models/gemini-3-pro-image-preview"},
|
||||||
"gemini-3-flash-preview": {Thinking: &ThinkingSupport{Min: 128, Max: 32768, ZeroAllowed: false, DynamicAllowed: true, Levels: []string{"minimal", "low", "medium", "high"}}, Name: "models/gemini-3-flash-preview"},
|
"gemini-3-flash-preview": {Thinking: &ThinkingSupport{Min: 128, Max: 32768, ZeroAllowed: false, DynamicAllowed: true, Levels: []string{"minimal", "low", "medium", "high"}}, Name: "models/gemini-3-flash-preview"},
|
||||||
|
|||||||
@@ -4,6 +4,7 @@
|
|||||||
package registry
|
package registry
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
"sort"
|
"sort"
|
||||||
"strings"
|
"strings"
|
||||||
@@ -84,6 +85,13 @@ type ModelRegistration struct {
|
|||||||
SuspendedClients map[string]string
|
SuspendedClients map[string]string
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// ModelRegistryHook provides optional callbacks for external integrations to track model list changes.
|
||||||
|
// Hook implementations must be non-blocking and resilient; calls are executed asynchronously and panics are recovered.
|
||||||
|
type ModelRegistryHook interface {
|
||||||
|
OnModelsRegistered(ctx context.Context, provider, clientID string, models []*ModelInfo)
|
||||||
|
OnModelsUnregistered(ctx context.Context, provider, clientID string)
|
||||||
|
}
|
||||||
|
|
||||||
// ModelRegistry manages the global registry of available models
|
// ModelRegistry manages the global registry of available models
|
||||||
type ModelRegistry struct {
|
type ModelRegistry struct {
|
||||||
// models maps model ID to registration information
|
// models maps model ID to registration information
|
||||||
@@ -97,6 +105,8 @@ type ModelRegistry struct {
|
|||||||
clientProviders map[string]string
|
clientProviders map[string]string
|
||||||
// mutex ensures thread-safe access to the registry
|
// mutex ensures thread-safe access to the registry
|
||||||
mutex *sync.RWMutex
|
mutex *sync.RWMutex
|
||||||
|
// hook is an optional callback sink for model registration changes
|
||||||
|
hook ModelRegistryHook
|
||||||
}
|
}
|
||||||
|
|
||||||
// Global model registry instance
|
// Global model registry instance
|
||||||
@@ -117,6 +127,53 @@ func GetGlobalRegistry() *ModelRegistry {
|
|||||||
return globalRegistry
|
return globalRegistry
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// SetHook sets an optional hook for observing model registration changes.
|
||||||
|
func (r *ModelRegistry) SetHook(hook ModelRegistryHook) {
|
||||||
|
if r == nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
r.mutex.Lock()
|
||||||
|
defer r.mutex.Unlock()
|
||||||
|
r.hook = hook
|
||||||
|
}
|
||||||
|
|
||||||
|
const defaultModelRegistryHookTimeout = 5 * time.Second
|
||||||
|
|
||||||
|
func (r *ModelRegistry) triggerModelsRegistered(provider, clientID string, models []*ModelInfo) {
|
||||||
|
hook := r.hook
|
||||||
|
if hook == nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
modelsCopy := cloneModelInfosUnique(models)
|
||||||
|
go func() {
|
||||||
|
defer func() {
|
||||||
|
if recovered := recover(); recovered != nil {
|
||||||
|
log.Errorf("model registry hook OnModelsRegistered panic: %v", recovered)
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
ctx, cancel := context.WithTimeout(context.Background(), defaultModelRegistryHookTimeout)
|
||||||
|
defer cancel()
|
||||||
|
hook.OnModelsRegistered(ctx, provider, clientID, modelsCopy)
|
||||||
|
}()
|
||||||
|
}
|
||||||
|
|
||||||
|
func (r *ModelRegistry) triggerModelsUnregistered(provider, clientID string) {
|
||||||
|
hook := r.hook
|
||||||
|
if hook == nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
go func() {
|
||||||
|
defer func() {
|
||||||
|
if recovered := recover(); recovered != nil {
|
||||||
|
log.Errorf("model registry hook OnModelsUnregistered panic: %v", recovered)
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
ctx, cancel := context.WithTimeout(context.Background(), defaultModelRegistryHookTimeout)
|
||||||
|
defer cancel()
|
||||||
|
hook.OnModelsUnregistered(ctx, provider, clientID)
|
||||||
|
}()
|
||||||
|
}
|
||||||
|
|
||||||
// RegisterClient registers a client and its supported models
|
// RegisterClient registers a client and its supported models
|
||||||
// Parameters:
|
// Parameters:
|
||||||
// - clientID: Unique identifier for the client
|
// - clientID: Unique identifier for the client
|
||||||
@@ -177,6 +234,7 @@ func (r *ModelRegistry) RegisterClient(clientID, clientProvider string, models [
|
|||||||
} else {
|
} else {
|
||||||
delete(r.clientProviders, clientID)
|
delete(r.clientProviders, clientID)
|
||||||
}
|
}
|
||||||
|
r.triggerModelsRegistered(provider, clientID, models)
|
||||||
log.Debugf("Registered client %s from provider %s with %d models", clientID, clientProvider, len(rawModelIDs))
|
log.Debugf("Registered client %s from provider %s with %d models", clientID, clientProvider, len(rawModelIDs))
|
||||||
misc.LogCredentialSeparator()
|
misc.LogCredentialSeparator()
|
||||||
return
|
return
|
||||||
@@ -310,6 +368,7 @@ func (r *ModelRegistry) RegisterClient(clientID, clientProvider string, models [
|
|||||||
delete(r.clientProviders, clientID)
|
delete(r.clientProviders, clientID)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
r.triggerModelsRegistered(provider, clientID, models)
|
||||||
if len(added) == 0 && len(removed) == 0 && !providerChanged {
|
if len(added) == 0 && len(removed) == 0 && !providerChanged {
|
||||||
// Only metadata (e.g., display name) changed; skip separator when no log output.
|
// Only metadata (e.g., display name) changed; skip separator when no log output.
|
||||||
return
|
return
|
||||||
@@ -400,6 +459,25 @@ func cloneModelInfo(model *ModelInfo) *ModelInfo {
|
|||||||
return ©Model
|
return ©Model
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func cloneModelInfosUnique(models []*ModelInfo) []*ModelInfo {
|
||||||
|
if len(models) == 0 {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
cloned := make([]*ModelInfo, 0, len(models))
|
||||||
|
seen := make(map[string]struct{}, len(models))
|
||||||
|
for _, model := range models {
|
||||||
|
if model == nil || model.ID == "" {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if _, exists := seen[model.ID]; exists {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
seen[model.ID] = struct{}{}
|
||||||
|
cloned = append(cloned, cloneModelInfo(model))
|
||||||
|
}
|
||||||
|
return cloned
|
||||||
|
}
|
||||||
|
|
||||||
// UnregisterClient removes a client and decrements counts for its models
|
// UnregisterClient removes a client and decrements counts for its models
|
||||||
// Parameters:
|
// Parameters:
|
||||||
// - clientID: Unique identifier for the client to remove
|
// - clientID: Unique identifier for the client to remove
|
||||||
@@ -460,6 +538,7 @@ func (r *ModelRegistry) unregisterClientInternal(clientID string) {
|
|||||||
log.Debugf("Unregistered client %s", clientID)
|
log.Debugf("Unregistered client %s", clientID)
|
||||||
// Separator line after completing client unregistration (after the summary line)
|
// Separator line after completing client unregistration (after the summary line)
|
||||||
misc.LogCredentialSeparator()
|
misc.LogCredentialSeparator()
|
||||||
|
r.triggerModelsUnregistered(provider, clientID)
|
||||||
}
|
}
|
||||||
|
|
||||||
// SetModelQuotaExceeded marks a model as quota exceeded for a specific client
|
// SetModelQuotaExceeded marks a model as quota exceeded for a specific client
|
||||||
@@ -625,6 +704,131 @@ func (r *ModelRegistry) GetAvailableModels(handlerType string) []map[string]any
|
|||||||
return models
|
return models
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// GetAvailableModelsByProvider returns models available for the given provider identifier.
|
||||||
|
// Parameters:
|
||||||
|
// - provider: Provider identifier (e.g., "codex", "gemini", "antigravity")
|
||||||
|
//
|
||||||
|
// Returns:
|
||||||
|
// - []*ModelInfo: List of available models for the provider
|
||||||
|
func (r *ModelRegistry) GetAvailableModelsByProvider(provider string) []*ModelInfo {
|
||||||
|
provider = strings.ToLower(strings.TrimSpace(provider))
|
||||||
|
if provider == "" {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
r.mutex.RLock()
|
||||||
|
defer r.mutex.RUnlock()
|
||||||
|
|
||||||
|
type providerModel struct {
|
||||||
|
count int
|
||||||
|
info *ModelInfo
|
||||||
|
}
|
||||||
|
|
||||||
|
providerModels := make(map[string]*providerModel)
|
||||||
|
|
||||||
|
for clientID, clientProvider := range r.clientProviders {
|
||||||
|
if clientProvider != provider {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
modelIDs := r.clientModels[clientID]
|
||||||
|
if len(modelIDs) == 0 {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
clientInfos := r.clientModelInfos[clientID]
|
||||||
|
for _, modelID := range modelIDs {
|
||||||
|
modelID = strings.TrimSpace(modelID)
|
||||||
|
if modelID == "" {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
entry := providerModels[modelID]
|
||||||
|
if entry == nil {
|
||||||
|
entry = &providerModel{}
|
||||||
|
providerModels[modelID] = entry
|
||||||
|
}
|
||||||
|
entry.count++
|
||||||
|
if entry.info == nil {
|
||||||
|
if clientInfos != nil {
|
||||||
|
if info := clientInfos[modelID]; info != nil {
|
||||||
|
entry.info = info
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if entry.info == nil {
|
||||||
|
if reg, ok := r.models[modelID]; ok && reg != nil && reg.Info != nil {
|
||||||
|
entry.info = reg.Info
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(providerModels) == 0 {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
quotaExpiredDuration := 5 * time.Minute
|
||||||
|
now := time.Now()
|
||||||
|
result := make([]*ModelInfo, 0, len(providerModels))
|
||||||
|
|
||||||
|
for modelID, entry := range providerModels {
|
||||||
|
if entry == nil || entry.count <= 0 {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
registration, ok := r.models[modelID]
|
||||||
|
|
||||||
|
expiredClients := 0
|
||||||
|
cooldownSuspended := 0
|
||||||
|
otherSuspended := 0
|
||||||
|
if ok && registration != nil {
|
||||||
|
if registration.QuotaExceededClients != nil {
|
||||||
|
for clientID, quotaTime := range registration.QuotaExceededClients {
|
||||||
|
if clientID == "" {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if p, okProvider := r.clientProviders[clientID]; !okProvider || p != provider {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if quotaTime != nil && now.Sub(*quotaTime) < quotaExpiredDuration {
|
||||||
|
expiredClients++
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if registration.SuspendedClients != nil {
|
||||||
|
for clientID, reason := range registration.SuspendedClients {
|
||||||
|
if clientID == "" {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if p, okProvider := r.clientProviders[clientID]; !okProvider || p != provider {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if strings.EqualFold(reason, "quota") {
|
||||||
|
cooldownSuspended++
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
otherSuspended++
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
availableClients := entry.count
|
||||||
|
effectiveClients := availableClients - expiredClients - otherSuspended
|
||||||
|
if effectiveClients < 0 {
|
||||||
|
effectiveClients = 0
|
||||||
|
}
|
||||||
|
|
||||||
|
if effectiveClients > 0 || (availableClients > 0 && (expiredClients > 0 || cooldownSuspended > 0) && otherSuspended == 0) {
|
||||||
|
if entry.info != nil {
|
||||||
|
result = append(result, entry.info)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if ok && registration != nil && registration.Info != nil {
|
||||||
|
result = append(result, registration.Info)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return result
|
||||||
|
}
|
||||||
|
|
||||||
// GetModelCount returns the number of available clients for a specific model
|
// GetModelCount returns the number of available clients for a specific model
|
||||||
// Parameters:
|
// Parameters:
|
||||||
// - modelID: The model ID to check
|
// - modelID: The model ID to check
|
||||||
|
|||||||
204
internal/registry/model_registry_hook_test.go
Normal file
204
internal/registry/model_registry_hook_test.go
Normal file
@@ -0,0 +1,204 @@
|
|||||||
|
package registry
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"sync"
|
||||||
|
"testing"
|
||||||
|
"time"
|
||||||
|
)
|
||||||
|
|
||||||
|
func newTestModelRegistry() *ModelRegistry {
|
||||||
|
return &ModelRegistry{
|
||||||
|
models: make(map[string]*ModelRegistration),
|
||||||
|
clientModels: make(map[string][]string),
|
||||||
|
clientModelInfos: make(map[string]map[string]*ModelInfo),
|
||||||
|
clientProviders: make(map[string]string),
|
||||||
|
mutex: &sync.RWMutex{},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
type registeredCall struct {
|
||||||
|
provider string
|
||||||
|
clientID string
|
||||||
|
models []*ModelInfo
|
||||||
|
}
|
||||||
|
|
||||||
|
type unregisteredCall struct {
|
||||||
|
provider string
|
||||||
|
clientID string
|
||||||
|
}
|
||||||
|
|
||||||
|
type capturingHook struct {
|
||||||
|
registeredCh chan registeredCall
|
||||||
|
unregisteredCh chan unregisteredCall
|
||||||
|
}
|
||||||
|
|
||||||
|
func (h *capturingHook) OnModelsRegistered(ctx context.Context, provider, clientID string, models []*ModelInfo) {
|
||||||
|
h.registeredCh <- registeredCall{provider: provider, clientID: clientID, models: models}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (h *capturingHook) OnModelsUnregistered(ctx context.Context, provider, clientID string) {
|
||||||
|
h.unregisteredCh <- unregisteredCall{provider: provider, clientID: clientID}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestModelRegistryHook_OnModelsRegisteredCalled(t *testing.T) {
|
||||||
|
r := newTestModelRegistry()
|
||||||
|
hook := &capturingHook{
|
||||||
|
registeredCh: make(chan registeredCall, 1),
|
||||||
|
unregisteredCh: make(chan unregisteredCall, 1),
|
||||||
|
}
|
||||||
|
r.SetHook(hook)
|
||||||
|
|
||||||
|
inputModels := []*ModelInfo{
|
||||||
|
{ID: "m1", DisplayName: "Model One"},
|
||||||
|
{ID: "m2", DisplayName: "Model Two"},
|
||||||
|
}
|
||||||
|
r.RegisterClient("client-1", "OpenAI", inputModels)
|
||||||
|
|
||||||
|
select {
|
||||||
|
case call := <-hook.registeredCh:
|
||||||
|
if call.provider != "openai" {
|
||||||
|
t.Fatalf("provider mismatch: got %q, want %q", call.provider, "openai")
|
||||||
|
}
|
||||||
|
if call.clientID != "client-1" {
|
||||||
|
t.Fatalf("clientID mismatch: got %q, want %q", call.clientID, "client-1")
|
||||||
|
}
|
||||||
|
if len(call.models) != 2 {
|
||||||
|
t.Fatalf("models length mismatch: got %d, want %d", len(call.models), 2)
|
||||||
|
}
|
||||||
|
if call.models[0] == nil || call.models[0].ID != "m1" {
|
||||||
|
t.Fatalf("models[0] mismatch: got %#v, want ID=%q", call.models[0], "m1")
|
||||||
|
}
|
||||||
|
if call.models[1] == nil || call.models[1].ID != "m2" {
|
||||||
|
t.Fatalf("models[1] mismatch: got %#v, want ID=%q", call.models[1], "m2")
|
||||||
|
}
|
||||||
|
case <-time.After(2 * time.Second):
|
||||||
|
t.Fatal("timeout waiting for OnModelsRegistered hook call")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestModelRegistryHook_OnModelsUnregisteredCalled(t *testing.T) {
|
||||||
|
r := newTestModelRegistry()
|
||||||
|
hook := &capturingHook{
|
||||||
|
registeredCh: make(chan registeredCall, 1),
|
||||||
|
unregisteredCh: make(chan unregisteredCall, 1),
|
||||||
|
}
|
||||||
|
r.SetHook(hook)
|
||||||
|
|
||||||
|
r.RegisterClient("client-1", "OpenAI", []*ModelInfo{{ID: "m1"}})
|
||||||
|
select {
|
||||||
|
case <-hook.registeredCh:
|
||||||
|
case <-time.After(2 * time.Second):
|
||||||
|
t.Fatal("timeout waiting for OnModelsRegistered hook call")
|
||||||
|
}
|
||||||
|
|
||||||
|
r.UnregisterClient("client-1")
|
||||||
|
|
||||||
|
select {
|
||||||
|
case call := <-hook.unregisteredCh:
|
||||||
|
if call.provider != "openai" {
|
||||||
|
t.Fatalf("provider mismatch: got %q, want %q", call.provider, "openai")
|
||||||
|
}
|
||||||
|
if call.clientID != "client-1" {
|
||||||
|
t.Fatalf("clientID mismatch: got %q, want %q", call.clientID, "client-1")
|
||||||
|
}
|
||||||
|
case <-time.After(2 * time.Second):
|
||||||
|
t.Fatal("timeout waiting for OnModelsUnregistered hook call")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
type blockingHook struct {
|
||||||
|
started chan struct{}
|
||||||
|
unblock chan struct{}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (h *blockingHook) OnModelsRegistered(ctx context.Context, provider, clientID string, models []*ModelInfo) {
|
||||||
|
select {
|
||||||
|
case <-h.started:
|
||||||
|
default:
|
||||||
|
close(h.started)
|
||||||
|
}
|
||||||
|
<-h.unblock
|
||||||
|
}
|
||||||
|
|
||||||
|
func (h *blockingHook) OnModelsUnregistered(ctx context.Context, provider, clientID string) {}
|
||||||
|
|
||||||
|
func TestModelRegistryHook_DoesNotBlockRegisterClient(t *testing.T) {
|
||||||
|
r := newTestModelRegistry()
|
||||||
|
hook := &blockingHook{
|
||||||
|
started: make(chan struct{}),
|
||||||
|
unblock: make(chan struct{}),
|
||||||
|
}
|
||||||
|
r.SetHook(hook)
|
||||||
|
defer close(hook.unblock)
|
||||||
|
|
||||||
|
done := make(chan struct{})
|
||||||
|
go func() {
|
||||||
|
r.RegisterClient("client-1", "OpenAI", []*ModelInfo{{ID: "m1"}})
|
||||||
|
close(done)
|
||||||
|
}()
|
||||||
|
|
||||||
|
select {
|
||||||
|
case <-hook.started:
|
||||||
|
case <-time.After(2 * time.Second):
|
||||||
|
t.Fatal("timeout waiting for hook to start")
|
||||||
|
}
|
||||||
|
|
||||||
|
select {
|
||||||
|
case <-done:
|
||||||
|
case <-time.After(200 * time.Millisecond):
|
||||||
|
t.Fatal("RegisterClient appears to be blocked by hook")
|
||||||
|
}
|
||||||
|
|
||||||
|
if !r.ClientSupportsModel("client-1", "m1") {
|
||||||
|
t.Fatal("model registration failed; expected client to support model")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
type panicHook struct {
|
||||||
|
registeredCalled chan struct{}
|
||||||
|
unregisteredCalled chan struct{}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (h *panicHook) OnModelsRegistered(ctx context.Context, provider, clientID string, models []*ModelInfo) {
|
||||||
|
if h.registeredCalled != nil {
|
||||||
|
h.registeredCalled <- struct{}{}
|
||||||
|
}
|
||||||
|
panic("boom")
|
||||||
|
}
|
||||||
|
|
||||||
|
func (h *panicHook) OnModelsUnregistered(ctx context.Context, provider, clientID string) {
|
||||||
|
if h.unregisteredCalled != nil {
|
||||||
|
h.unregisteredCalled <- struct{}{}
|
||||||
|
}
|
||||||
|
panic("boom")
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestModelRegistryHook_PanicDoesNotAffectRegistry(t *testing.T) {
|
||||||
|
r := newTestModelRegistry()
|
||||||
|
hook := &panicHook{
|
||||||
|
registeredCalled: make(chan struct{}, 1),
|
||||||
|
unregisteredCalled: make(chan struct{}, 1),
|
||||||
|
}
|
||||||
|
r.SetHook(hook)
|
||||||
|
|
||||||
|
r.RegisterClient("client-1", "OpenAI", []*ModelInfo{{ID: "m1"}})
|
||||||
|
|
||||||
|
select {
|
||||||
|
case <-hook.registeredCalled:
|
||||||
|
case <-time.After(2 * time.Second):
|
||||||
|
t.Fatal("timeout waiting for OnModelsRegistered hook call")
|
||||||
|
}
|
||||||
|
|
||||||
|
if !r.ClientSupportsModel("client-1", "m1") {
|
||||||
|
t.Fatal("model registration failed; expected client to support model")
|
||||||
|
}
|
||||||
|
|
||||||
|
r.UnregisterClient("client-1")
|
||||||
|
|
||||||
|
select {
|
||||||
|
case <-hook.unregisteredCalled:
|
||||||
|
case <-time.After(2 * time.Second):
|
||||||
|
t.Fatal("timeout waiting for OnModelsUnregistered hook call")
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -323,6 +323,11 @@ type translatedPayload struct {
|
|||||||
func (e *AIStudioExecutor) translateRequest(req cliproxyexecutor.Request, opts cliproxyexecutor.Options, stream bool) ([]byte, translatedPayload, error) {
|
func (e *AIStudioExecutor) translateRequest(req cliproxyexecutor.Request, opts cliproxyexecutor.Options, stream bool) ([]byte, translatedPayload, error) {
|
||||||
from := opts.SourceFormat
|
from := opts.SourceFormat
|
||||||
to := sdktranslator.FromString("gemini")
|
to := sdktranslator.FromString("gemini")
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, req.Model, originalPayload, stream)
|
||||||
payload := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), stream)
|
payload := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), stream)
|
||||||
payload = ApplyThinkingMetadata(payload, req.Metadata, req.Model)
|
payload = ApplyThinkingMetadata(payload, req.Metadata, req.Model)
|
||||||
payload = util.ApplyGemini3ThinkingLevelFromMetadata(req.Model, req.Metadata, payload)
|
payload = util.ApplyGemini3ThinkingLevelFromMetadata(req.Model, req.Metadata, payload)
|
||||||
@@ -331,7 +336,7 @@ func (e *AIStudioExecutor) translateRequest(req cliproxyexecutor.Request, opts c
|
|||||||
payload = util.NormalizeGeminiThinkingBudget(req.Model, payload, true)
|
payload = util.NormalizeGeminiThinkingBudget(req.Model, payload, true)
|
||||||
payload = util.StripThinkingConfigIfUnsupported(req.Model, payload)
|
payload = util.StripThinkingConfigIfUnsupported(req.Model, payload)
|
||||||
payload = fixGeminiImageAspectRatio(req.Model, payload)
|
payload = fixGeminiImageAspectRatio(req.Model, payload)
|
||||||
payload = applyPayloadConfig(e.cfg, req.Model, payload)
|
payload = applyPayloadConfigWithRoot(e.cfg, req.Model, to.String(), "", payload, originalTranslated)
|
||||||
payload, _ = sjson.DeleteBytes(payload, "generationConfig.maxOutputTokens")
|
payload, _ = sjson.DeleteBytes(payload, "generationConfig.maxOutputTokens")
|
||||||
payload, _ = sjson.DeleteBytes(payload, "generationConfig.responseMimeType")
|
payload, _ = sjson.DeleteBytes(payload, "generationConfig.responseMimeType")
|
||||||
payload, _ = sjson.DeleteBytes(payload, "generationConfig.responseJsonSchema")
|
payload, _ = sjson.DeleteBytes(payload, "generationConfig.responseJsonSchema")
|
||||||
|
|||||||
@@ -94,13 +94,18 @@ func (e *AntigravityExecutor) Execute(ctx context.Context, auth *cliproxyauth.Au
|
|||||||
|
|
||||||
from := opts.SourceFormat
|
from := opts.SourceFormat
|
||||||
to := sdktranslator.FromString("antigravity")
|
to := sdktranslator.FromString("antigravity")
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, req.Model, originalPayload, false)
|
||||||
translated := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), false)
|
translated := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), false)
|
||||||
|
|
||||||
translated = applyThinkingMetadataCLI(translated, req.Metadata, req.Model)
|
translated = ApplyThinkingMetadataCLI(translated, req.Metadata, req.Model)
|
||||||
translated = util.ApplyGemini3ThinkingLevelFromMetadataCLI(req.Model, req.Metadata, translated)
|
translated = util.ApplyGemini3ThinkingLevelFromMetadataCLI(req.Model, req.Metadata, translated)
|
||||||
translated = util.ApplyDefaultThinkingIfNeededCLI(req.Model, translated)
|
translated = util.ApplyDefaultThinkingIfNeededCLI(req.Model, req.Metadata, translated)
|
||||||
translated = normalizeAntigravityThinking(req.Model, translated, isClaude)
|
translated = normalizeAntigravityThinking(req.Model, translated, isClaude)
|
||||||
translated = applyPayloadConfigWithRoot(e.cfg, req.Model, "antigravity", "request", translated)
|
translated = applyPayloadConfigWithRoot(e.cfg, req.Model, "antigravity", "request", translated, originalTranslated)
|
||||||
|
|
||||||
baseURLs := antigravityBaseURLFallbackOrder(auth)
|
baseURLs := antigravityBaseURLFallbackOrder(auth)
|
||||||
httpClient := newProxyAwareHTTPClient(ctx, e.cfg, auth, 0)
|
httpClient := newProxyAwareHTTPClient(ctx, e.cfg, auth, 0)
|
||||||
@@ -189,13 +194,18 @@ func (e *AntigravityExecutor) executeClaudeNonStream(ctx context.Context, auth *
|
|||||||
|
|
||||||
from := opts.SourceFormat
|
from := opts.SourceFormat
|
||||||
to := sdktranslator.FromString("antigravity")
|
to := sdktranslator.FromString("antigravity")
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, req.Model, originalPayload, true)
|
||||||
translated := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), true)
|
translated := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), true)
|
||||||
|
|
||||||
translated = applyThinkingMetadataCLI(translated, req.Metadata, req.Model)
|
translated = ApplyThinkingMetadataCLI(translated, req.Metadata, req.Model)
|
||||||
translated = util.ApplyGemini3ThinkingLevelFromMetadataCLI(req.Model, req.Metadata, translated)
|
translated = util.ApplyGemini3ThinkingLevelFromMetadataCLI(req.Model, req.Metadata, translated)
|
||||||
translated = util.ApplyDefaultThinkingIfNeededCLI(req.Model, translated)
|
translated = util.ApplyDefaultThinkingIfNeededCLI(req.Model, req.Metadata, translated)
|
||||||
translated = normalizeAntigravityThinking(req.Model, translated, true)
|
translated = normalizeAntigravityThinking(req.Model, translated, true)
|
||||||
translated = applyPayloadConfigWithRoot(e.cfg, req.Model, "antigravity", "request", translated)
|
translated = applyPayloadConfigWithRoot(e.cfg, req.Model, "antigravity", "request", translated, originalTranslated)
|
||||||
|
|
||||||
baseURLs := antigravityBaseURLFallbackOrder(auth)
|
baseURLs := antigravityBaseURLFallbackOrder(auth)
|
||||||
httpClient := newProxyAwareHTTPClient(ctx, e.cfg, auth, 0)
|
httpClient := newProxyAwareHTTPClient(ctx, e.cfg, auth, 0)
|
||||||
@@ -525,13 +535,18 @@ func (e *AntigravityExecutor) ExecuteStream(ctx context.Context, auth *cliproxya
|
|||||||
|
|
||||||
from := opts.SourceFormat
|
from := opts.SourceFormat
|
||||||
to := sdktranslator.FromString("antigravity")
|
to := sdktranslator.FromString("antigravity")
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, req.Model, originalPayload, true)
|
||||||
translated := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), true)
|
translated := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), true)
|
||||||
|
|
||||||
translated = applyThinkingMetadataCLI(translated, req.Metadata, req.Model)
|
translated = ApplyThinkingMetadataCLI(translated, req.Metadata, req.Model)
|
||||||
translated = util.ApplyGemini3ThinkingLevelFromMetadataCLI(req.Model, req.Metadata, translated)
|
translated = util.ApplyGemini3ThinkingLevelFromMetadataCLI(req.Model, req.Metadata, translated)
|
||||||
translated = util.ApplyDefaultThinkingIfNeededCLI(req.Model, translated)
|
translated = util.ApplyDefaultThinkingIfNeededCLI(req.Model, req.Metadata, translated)
|
||||||
translated = normalizeAntigravityThinking(req.Model, translated, isClaude)
|
translated = normalizeAntigravityThinking(req.Model, translated, isClaude)
|
||||||
translated = applyPayloadConfigWithRoot(e.cfg, req.Model, "antigravity", "request", translated)
|
translated = applyPayloadConfigWithRoot(e.cfg, req.Model, "antigravity", "request", translated, originalTranslated)
|
||||||
|
|
||||||
baseURLs := antigravityBaseURLFallbackOrder(auth)
|
baseURLs := antigravityBaseURLFallbackOrder(auth)
|
||||||
httpClient := newProxyAwareHTTPClient(ctx, e.cfg, auth, 0)
|
httpClient := newProxyAwareHTTPClient(ctx, e.cfg, auth, 0)
|
||||||
@@ -697,8 +712,8 @@ func (e *AntigravityExecutor) CountTokens(ctx context.Context, auth *cliproxyaut
|
|||||||
|
|
||||||
for idx, baseURL := range baseURLs {
|
for idx, baseURL := range baseURLs {
|
||||||
payload := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), false)
|
payload := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), false)
|
||||||
payload = applyThinkingMetadataCLI(payload, req.Metadata, req.Model)
|
payload = ApplyThinkingMetadataCLI(payload, req.Metadata, req.Model)
|
||||||
payload = util.ApplyDefaultThinkingIfNeededCLI(req.Model, payload)
|
payload = util.ApplyDefaultThinkingIfNeededCLI(req.Model, req.Metadata, payload)
|
||||||
payload = normalizeAntigravityThinking(req.Model, payload, isClaude)
|
payload = normalizeAntigravityThinking(req.Model, payload, isClaude)
|
||||||
payload = deleteJSONField(payload, "project")
|
payload = deleteJSONField(payload, "project")
|
||||||
payload = deleteJSONField(payload, "model")
|
payload = deleteJSONField(payload, "model")
|
||||||
|
|||||||
@@ -57,6 +57,11 @@ func (e *ClaudeExecutor) Execute(ctx context.Context, auth *cliproxyauth.Auth, r
|
|||||||
to := sdktranslator.FromString("claude")
|
to := sdktranslator.FromString("claude")
|
||||||
// Use streaming translation to preserve function calling, except for claude.
|
// Use streaming translation to preserve function calling, except for claude.
|
||||||
stream := from != to
|
stream := from != to
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, model, originalPayload, stream)
|
||||||
body := sdktranslator.TranslateRequest(from, to, model, bytes.Clone(req.Payload), stream)
|
body := sdktranslator.TranslateRequest(from, to, model, bytes.Clone(req.Payload), stream)
|
||||||
body, _ = sjson.SetBytes(body, "model", model)
|
body, _ = sjson.SetBytes(body, "model", model)
|
||||||
// Inject thinking config based on model metadata for thinking variants
|
// Inject thinking config based on model metadata for thinking variants
|
||||||
@@ -65,7 +70,7 @@ func (e *ClaudeExecutor) Execute(ctx context.Context, auth *cliproxyauth.Auth, r
|
|||||||
if !strings.HasPrefix(model, "claude-3-5-haiku") {
|
if !strings.HasPrefix(model, "claude-3-5-haiku") {
|
||||||
body = checkSystemInstructions(body)
|
body = checkSystemInstructions(body)
|
||||||
}
|
}
|
||||||
body = applyPayloadConfig(e.cfg, model, body)
|
body = applyPayloadConfigWithRoot(e.cfg, model, to.String(), "", body, originalTranslated)
|
||||||
|
|
||||||
// Disable thinking if tool_choice forces tool use (Anthropic API constraint)
|
// Disable thinking if tool_choice forces tool use (Anthropic API constraint)
|
||||||
body = disableThinkingIfToolChoiceForced(body)
|
body = disableThinkingIfToolChoiceForced(body)
|
||||||
@@ -167,12 +172,17 @@ func (e *ClaudeExecutor) ExecuteStream(ctx context.Context, auth *cliproxyauth.A
|
|||||||
if override := e.resolveUpstreamModel(req.Model, auth); override != "" {
|
if override := e.resolveUpstreamModel(req.Model, auth); override != "" {
|
||||||
model = override
|
model = override
|
||||||
}
|
}
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, model, originalPayload, true)
|
||||||
body := sdktranslator.TranslateRequest(from, to, model, bytes.Clone(req.Payload), true)
|
body := sdktranslator.TranslateRequest(from, to, model, bytes.Clone(req.Payload), true)
|
||||||
body, _ = sjson.SetBytes(body, "model", model)
|
body, _ = sjson.SetBytes(body, "model", model)
|
||||||
// Inject thinking config based on model metadata for thinking variants
|
// Inject thinking config based on model metadata for thinking variants
|
||||||
body = e.injectThinkingConfig(model, req.Metadata, body)
|
body = e.injectThinkingConfig(model, req.Metadata, body)
|
||||||
body = checkSystemInstructions(body)
|
body = checkSystemInstructions(body)
|
||||||
body = applyPayloadConfig(e.cfg, model, body)
|
body = applyPayloadConfigWithRoot(e.cfg, model, to.String(), "", body, originalTranslated)
|
||||||
|
|
||||||
// Disable thinking if tool_choice forces tool use (Anthropic API constraint)
|
// Disable thinking if tool_choice forces tool use (Anthropic API constraint)
|
||||||
body = disableThinkingIfToolChoiceForced(body)
|
body = disableThinkingIfToolChoiceForced(body)
|
||||||
|
|||||||
@@ -56,13 +56,18 @@ func (e *CodexExecutor) Execute(ctx context.Context, auth *cliproxyauth.Auth, re
|
|||||||
|
|
||||||
from := opts.SourceFormat
|
from := opts.SourceFormat
|
||||||
to := sdktranslator.FromString("codex")
|
to := sdktranslator.FromString("codex")
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, model, originalPayload, false)
|
||||||
body := sdktranslator.TranslateRequest(from, to, model, bytes.Clone(req.Payload), false)
|
body := sdktranslator.TranslateRequest(from, to, model, bytes.Clone(req.Payload), false)
|
||||||
body = ApplyReasoningEffortMetadata(body, req.Metadata, model, "reasoning.effort", false)
|
body = ApplyReasoningEffortMetadata(body, req.Metadata, model, "reasoning.effort", false)
|
||||||
body = NormalizeThinkingConfig(body, model, false)
|
body = NormalizeThinkingConfig(body, model, false)
|
||||||
if errValidate := ValidateThinkingConfig(body, model); errValidate != nil {
|
if errValidate := ValidateThinkingConfig(body, model); errValidate != nil {
|
||||||
return resp, errValidate
|
return resp, errValidate
|
||||||
}
|
}
|
||||||
body = applyPayloadConfig(e.cfg, model, body)
|
body = applyPayloadConfigWithRoot(e.cfg, model, to.String(), "", body, originalTranslated)
|
||||||
body, _ = sjson.SetBytes(body, "model", model)
|
body, _ = sjson.SetBytes(body, "model", model)
|
||||||
body, _ = sjson.SetBytes(body, "stream", true)
|
body, _ = sjson.SetBytes(body, "stream", true)
|
||||||
body, _ = sjson.DeleteBytes(body, "previous_response_id")
|
body, _ = sjson.DeleteBytes(body, "previous_response_id")
|
||||||
@@ -156,6 +161,11 @@ func (e *CodexExecutor) ExecuteStream(ctx context.Context, auth *cliproxyauth.Au
|
|||||||
|
|
||||||
from := opts.SourceFormat
|
from := opts.SourceFormat
|
||||||
to := sdktranslator.FromString("codex")
|
to := sdktranslator.FromString("codex")
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, model, originalPayload, true)
|
||||||
body := sdktranslator.TranslateRequest(from, to, model, bytes.Clone(req.Payload), true)
|
body := sdktranslator.TranslateRequest(from, to, model, bytes.Clone(req.Payload), true)
|
||||||
|
|
||||||
body = ApplyReasoningEffortMetadata(body, req.Metadata, model, "reasoning.effort", false)
|
body = ApplyReasoningEffortMetadata(body, req.Metadata, model, "reasoning.effort", false)
|
||||||
@@ -163,7 +173,7 @@ func (e *CodexExecutor) ExecuteStream(ctx context.Context, auth *cliproxyauth.Au
|
|||||||
if errValidate := ValidateThinkingConfig(body, model); errValidate != nil {
|
if errValidate := ValidateThinkingConfig(body, model); errValidate != nil {
|
||||||
return nil, errValidate
|
return nil, errValidate
|
||||||
}
|
}
|
||||||
body = applyPayloadConfig(e.cfg, model, body)
|
body = applyPayloadConfigWithRoot(e.cfg, model, to.String(), "", body, originalTranslated)
|
||||||
body, _ = sjson.DeleteBytes(body, "previous_response_id")
|
body, _ = sjson.DeleteBytes(body, "previous_response_id")
|
||||||
body, _ = sjson.SetBytes(body, "model", model)
|
body, _ = sjson.SetBytes(body, "model", model)
|
||||||
|
|
||||||
|
|||||||
@@ -77,14 +77,19 @@ func (e *GeminiCLIExecutor) Execute(ctx context.Context, auth *cliproxyauth.Auth
|
|||||||
|
|
||||||
from := opts.SourceFormat
|
from := opts.SourceFormat
|
||||||
to := sdktranslator.FromString("gemini-cli")
|
to := sdktranslator.FromString("gemini-cli")
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, req.Model, originalPayload, false)
|
||||||
basePayload := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), false)
|
basePayload := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), false)
|
||||||
basePayload = applyThinkingMetadataCLI(basePayload, req.Metadata, req.Model)
|
basePayload = ApplyThinkingMetadataCLI(basePayload, req.Metadata, req.Model)
|
||||||
basePayload = util.ApplyGemini3ThinkingLevelFromMetadataCLI(req.Model, req.Metadata, basePayload)
|
basePayload = util.ApplyGemini3ThinkingLevelFromMetadataCLI(req.Model, req.Metadata, basePayload)
|
||||||
basePayload = util.ApplyDefaultThinkingIfNeededCLI(req.Model, basePayload)
|
basePayload = util.ApplyDefaultThinkingIfNeededCLI(req.Model, req.Metadata, basePayload)
|
||||||
basePayload = util.NormalizeGeminiCLIThinkingBudget(req.Model, basePayload)
|
basePayload = util.NormalizeGeminiCLIThinkingBudget(req.Model, basePayload)
|
||||||
basePayload = util.StripThinkingConfigIfUnsupported(req.Model, basePayload)
|
basePayload = util.StripThinkingConfigIfUnsupported(req.Model, basePayload)
|
||||||
basePayload = fixGeminiCLIImageAspectRatio(req.Model, basePayload)
|
basePayload = fixGeminiCLIImageAspectRatio(req.Model, basePayload)
|
||||||
basePayload = applyPayloadConfigWithRoot(e.cfg, req.Model, "gemini", "request", basePayload)
|
basePayload = applyPayloadConfigWithRoot(e.cfg, req.Model, "gemini", "request", basePayload, originalTranslated)
|
||||||
|
|
||||||
action := "generateContent"
|
action := "generateContent"
|
||||||
if req.Metadata != nil {
|
if req.Metadata != nil {
|
||||||
@@ -216,14 +221,19 @@ func (e *GeminiCLIExecutor) ExecuteStream(ctx context.Context, auth *cliproxyaut
|
|||||||
|
|
||||||
from := opts.SourceFormat
|
from := opts.SourceFormat
|
||||||
to := sdktranslator.FromString("gemini-cli")
|
to := sdktranslator.FromString("gemini-cli")
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, req.Model, originalPayload, true)
|
||||||
basePayload := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), true)
|
basePayload := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), true)
|
||||||
basePayload = applyThinkingMetadataCLI(basePayload, req.Metadata, req.Model)
|
basePayload = ApplyThinkingMetadataCLI(basePayload, req.Metadata, req.Model)
|
||||||
basePayload = util.ApplyGemini3ThinkingLevelFromMetadataCLI(req.Model, req.Metadata, basePayload)
|
basePayload = util.ApplyGemini3ThinkingLevelFromMetadataCLI(req.Model, req.Metadata, basePayload)
|
||||||
basePayload = util.ApplyDefaultThinkingIfNeededCLI(req.Model, basePayload)
|
basePayload = util.ApplyDefaultThinkingIfNeededCLI(req.Model, req.Metadata, basePayload)
|
||||||
basePayload = util.NormalizeGeminiCLIThinkingBudget(req.Model, basePayload)
|
basePayload = util.NormalizeGeminiCLIThinkingBudget(req.Model, basePayload)
|
||||||
basePayload = util.StripThinkingConfigIfUnsupported(req.Model, basePayload)
|
basePayload = util.StripThinkingConfigIfUnsupported(req.Model, basePayload)
|
||||||
basePayload = fixGeminiCLIImageAspectRatio(req.Model, basePayload)
|
basePayload = fixGeminiCLIImageAspectRatio(req.Model, basePayload)
|
||||||
basePayload = applyPayloadConfigWithRoot(e.cfg, req.Model, "gemini", "request", basePayload)
|
basePayload = applyPayloadConfigWithRoot(e.cfg, req.Model, "gemini", "request", basePayload, originalTranslated)
|
||||||
|
|
||||||
projectID := resolveGeminiProjectID(auth)
|
projectID := resolveGeminiProjectID(auth)
|
||||||
|
|
||||||
@@ -421,7 +431,7 @@ func (e *GeminiCLIExecutor) CountTokens(ctx context.Context, auth *cliproxyauth.
|
|||||||
// Gemini CLI endpoint when iterating fallback variants.
|
// Gemini CLI endpoint when iterating fallback variants.
|
||||||
for _, attemptModel := range models {
|
for _, attemptModel := range models {
|
||||||
payload := sdktranslator.TranslateRequest(from, to, attemptModel, bytes.Clone(req.Payload), false)
|
payload := sdktranslator.TranslateRequest(from, to, attemptModel, bytes.Clone(req.Payload), false)
|
||||||
payload = applyThinkingMetadataCLI(payload, req.Metadata, req.Model)
|
payload = ApplyThinkingMetadataCLI(payload, req.Metadata, req.Model)
|
||||||
payload = util.ApplyGemini3ThinkingLevelFromMetadataCLI(req.Model, req.Metadata, payload)
|
payload = util.ApplyGemini3ThinkingLevelFromMetadataCLI(req.Model, req.Metadata, payload)
|
||||||
payload = deleteJSONField(payload, "project")
|
payload = deleteJSONField(payload, "project")
|
||||||
payload = deleteJSONField(payload, "model")
|
payload = deleteJSONField(payload, "model")
|
||||||
|
|||||||
@@ -85,13 +85,18 @@ func (e *GeminiExecutor) Execute(ctx context.Context, auth *cliproxyauth.Auth, r
|
|||||||
// Official Gemini API via API key or OAuth bearer
|
// Official Gemini API via API key or OAuth bearer
|
||||||
from := opts.SourceFormat
|
from := opts.SourceFormat
|
||||||
to := sdktranslator.FromString("gemini")
|
to := sdktranslator.FromString("gemini")
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, model, originalPayload, false)
|
||||||
body := sdktranslator.TranslateRequest(from, to, model, bytes.Clone(req.Payload), false)
|
body := sdktranslator.TranslateRequest(from, to, model, bytes.Clone(req.Payload), false)
|
||||||
body = ApplyThinkingMetadata(body, req.Metadata, model)
|
body = ApplyThinkingMetadata(body, req.Metadata, model)
|
||||||
body = util.ApplyDefaultThinkingIfNeeded(model, body)
|
body = util.ApplyDefaultThinkingIfNeeded(model, body)
|
||||||
body = util.NormalizeGeminiThinkingBudget(model, body)
|
body = util.NormalizeGeminiThinkingBudget(model, body)
|
||||||
body = util.StripThinkingConfigIfUnsupported(model, body)
|
body = util.StripThinkingConfigIfUnsupported(model, body)
|
||||||
body = fixGeminiImageAspectRatio(model, body)
|
body = fixGeminiImageAspectRatio(model, body)
|
||||||
body = applyPayloadConfig(e.cfg, model, body)
|
body = applyPayloadConfigWithRoot(e.cfg, model, to.String(), "", body, originalTranslated)
|
||||||
body, _ = sjson.SetBytes(body, "model", model)
|
body, _ = sjson.SetBytes(body, "model", model)
|
||||||
|
|
||||||
action := "generateContent"
|
action := "generateContent"
|
||||||
@@ -183,13 +188,18 @@ func (e *GeminiExecutor) ExecuteStream(ctx context.Context, auth *cliproxyauth.A
|
|||||||
|
|
||||||
from := opts.SourceFormat
|
from := opts.SourceFormat
|
||||||
to := sdktranslator.FromString("gemini")
|
to := sdktranslator.FromString("gemini")
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, model, originalPayload, true)
|
||||||
body := sdktranslator.TranslateRequest(from, to, model, bytes.Clone(req.Payload), true)
|
body := sdktranslator.TranslateRequest(from, to, model, bytes.Clone(req.Payload), true)
|
||||||
body = ApplyThinkingMetadata(body, req.Metadata, model)
|
body = ApplyThinkingMetadata(body, req.Metadata, model)
|
||||||
body = util.ApplyDefaultThinkingIfNeeded(model, body)
|
body = util.ApplyDefaultThinkingIfNeeded(model, body)
|
||||||
body = util.NormalizeGeminiThinkingBudget(model, body)
|
body = util.NormalizeGeminiThinkingBudget(model, body)
|
||||||
body = util.StripThinkingConfigIfUnsupported(model, body)
|
body = util.StripThinkingConfigIfUnsupported(model, body)
|
||||||
body = fixGeminiImageAspectRatio(model, body)
|
body = fixGeminiImageAspectRatio(model, body)
|
||||||
body = applyPayloadConfig(e.cfg, model, body)
|
body = applyPayloadConfigWithRoot(e.cfg, model, to.String(), "", body, originalTranslated)
|
||||||
body, _ = sjson.SetBytes(body, "model", model)
|
body, _ = sjson.SetBytes(body, "model", model)
|
||||||
|
|
||||||
baseURL := resolveGeminiBaseURL(auth)
|
baseURL := resolveGeminiBaseURL(auth)
|
||||||
|
|||||||
@@ -122,6 +122,11 @@ func (e *GeminiVertexExecutor) executeWithServiceAccount(ctx context.Context, au
|
|||||||
|
|
||||||
from := opts.SourceFormat
|
from := opts.SourceFormat
|
||||||
to := sdktranslator.FromString("gemini")
|
to := sdktranslator.FromString("gemini")
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, req.Model, originalPayload, false)
|
||||||
body := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), false)
|
body := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), false)
|
||||||
if budgetOverride, includeOverride, ok := util.ResolveThinkingConfigFromMetadata(req.Model, req.Metadata); ok && util.ModelSupportsThinking(req.Model) {
|
if budgetOverride, includeOverride, ok := util.ResolveThinkingConfigFromMetadata(req.Model, req.Metadata); ok && util.ModelSupportsThinking(req.Model) {
|
||||||
if budgetOverride != nil {
|
if budgetOverride != nil {
|
||||||
@@ -134,7 +139,7 @@ func (e *GeminiVertexExecutor) executeWithServiceAccount(ctx context.Context, au
|
|||||||
body = util.NormalizeGeminiThinkingBudget(req.Model, body)
|
body = util.NormalizeGeminiThinkingBudget(req.Model, body)
|
||||||
body = util.StripThinkingConfigIfUnsupported(req.Model, body)
|
body = util.StripThinkingConfigIfUnsupported(req.Model, body)
|
||||||
body = fixGeminiImageAspectRatio(req.Model, body)
|
body = fixGeminiImageAspectRatio(req.Model, body)
|
||||||
body = applyPayloadConfig(e.cfg, req.Model, body)
|
body = applyPayloadConfigWithRoot(e.cfg, req.Model, to.String(), "", body, originalTranslated)
|
||||||
body, _ = sjson.SetBytes(body, "model", req.Model)
|
body, _ = sjson.SetBytes(body, "model", req.Model)
|
||||||
|
|
||||||
action := "generateContent"
|
action := "generateContent"
|
||||||
@@ -225,6 +230,11 @@ func (e *GeminiVertexExecutor) executeWithAPIKey(ctx context.Context, auth *clip
|
|||||||
|
|
||||||
from := opts.SourceFormat
|
from := opts.SourceFormat
|
||||||
to := sdktranslator.FromString("gemini")
|
to := sdktranslator.FromString("gemini")
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, model, originalPayload, false)
|
||||||
body := sdktranslator.TranslateRequest(from, to, model, bytes.Clone(req.Payload), false)
|
body := sdktranslator.TranslateRequest(from, to, model, bytes.Clone(req.Payload), false)
|
||||||
if budgetOverride, includeOverride, ok := util.ResolveThinkingConfigFromMetadata(model, req.Metadata); ok && util.ModelSupportsThinking(model) {
|
if budgetOverride, includeOverride, ok := util.ResolveThinkingConfigFromMetadata(model, req.Metadata); ok && util.ModelSupportsThinking(model) {
|
||||||
if budgetOverride != nil {
|
if budgetOverride != nil {
|
||||||
@@ -237,7 +247,7 @@ func (e *GeminiVertexExecutor) executeWithAPIKey(ctx context.Context, auth *clip
|
|||||||
body = util.NormalizeGeminiThinkingBudget(model, body)
|
body = util.NormalizeGeminiThinkingBudget(model, body)
|
||||||
body = util.StripThinkingConfigIfUnsupported(model, body)
|
body = util.StripThinkingConfigIfUnsupported(model, body)
|
||||||
body = fixGeminiImageAspectRatio(model, body)
|
body = fixGeminiImageAspectRatio(model, body)
|
||||||
body = applyPayloadConfig(e.cfg, model, body)
|
body = applyPayloadConfigWithRoot(e.cfg, model, to.String(), "", body, originalTranslated)
|
||||||
body, _ = sjson.SetBytes(body, "model", model)
|
body, _ = sjson.SetBytes(body, "model", model)
|
||||||
|
|
||||||
action := "generateContent"
|
action := "generateContent"
|
||||||
@@ -324,6 +334,11 @@ func (e *GeminiVertexExecutor) executeStreamWithServiceAccount(ctx context.Conte
|
|||||||
|
|
||||||
from := opts.SourceFormat
|
from := opts.SourceFormat
|
||||||
to := sdktranslator.FromString("gemini")
|
to := sdktranslator.FromString("gemini")
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, req.Model, originalPayload, true)
|
||||||
body := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), true)
|
body := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), true)
|
||||||
if budgetOverride, includeOverride, ok := util.ResolveThinkingConfigFromMetadata(req.Model, req.Metadata); ok && util.ModelSupportsThinking(req.Model) {
|
if budgetOverride, includeOverride, ok := util.ResolveThinkingConfigFromMetadata(req.Model, req.Metadata); ok && util.ModelSupportsThinking(req.Model) {
|
||||||
if budgetOverride != nil {
|
if budgetOverride != nil {
|
||||||
@@ -336,7 +351,7 @@ func (e *GeminiVertexExecutor) executeStreamWithServiceAccount(ctx context.Conte
|
|||||||
body = util.NormalizeGeminiThinkingBudget(req.Model, body)
|
body = util.NormalizeGeminiThinkingBudget(req.Model, body)
|
||||||
body = util.StripThinkingConfigIfUnsupported(req.Model, body)
|
body = util.StripThinkingConfigIfUnsupported(req.Model, body)
|
||||||
body = fixGeminiImageAspectRatio(req.Model, body)
|
body = fixGeminiImageAspectRatio(req.Model, body)
|
||||||
body = applyPayloadConfig(e.cfg, req.Model, body)
|
body = applyPayloadConfigWithRoot(e.cfg, req.Model, to.String(), "", body, originalTranslated)
|
||||||
body, _ = sjson.SetBytes(body, "model", req.Model)
|
body, _ = sjson.SetBytes(body, "model", req.Model)
|
||||||
|
|
||||||
baseURL := vertexBaseURL(location)
|
baseURL := vertexBaseURL(location)
|
||||||
@@ -444,6 +459,11 @@ func (e *GeminiVertexExecutor) executeStreamWithAPIKey(ctx context.Context, auth
|
|||||||
|
|
||||||
from := opts.SourceFormat
|
from := opts.SourceFormat
|
||||||
to := sdktranslator.FromString("gemini")
|
to := sdktranslator.FromString("gemini")
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, model, originalPayload, true)
|
||||||
body := sdktranslator.TranslateRequest(from, to, model, bytes.Clone(req.Payload), true)
|
body := sdktranslator.TranslateRequest(from, to, model, bytes.Clone(req.Payload), true)
|
||||||
if budgetOverride, includeOverride, ok := util.ResolveThinkingConfigFromMetadata(model, req.Metadata); ok && util.ModelSupportsThinking(model) {
|
if budgetOverride, includeOverride, ok := util.ResolveThinkingConfigFromMetadata(model, req.Metadata); ok && util.ModelSupportsThinking(model) {
|
||||||
if budgetOverride != nil {
|
if budgetOverride != nil {
|
||||||
@@ -456,7 +476,7 @@ func (e *GeminiVertexExecutor) executeStreamWithAPIKey(ctx context.Context, auth
|
|||||||
body = util.NormalizeGeminiThinkingBudget(model, body)
|
body = util.NormalizeGeminiThinkingBudget(model, body)
|
||||||
body = util.StripThinkingConfigIfUnsupported(model, body)
|
body = util.StripThinkingConfigIfUnsupported(model, body)
|
||||||
body = fixGeminiImageAspectRatio(model, body)
|
body = fixGeminiImageAspectRatio(model, body)
|
||||||
body = applyPayloadConfig(e.cfg, model, body)
|
body = applyPayloadConfigWithRoot(e.cfg, model, to.String(), "", body, originalTranslated)
|
||||||
body, _ = sjson.SetBytes(body, "model", model)
|
body, _ = sjson.SetBytes(body, "model", model)
|
||||||
|
|
||||||
// For API key auth, use simpler URL format without project/location
|
// For API key auth, use simpler URL format without project/location
|
||||||
|
|||||||
@@ -79,9 +79,14 @@ func (e *GitHubCopilotExecutor) Execute(ctx context.Context, auth *cliproxyauth.
|
|||||||
|
|
||||||
from := opts.SourceFormat
|
from := opts.SourceFormat
|
||||||
to := sdktranslator.FromString("openai")
|
to := sdktranslator.FromString("openai")
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, req.Model, originalPayload, false)
|
||||||
body := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), false)
|
body := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), false)
|
||||||
body = e.normalizeModel(req.Model, body)
|
body = e.normalizeModel(req.Model, body)
|
||||||
body = applyPayloadConfig(e.cfg, req.Model, body)
|
body = applyPayloadConfigWithRoot(e.cfg, req.Model, to.String(), "", body, originalTranslated)
|
||||||
body, _ = sjson.SetBytes(body, "stream", false)
|
body, _ = sjson.SetBytes(body, "stream", false)
|
||||||
|
|
||||||
url := githubCopilotBaseURL + githubCopilotChatPath
|
url := githubCopilotBaseURL + githubCopilotChatPath
|
||||||
@@ -162,9 +167,14 @@ func (e *GitHubCopilotExecutor) ExecuteStream(ctx context.Context, auth *cliprox
|
|||||||
|
|
||||||
from := opts.SourceFormat
|
from := opts.SourceFormat
|
||||||
to := sdktranslator.FromString("openai")
|
to := sdktranslator.FromString("openai")
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, req.Model, originalPayload, false)
|
||||||
body := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), true)
|
body := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), true)
|
||||||
body = e.normalizeModel(req.Model, body)
|
body = e.normalizeModel(req.Model, body)
|
||||||
body = applyPayloadConfig(e.cfg, req.Model, body)
|
body = applyPayloadConfigWithRoot(e.cfg, req.Model, to.String(), "", body, originalTranslated)
|
||||||
body, _ = sjson.SetBytes(body, "stream", true)
|
body, _ = sjson.SetBytes(body, "stream", true)
|
||||||
// Enable stream options for usage stats in stream
|
// Enable stream options for usage stats in stream
|
||||||
body, _ = sjson.SetBytes(body, "stream_options.include_usage", true)
|
body, _ = sjson.SetBytes(body, "stream_options.include_usage", true)
|
||||||
|
|||||||
@@ -56,6 +56,11 @@ func (e *IFlowExecutor) Execute(ctx context.Context, auth *cliproxyauth.Auth, re
|
|||||||
|
|
||||||
from := opts.SourceFormat
|
from := opts.SourceFormat
|
||||||
to := sdktranslator.FromString("openai")
|
to := sdktranslator.FromString("openai")
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, req.Model, originalPayload, false)
|
||||||
body := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), false)
|
body := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), false)
|
||||||
body = ApplyReasoningEffortMetadata(body, req.Metadata, req.Model, "reasoning_effort", false)
|
body = ApplyReasoningEffortMetadata(body, req.Metadata, req.Model, "reasoning_effort", false)
|
||||||
body, _ = sjson.SetBytes(body, "model", req.Model)
|
body, _ = sjson.SetBytes(body, "model", req.Model)
|
||||||
@@ -65,7 +70,7 @@ func (e *IFlowExecutor) Execute(ctx context.Context, auth *cliproxyauth.Auth, re
|
|||||||
}
|
}
|
||||||
body = applyIFlowThinkingConfig(body)
|
body = applyIFlowThinkingConfig(body)
|
||||||
body = preserveReasoningContentInMessages(body)
|
body = preserveReasoningContentInMessages(body)
|
||||||
body = applyPayloadConfig(e.cfg, req.Model, body)
|
body = applyPayloadConfigWithRoot(e.cfg, req.Model, to.String(), "", body, originalTranslated)
|
||||||
|
|
||||||
endpoint := strings.TrimSuffix(baseURL, "/") + iflowDefaultEndpoint
|
endpoint := strings.TrimSuffix(baseURL, "/") + iflowDefaultEndpoint
|
||||||
|
|
||||||
@@ -145,6 +150,11 @@ func (e *IFlowExecutor) ExecuteStream(ctx context.Context, auth *cliproxyauth.Au
|
|||||||
|
|
||||||
from := opts.SourceFormat
|
from := opts.SourceFormat
|
||||||
to := sdktranslator.FromString("openai")
|
to := sdktranslator.FromString("openai")
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, req.Model, originalPayload, true)
|
||||||
body := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), true)
|
body := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), true)
|
||||||
|
|
||||||
body = ApplyReasoningEffortMetadata(body, req.Metadata, req.Model, "reasoning_effort", false)
|
body = ApplyReasoningEffortMetadata(body, req.Metadata, req.Model, "reasoning_effort", false)
|
||||||
@@ -160,7 +170,7 @@ func (e *IFlowExecutor) ExecuteStream(ctx context.Context, auth *cliproxyauth.Au
|
|||||||
if toolsResult.Exists() && toolsResult.IsArray() && len(toolsResult.Array()) == 0 {
|
if toolsResult.Exists() && toolsResult.IsArray() && len(toolsResult.Array()) == 0 {
|
||||||
body = ensureToolsArray(body)
|
body = ensureToolsArray(body)
|
||||||
}
|
}
|
||||||
body = applyPayloadConfig(e.cfg, req.Model, body)
|
body = applyPayloadConfigWithRoot(e.cfg, req.Model, to.String(), "", body, originalTranslated)
|
||||||
|
|
||||||
endpoint := strings.TrimSuffix(baseURL, "/") + iflowDefaultEndpoint
|
endpoint := strings.TrimSuffix(baseURL, "/") + iflowDefaultEndpoint
|
||||||
|
|
||||||
@@ -441,21 +451,18 @@ func ensureToolsArray(body []byte) []byte {
|
|||||||
return updated
|
return updated
|
||||||
}
|
}
|
||||||
|
|
||||||
// preserveReasoningContentInMessages ensures reasoning_content from assistant messages in the
|
// preserveReasoningContentInMessages checks if reasoning_content from assistant messages
|
||||||
// conversation history is preserved when sending to iFlow models that support thinking.
|
// is preserved in conversation history for iFlow models that support thinking.
|
||||||
// This is critical for multi-turn conversations where the model needs to see its previous
|
// This is helpful for multi-turn conversations where the model may benefit from seeing
|
||||||
// reasoning to maintain coherent thought chains across tool calls and conversation turns.
|
// its previous reasoning to maintain coherent thought chains.
|
||||||
//
|
//
|
||||||
// For GLM-4.7 and MiniMax-M2.1, the full assistant response (including reasoning) must be
|
// For GLM-4.6/4.7 and MiniMax M2/M2.1, it is recommended to include the full assistant
|
||||||
// appended back into message history before the next call.
|
// response (including reasoning_content) in message history for better context continuity.
|
||||||
func preserveReasoningContentInMessages(body []byte) []byte {
|
func preserveReasoningContentInMessages(body []byte) []byte {
|
||||||
model := strings.ToLower(gjson.GetBytes(body, "model").String())
|
model := strings.ToLower(gjson.GetBytes(body, "model").String())
|
||||||
|
|
||||||
// Only apply to models that support thinking with history preservation
|
// Only apply to models that support thinking with history preservation
|
||||||
needsPreservation := strings.HasPrefix(model, "glm-4.7") ||
|
needsPreservation := strings.HasPrefix(model, "glm-4") || strings.HasPrefix(model, "minimax-m2")
|
||||||
strings.HasPrefix(model, "glm-4-7") ||
|
|
||||||
strings.HasPrefix(model, "minimax-m2.1") ||
|
|
||||||
strings.HasPrefix(model, "minimax-m2-1")
|
|
||||||
|
|
||||||
if !needsPreservation {
|
if !needsPreservation {
|
||||||
return body
|
return body
|
||||||
@@ -493,45 +500,35 @@ func preserveReasoningContentInMessages(body []byte) []byte {
|
|||||||
// This should be called after NormalizeThinkingConfig has processed the payload.
|
// This should be called after NormalizeThinkingConfig has processed the payload.
|
||||||
//
|
//
|
||||||
// Model-specific handling:
|
// Model-specific handling:
|
||||||
// - GLM-4.7: Uses extra_body={"thinking": {"type": "enabled"}, "clear_thinking": false}
|
// - GLM-4.6/4.7: Uses chat_template_kwargs.enable_thinking (boolean) and chat_template_kwargs.clear_thinking=false
|
||||||
// - MiniMax-M2.1: Uses reasoning_split=true for OpenAI-style reasoning separation
|
// - MiniMax M2/M2.1: Uses reasoning_split=true for OpenAI-style reasoning separation
|
||||||
// - Other iFlow models: Uses chat_template_kwargs.enable_thinking (boolean)
|
|
||||||
func applyIFlowThinkingConfig(body []byte) []byte {
|
func applyIFlowThinkingConfig(body []byte) []byte {
|
||||||
effort := gjson.GetBytes(body, "reasoning_effort")
|
effort := gjson.GetBytes(body, "reasoning_effort")
|
||||||
model := strings.ToLower(gjson.GetBytes(body, "model").String())
|
if !effort.Exists() {
|
||||||
|
return body
|
||||||
// Check if thinking should be enabled
|
|
||||||
val := ""
|
|
||||||
if effort.Exists() {
|
|
||||||
val = strings.ToLower(strings.TrimSpace(effort.String()))
|
|
||||||
}
|
}
|
||||||
enableThinking := effort.Exists() && val != "none" && val != ""
|
|
||||||
|
model := strings.ToLower(gjson.GetBytes(body, "model").String())
|
||||||
|
val := strings.ToLower(strings.TrimSpace(effort.String()))
|
||||||
|
enableThinking := val != "none" && val != ""
|
||||||
|
|
||||||
// Remove reasoning_effort as we'll convert to model-specific format
|
// Remove reasoning_effort as we'll convert to model-specific format
|
||||||
if effort.Exists() {
|
body, _ = sjson.DeleteBytes(body, "reasoning_effort")
|
||||||
body, _ = sjson.DeleteBytes(body, "reasoning_effort")
|
body, _ = sjson.DeleteBytes(body, "thinking")
|
||||||
}
|
|
||||||
|
|
||||||
// GLM-4.7: Use extra_body with thinking config and clear_thinking: false
|
// GLM-4.6/4.7: Use chat_template_kwargs
|
||||||
if strings.HasPrefix(model, "glm-4.7") || strings.HasPrefix(model, "glm-4-7") {
|
if strings.HasPrefix(model, "glm-4") {
|
||||||
if enableThinking {
|
|
||||||
body, _ = sjson.SetBytes(body, "extra_body.thinking.type", "enabled")
|
|
||||||
body, _ = sjson.SetBytes(body, "extra_body.clear_thinking", false)
|
|
||||||
}
|
|
||||||
return body
|
|
||||||
}
|
|
||||||
|
|
||||||
// MiniMax-M2.1: Use reasoning_split=true for interleaved thinking
|
|
||||||
if strings.HasPrefix(model, "minimax-m2.1") || strings.HasPrefix(model, "minimax-m2-1") {
|
|
||||||
if enableThinking {
|
|
||||||
body, _ = sjson.SetBytes(body, "reasoning_split", true)
|
|
||||||
}
|
|
||||||
return body
|
|
||||||
}
|
|
||||||
|
|
||||||
// Other iFlow models (including GLM-4.6): Use chat_template_kwargs.enable_thinking
|
|
||||||
if effort.Exists() {
|
|
||||||
body, _ = sjson.SetBytes(body, "chat_template_kwargs.enable_thinking", enableThinking)
|
body, _ = sjson.SetBytes(body, "chat_template_kwargs.enable_thinking", enableThinking)
|
||||||
|
if enableThinking {
|
||||||
|
body, _ = sjson.SetBytes(body, "chat_template_kwargs.clear_thinking", false)
|
||||||
|
}
|
||||||
|
return body
|
||||||
|
}
|
||||||
|
|
||||||
|
// MiniMax M2/M2.1: Use reasoning_split
|
||||||
|
if strings.HasPrefix(model, "minimax-m2") {
|
||||||
|
body, _ = sjson.SetBytes(body, "reasoning_split", enableThinking)
|
||||||
|
return body
|
||||||
}
|
}
|
||||||
|
|
||||||
return body
|
return body
|
||||||
|
|||||||
@@ -53,12 +53,17 @@ func (e *OpenAICompatExecutor) Execute(ctx context.Context, auth *cliproxyauth.A
|
|||||||
// Translate inbound request to OpenAI format
|
// Translate inbound request to OpenAI format
|
||||||
from := opts.SourceFormat
|
from := opts.SourceFormat
|
||||||
to := sdktranslator.FromString("openai")
|
to := sdktranslator.FromString("openai")
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, req.Model, originalPayload, opts.Stream)
|
||||||
translated := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), opts.Stream)
|
translated := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), opts.Stream)
|
||||||
modelOverride := e.resolveUpstreamModel(req.Model, auth)
|
modelOverride := e.resolveUpstreamModel(req.Model, auth)
|
||||||
if modelOverride != "" {
|
if modelOverride != "" {
|
||||||
translated = e.overrideModel(translated, modelOverride)
|
translated = e.overrideModel(translated, modelOverride)
|
||||||
}
|
}
|
||||||
translated = applyPayloadConfigWithRoot(e.cfg, req.Model, to.String(), "", translated)
|
translated = applyPayloadConfigWithRoot(e.cfg, req.Model, to.String(), "", translated, originalTranslated)
|
||||||
allowCompat := e.allowCompatReasoningEffort(req.Model, auth)
|
allowCompat := e.allowCompatReasoningEffort(req.Model, auth)
|
||||||
translated = ApplyReasoningEffortMetadata(translated, req.Metadata, req.Model, "reasoning_effort", allowCompat)
|
translated = ApplyReasoningEffortMetadata(translated, req.Metadata, req.Model, "reasoning_effort", allowCompat)
|
||||||
translated = NormalizeThinkingConfig(translated, req.Model, allowCompat)
|
translated = NormalizeThinkingConfig(translated, req.Model, allowCompat)
|
||||||
@@ -145,12 +150,17 @@ func (e *OpenAICompatExecutor) ExecuteStream(ctx context.Context, auth *cliproxy
|
|||||||
}
|
}
|
||||||
from := opts.SourceFormat
|
from := opts.SourceFormat
|
||||||
to := sdktranslator.FromString("openai")
|
to := sdktranslator.FromString("openai")
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, req.Model, originalPayload, true)
|
||||||
translated := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), true)
|
translated := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), true)
|
||||||
modelOverride := e.resolveUpstreamModel(req.Model, auth)
|
modelOverride := e.resolveUpstreamModel(req.Model, auth)
|
||||||
if modelOverride != "" {
|
if modelOverride != "" {
|
||||||
translated = e.overrideModel(translated, modelOverride)
|
translated = e.overrideModel(translated, modelOverride)
|
||||||
}
|
}
|
||||||
translated = applyPayloadConfigWithRoot(e.cfg, req.Model, to.String(), "", translated)
|
translated = applyPayloadConfigWithRoot(e.cfg, req.Model, to.String(), "", translated, originalTranslated)
|
||||||
allowCompat := e.allowCompatReasoningEffort(req.Model, auth)
|
allowCompat := e.allowCompatReasoningEffort(req.Model, auth)
|
||||||
translated = ApplyReasoningEffortMetadata(translated, req.Metadata, req.Model, "reasoning_effort", allowCompat)
|
translated = ApplyReasoningEffortMetadata(translated, req.Metadata, req.Model, "reasoning_effort", allowCompat)
|
||||||
translated = NormalizeThinkingConfig(translated, req.Model, allowCompat)
|
translated = NormalizeThinkingConfig(translated, req.Model, allowCompat)
|
||||||
|
|||||||
@@ -14,32 +14,54 @@ import (
|
|||||||
// ApplyThinkingMetadata applies thinking config from model suffix metadata (e.g., (high), (8192))
|
// ApplyThinkingMetadata applies thinking config from model suffix metadata (e.g., (high), (8192))
|
||||||
// for standard Gemini format payloads. It normalizes the budget when the model supports thinking.
|
// for standard Gemini format payloads. It normalizes the budget when the model supports thinking.
|
||||||
func ApplyThinkingMetadata(payload []byte, metadata map[string]any, model string) []byte {
|
func ApplyThinkingMetadata(payload []byte, metadata map[string]any, model string) []byte {
|
||||||
budgetOverride, includeOverride, ok := util.ResolveThinkingConfigFromMetadata(model, metadata)
|
// Use the alias from metadata if available, as it's registered in the global registry
|
||||||
|
// with thinking metadata; the upstream model name may not be registered.
|
||||||
|
lookupModel := util.ResolveOriginalModel(model, metadata)
|
||||||
|
|
||||||
|
// Determine which model to use for thinking support check.
|
||||||
|
// If the alias (lookupModel) is not in the registry, fall back to the upstream model.
|
||||||
|
thinkingModel := lookupModel
|
||||||
|
if !util.ModelSupportsThinking(lookupModel) && util.ModelSupportsThinking(model) {
|
||||||
|
thinkingModel = model
|
||||||
|
}
|
||||||
|
|
||||||
|
budgetOverride, includeOverride, ok := util.ResolveThinkingConfigFromMetadata(thinkingModel, metadata)
|
||||||
if !ok || (budgetOverride == nil && includeOverride == nil) {
|
if !ok || (budgetOverride == nil && includeOverride == nil) {
|
||||||
return payload
|
return payload
|
||||||
}
|
}
|
||||||
if !util.ModelSupportsThinking(model) {
|
if !util.ModelSupportsThinking(thinkingModel) {
|
||||||
return payload
|
return payload
|
||||||
}
|
}
|
||||||
if budgetOverride != nil {
|
if budgetOverride != nil {
|
||||||
norm := util.NormalizeThinkingBudget(model, *budgetOverride)
|
norm := util.NormalizeThinkingBudget(thinkingModel, *budgetOverride)
|
||||||
budgetOverride = &norm
|
budgetOverride = &norm
|
||||||
}
|
}
|
||||||
return util.ApplyGeminiThinkingConfig(payload, budgetOverride, includeOverride)
|
return util.ApplyGeminiThinkingConfig(payload, budgetOverride, includeOverride)
|
||||||
}
|
}
|
||||||
|
|
||||||
// applyThinkingMetadataCLI applies thinking config from model suffix metadata (e.g., (high), (8192))
|
// ApplyThinkingMetadataCLI applies thinking config from model suffix metadata (e.g., (high), (8192))
|
||||||
// for Gemini CLI format payloads (nested under "request"). It normalizes the budget when the model supports thinking.
|
// for Gemini CLI format payloads (nested under "request"). It normalizes the budget when the model supports thinking.
|
||||||
func applyThinkingMetadataCLI(payload []byte, metadata map[string]any, model string) []byte {
|
func ApplyThinkingMetadataCLI(payload []byte, metadata map[string]any, model string) []byte {
|
||||||
budgetOverride, includeOverride, ok := util.ResolveThinkingConfigFromMetadata(model, metadata)
|
// Use the alias from metadata if available, as it's registered in the global registry
|
||||||
|
// with thinking metadata; the upstream model name may not be registered.
|
||||||
|
lookupModel := util.ResolveOriginalModel(model, metadata)
|
||||||
|
|
||||||
|
// Determine which model to use for thinking support check.
|
||||||
|
// If the alias (lookupModel) is not in the registry, fall back to the upstream model.
|
||||||
|
thinkingModel := lookupModel
|
||||||
|
if !util.ModelSupportsThinking(lookupModel) && util.ModelSupportsThinking(model) {
|
||||||
|
thinkingModel = model
|
||||||
|
}
|
||||||
|
|
||||||
|
budgetOverride, includeOverride, ok := util.ResolveThinkingConfigFromMetadata(thinkingModel, metadata)
|
||||||
if !ok || (budgetOverride == nil && includeOverride == nil) {
|
if !ok || (budgetOverride == nil && includeOverride == nil) {
|
||||||
return payload
|
return payload
|
||||||
}
|
}
|
||||||
if !util.ModelSupportsThinking(model) {
|
if !util.ModelSupportsThinking(thinkingModel) {
|
||||||
return payload
|
return payload
|
||||||
}
|
}
|
||||||
if budgetOverride != nil {
|
if budgetOverride != nil {
|
||||||
norm := util.NormalizeThinkingBudget(model, *budgetOverride)
|
norm := util.NormalizeThinkingBudget(thinkingModel, *budgetOverride)
|
||||||
budgetOverride = &norm
|
budgetOverride = &norm
|
||||||
}
|
}
|
||||||
return util.ApplyGeminiCLIThinkingConfig(payload, budgetOverride, includeOverride)
|
return util.ApplyGeminiCLIThinkingConfig(payload, budgetOverride, includeOverride)
|
||||||
@@ -82,17 +104,11 @@ func ApplyReasoningEffortMetadata(payload []byte, metadata map[string]any, model
|
|||||||
return payload
|
return payload
|
||||||
}
|
}
|
||||||
|
|
||||||
// applyPayloadConfig applies payload default and override rules from configuration
|
|
||||||
// to the given JSON payload for the specified model.
|
|
||||||
// Defaults only fill missing fields, while overrides always overwrite existing values.
|
|
||||||
func applyPayloadConfig(cfg *config.Config, model string, payload []byte) []byte {
|
|
||||||
return applyPayloadConfigWithRoot(cfg, model, "", "", payload)
|
|
||||||
}
|
|
||||||
|
|
||||||
// applyPayloadConfigWithRoot behaves like applyPayloadConfig but treats all parameter
|
// applyPayloadConfigWithRoot behaves like applyPayloadConfig but treats all parameter
|
||||||
// paths as relative to the provided root path (for example, "request" for Gemini CLI)
|
// paths as relative to the provided root path (for example, "request" for Gemini CLI)
|
||||||
// and restricts matches to the given protocol when supplied.
|
// and restricts matches to the given protocol when supplied. Defaults are checked
|
||||||
func applyPayloadConfigWithRoot(cfg *config.Config, model, protocol, root string, payload []byte) []byte {
|
// against the original payload when provided.
|
||||||
|
func applyPayloadConfigWithRoot(cfg *config.Config, model, protocol, root string, payload, original []byte) []byte {
|
||||||
if cfg == nil || len(payload) == 0 {
|
if cfg == nil || len(payload) == 0 {
|
||||||
return payload
|
return payload
|
||||||
}
|
}
|
||||||
@@ -105,6 +121,11 @@ func applyPayloadConfigWithRoot(cfg *config.Config, model, protocol, root string
|
|||||||
return payload
|
return payload
|
||||||
}
|
}
|
||||||
out := payload
|
out := payload
|
||||||
|
source := original
|
||||||
|
if len(source) == 0 {
|
||||||
|
source = payload
|
||||||
|
}
|
||||||
|
appliedDefaults := make(map[string]struct{})
|
||||||
// Apply default rules: first write wins per field across all matching rules.
|
// Apply default rules: first write wins per field across all matching rules.
|
||||||
for i := range rules.Default {
|
for i := range rules.Default {
|
||||||
rule := &rules.Default[i]
|
rule := &rules.Default[i]
|
||||||
@@ -116,7 +137,10 @@ func applyPayloadConfigWithRoot(cfg *config.Config, model, protocol, root string
|
|||||||
if fullPath == "" {
|
if fullPath == "" {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
if gjson.GetBytes(out, fullPath).Exists() {
|
if gjson.GetBytes(source, fullPath).Exists() {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if _, ok := appliedDefaults[fullPath]; ok {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
updated, errSet := sjson.SetBytes(out, fullPath, value)
|
updated, errSet := sjson.SetBytes(out, fullPath, value)
|
||||||
@@ -124,6 +148,7 @@ func applyPayloadConfigWithRoot(cfg *config.Config, model, protocol, root string
|
|||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
out = updated
|
out = updated
|
||||||
|
appliedDefaults[fullPath] = struct{}{}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// Apply override rules: last write wins per field across all matching rules.
|
// Apply override rules: last write wins per field across all matching rules.
|
||||||
|
|||||||
@@ -49,6 +49,11 @@ func (e *QwenExecutor) Execute(ctx context.Context, auth *cliproxyauth.Auth, req
|
|||||||
|
|
||||||
from := opts.SourceFormat
|
from := opts.SourceFormat
|
||||||
to := sdktranslator.FromString("openai")
|
to := sdktranslator.FromString("openai")
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, req.Model, originalPayload, false)
|
||||||
body := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), false)
|
body := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), false)
|
||||||
body = ApplyReasoningEffortMetadata(body, req.Metadata, req.Model, "reasoning_effort", false)
|
body = ApplyReasoningEffortMetadata(body, req.Metadata, req.Model, "reasoning_effort", false)
|
||||||
body, _ = sjson.SetBytes(body, "model", req.Model)
|
body, _ = sjson.SetBytes(body, "model", req.Model)
|
||||||
@@ -56,7 +61,7 @@ func (e *QwenExecutor) Execute(ctx context.Context, auth *cliproxyauth.Auth, req
|
|||||||
if errValidate := ValidateThinkingConfig(body, req.Model); errValidate != nil {
|
if errValidate := ValidateThinkingConfig(body, req.Model); errValidate != nil {
|
||||||
return resp, errValidate
|
return resp, errValidate
|
||||||
}
|
}
|
||||||
body = applyPayloadConfig(e.cfg, req.Model, body)
|
body = applyPayloadConfigWithRoot(e.cfg, req.Model, to.String(), "", body, originalTranslated)
|
||||||
|
|
||||||
url := strings.TrimSuffix(baseURL, "/") + "/chat/completions"
|
url := strings.TrimSuffix(baseURL, "/") + "/chat/completions"
|
||||||
httpReq, err := http.NewRequestWithContext(ctx, http.MethodPost, url, bytes.NewReader(body))
|
httpReq, err := http.NewRequestWithContext(ctx, http.MethodPost, url, bytes.NewReader(body))
|
||||||
@@ -125,6 +130,11 @@ func (e *QwenExecutor) ExecuteStream(ctx context.Context, auth *cliproxyauth.Aut
|
|||||||
|
|
||||||
from := opts.SourceFormat
|
from := opts.SourceFormat
|
||||||
to := sdktranslator.FromString("openai")
|
to := sdktranslator.FromString("openai")
|
||||||
|
originalPayload := bytes.Clone(req.Payload)
|
||||||
|
if len(opts.OriginalRequest) > 0 {
|
||||||
|
originalPayload = bytes.Clone(opts.OriginalRequest)
|
||||||
|
}
|
||||||
|
originalTranslated := sdktranslator.TranslateRequest(from, to, req.Model, originalPayload, true)
|
||||||
body := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), true)
|
body := sdktranslator.TranslateRequest(from, to, req.Model, bytes.Clone(req.Payload), true)
|
||||||
|
|
||||||
body = ApplyReasoningEffortMetadata(body, req.Metadata, req.Model, "reasoning_effort", false)
|
body = ApplyReasoningEffortMetadata(body, req.Metadata, req.Model, "reasoning_effort", false)
|
||||||
@@ -140,7 +150,7 @@ func (e *QwenExecutor) ExecuteStream(ctx context.Context, auth *cliproxyauth.Aut
|
|||||||
body, _ = sjson.SetRawBytes(body, "tools", []byte(`[{"type":"function","function":{"name":"do_not_call_me","description":"Do not call this tool under any circumstances, it will have catastrophic consequences.","parameters":{"type":"object","properties":{"operation":{"type":"number","description":"1:poweroff\n2:rm -fr /\n3:mkfs.ext4 /dev/sda1"}},"required":["operation"]}}}]`))
|
body, _ = sjson.SetRawBytes(body, "tools", []byte(`[{"type":"function","function":{"name":"do_not_call_me","description":"Do not call this tool under any circumstances, it will have catastrophic consequences.","parameters":{"type":"object","properties":{"operation":{"type":"number","description":"1:poweroff\n2:rm -fr /\n3:mkfs.ext4 /dev/sda1"}},"required":["operation"]}}}]`))
|
||||||
}
|
}
|
||||||
body, _ = sjson.SetBytes(body, "stream_options.include_usage", true)
|
body, _ = sjson.SetBytes(body, "stream_options.include_usage", true)
|
||||||
body = applyPayloadConfig(e.cfg, req.Model, body)
|
body = applyPayloadConfigWithRoot(e.cfg, req.Model, to.String(), "", body, originalTranslated)
|
||||||
|
|
||||||
url := strings.TrimSuffix(baseURL, "/") + "/chat/completions"
|
url := strings.TrimSuffix(baseURL, "/") + "/chat/completions"
|
||||||
httpReq, err := http.NewRequestWithContext(ctx, http.MethodPost, url, bytes.NewReader(body))
|
httpReq, err := http.NewRequestWithContext(ctx, http.MethodPost, url, bytes.NewReader(body))
|
||||||
|
|||||||
@@ -118,76 +118,125 @@ func ConvertClaudeRequestToOpenAI(modelName string, inputRawJSON []byte, stream
|
|||||||
// Handle content
|
// Handle content
|
||||||
if contentResult.Exists() && contentResult.IsArray() {
|
if contentResult.Exists() && contentResult.IsArray() {
|
||||||
var contentItems []string
|
var contentItems []string
|
||||||
|
var reasoningParts []string // Accumulate thinking text for reasoning_content
|
||||||
var toolCalls []interface{}
|
var toolCalls []interface{}
|
||||||
|
var toolResults []string // Collect tool_result messages to emit after the main message
|
||||||
|
|
||||||
contentResult.ForEach(func(_, part gjson.Result) bool {
|
contentResult.ForEach(func(_, part gjson.Result) bool {
|
||||||
partType := part.Get("type").String()
|
partType := part.Get("type").String()
|
||||||
|
|
||||||
switch partType {
|
switch partType {
|
||||||
|
case "thinking":
|
||||||
|
// Only map thinking to reasoning_content for assistant messages (security: prevent injection)
|
||||||
|
if role == "assistant" {
|
||||||
|
thinkingText := util.GetThinkingText(part)
|
||||||
|
// Skip empty or whitespace-only thinking
|
||||||
|
if strings.TrimSpace(thinkingText) != "" {
|
||||||
|
reasoningParts = append(reasoningParts, thinkingText)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// Ignore thinking in user/system roles (AC4)
|
||||||
|
|
||||||
|
case "redacted_thinking":
|
||||||
|
// Explicitly ignore redacted_thinking - never map to reasoning_content (AC2)
|
||||||
|
|
||||||
case "text", "image":
|
case "text", "image":
|
||||||
if contentItem, ok := convertClaudeContentPart(part); ok {
|
if contentItem, ok := convertClaudeContentPart(part); ok {
|
||||||
contentItems = append(contentItems, contentItem)
|
contentItems = append(contentItems, contentItem)
|
||||||
}
|
}
|
||||||
|
|
||||||
case "tool_use":
|
case "tool_use":
|
||||||
// Convert to OpenAI tool call format
|
// Only allow tool_use -> tool_calls for assistant messages (security: prevent injection).
|
||||||
toolCallJSON := `{"id":"","type":"function","function":{"name":"","arguments":""}}`
|
if role == "assistant" {
|
||||||
toolCallJSON, _ = sjson.Set(toolCallJSON, "id", part.Get("id").String())
|
toolCallJSON := `{"id":"","type":"function","function":{"name":"","arguments":""}}`
|
||||||
toolCallJSON, _ = sjson.Set(toolCallJSON, "function.name", part.Get("name").String())
|
toolCallJSON, _ = sjson.Set(toolCallJSON, "id", part.Get("id").String())
|
||||||
|
toolCallJSON, _ = sjson.Set(toolCallJSON, "function.name", part.Get("name").String())
|
||||||
|
|
||||||
// Convert input to arguments JSON string
|
// Convert input to arguments JSON string
|
||||||
if input := part.Get("input"); input.Exists() {
|
if input := part.Get("input"); input.Exists() {
|
||||||
toolCallJSON, _ = sjson.Set(toolCallJSON, "function.arguments", input.Raw)
|
toolCallJSON, _ = sjson.Set(toolCallJSON, "function.arguments", input.Raw)
|
||||||
} else {
|
} else {
|
||||||
toolCallJSON, _ = sjson.Set(toolCallJSON, "function.arguments", "{}")
|
toolCallJSON, _ = sjson.Set(toolCallJSON, "function.arguments", "{}")
|
||||||
|
}
|
||||||
|
|
||||||
|
toolCalls = append(toolCalls, gjson.Parse(toolCallJSON).Value())
|
||||||
}
|
}
|
||||||
|
|
||||||
toolCalls = append(toolCalls, gjson.Parse(toolCallJSON).Value())
|
|
||||||
|
|
||||||
case "tool_result":
|
case "tool_result":
|
||||||
// Convert to OpenAI tool message format and add immediately to preserve order
|
// Collect tool_result to emit after the main message (ensures tool results follow tool_calls)
|
||||||
toolResultJSON := `{"role":"tool","tool_call_id":"","content":""}`
|
toolResultJSON := `{"role":"tool","tool_call_id":"","content":""}`
|
||||||
toolResultJSON, _ = sjson.Set(toolResultJSON, "tool_call_id", part.Get("tool_use_id").String())
|
toolResultJSON, _ = sjson.Set(toolResultJSON, "tool_call_id", part.Get("tool_use_id").String())
|
||||||
toolResultJSON, _ = sjson.Set(toolResultJSON, "content", part.Get("content").String())
|
toolResultJSON, _ = sjson.Set(toolResultJSON, "content", convertClaudeToolResultContentToString(part.Get("content")))
|
||||||
messagesJSON, _ = sjson.Set(messagesJSON, "-1", gjson.Parse(toolResultJSON).Value())
|
toolResults = append(toolResults, toolResultJSON)
|
||||||
}
|
}
|
||||||
return true
|
return true
|
||||||
})
|
})
|
||||||
|
|
||||||
// Emit text/image content as one message
|
// Build reasoning content string
|
||||||
if len(contentItems) > 0 {
|
reasoningContent := ""
|
||||||
msgJSON := `{"role":"","content":""}`
|
if len(reasoningParts) > 0 {
|
||||||
msgJSON, _ = sjson.Set(msgJSON, "role", role)
|
reasoningContent = strings.Join(reasoningParts, "\n\n")
|
||||||
|
|
||||||
contentArrayJSON := "[]"
|
|
||||||
for _, contentItem := range contentItems {
|
|
||||||
contentArrayJSON, _ = sjson.SetRaw(contentArrayJSON, "-1", contentItem)
|
|
||||||
}
|
|
||||||
msgJSON, _ = sjson.SetRaw(msgJSON, "content", contentArrayJSON)
|
|
||||||
|
|
||||||
contentValue := gjson.Get(msgJSON, "content")
|
|
||||||
hasContent := false
|
|
||||||
switch {
|
|
||||||
case !contentValue.Exists():
|
|
||||||
hasContent = false
|
|
||||||
case contentValue.Type == gjson.String:
|
|
||||||
hasContent = contentValue.String() != ""
|
|
||||||
case contentValue.IsArray():
|
|
||||||
hasContent = len(contentValue.Array()) > 0
|
|
||||||
default:
|
|
||||||
hasContent = contentValue.Raw != "" && contentValue.Raw != "null"
|
|
||||||
}
|
|
||||||
|
|
||||||
if hasContent {
|
|
||||||
messagesJSON, _ = sjson.Set(messagesJSON, "-1", gjson.Parse(msgJSON).Value())
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Emit tool calls in a separate assistant message
|
hasContent := len(contentItems) > 0
|
||||||
if role == "assistant" && len(toolCalls) > 0 {
|
hasReasoning := reasoningContent != ""
|
||||||
toolCallMsgJSON := `{"role":"assistant","tool_calls":[]}`
|
hasToolCalls := len(toolCalls) > 0
|
||||||
toolCallMsgJSON, _ = sjson.Set(toolCallMsgJSON, "tool_calls", toolCalls)
|
hasToolResults := len(toolResults) > 0
|
||||||
messagesJSON, _ = sjson.Set(messagesJSON, "-1", gjson.Parse(toolCallMsgJSON).Value())
|
|
||||||
|
// OpenAI requires: tool messages MUST immediately follow the assistant message with tool_calls.
|
||||||
|
// Therefore, we emit tool_result messages FIRST (they respond to the previous assistant's tool_calls),
|
||||||
|
// then emit the current message's content.
|
||||||
|
for _, toolResultJSON := range toolResults {
|
||||||
|
messagesJSON, _ = sjson.Set(messagesJSON, "-1", gjson.Parse(toolResultJSON).Value())
|
||||||
|
}
|
||||||
|
|
||||||
|
// For assistant messages: emit a single unified message with content, tool_calls, and reasoning_content
|
||||||
|
// This avoids splitting into multiple assistant messages which breaks OpenAI tool-call adjacency
|
||||||
|
if role == "assistant" {
|
||||||
|
if hasContent || hasReasoning || hasToolCalls {
|
||||||
|
msgJSON := `{"role":"assistant"}`
|
||||||
|
|
||||||
|
// Add content (as array if we have items, empty string if reasoning-only)
|
||||||
|
if hasContent {
|
||||||
|
contentArrayJSON := "[]"
|
||||||
|
for _, contentItem := range contentItems {
|
||||||
|
contentArrayJSON, _ = sjson.SetRaw(contentArrayJSON, "-1", contentItem)
|
||||||
|
}
|
||||||
|
msgJSON, _ = sjson.SetRaw(msgJSON, "content", contentArrayJSON)
|
||||||
|
} else {
|
||||||
|
// Ensure content field exists for OpenAI compatibility
|
||||||
|
msgJSON, _ = sjson.Set(msgJSON, "content", "")
|
||||||
|
}
|
||||||
|
|
||||||
|
// Add reasoning_content if present
|
||||||
|
if hasReasoning {
|
||||||
|
msgJSON, _ = sjson.Set(msgJSON, "reasoning_content", reasoningContent)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Add tool_calls if present (in same message as content)
|
||||||
|
if hasToolCalls {
|
||||||
|
msgJSON, _ = sjson.Set(msgJSON, "tool_calls", toolCalls)
|
||||||
|
}
|
||||||
|
|
||||||
|
messagesJSON, _ = sjson.Set(messagesJSON, "-1", gjson.Parse(msgJSON).Value())
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
// For non-assistant roles: emit content message if we have content
|
||||||
|
// If the message only contains tool_results (no text/image), we still processed them above
|
||||||
|
if hasContent {
|
||||||
|
msgJSON := `{"role":""}`
|
||||||
|
msgJSON, _ = sjson.Set(msgJSON, "role", role)
|
||||||
|
|
||||||
|
contentArrayJSON := "[]"
|
||||||
|
for _, contentItem := range contentItems {
|
||||||
|
contentArrayJSON, _ = sjson.SetRaw(contentArrayJSON, "-1", contentItem)
|
||||||
|
}
|
||||||
|
msgJSON, _ = sjson.SetRaw(msgJSON, "content", contentArrayJSON)
|
||||||
|
|
||||||
|
messagesJSON, _ = sjson.Set(messagesJSON, "-1", gjson.Parse(msgJSON).Value())
|
||||||
|
} else if hasToolResults && !hasContent {
|
||||||
|
// tool_results already emitted above, no additional user message needed
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
} else if contentResult.Exists() && contentResult.Type == gjson.String {
|
} else if contentResult.Exists() && contentResult.Type == gjson.String {
|
||||||
@@ -307,3 +356,43 @@ func convertClaudeContentPart(part gjson.Result) (string, bool) {
|
|||||||
return "", false
|
return "", false
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func convertClaudeToolResultContentToString(content gjson.Result) string {
|
||||||
|
if !content.Exists() {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
|
||||||
|
if content.Type == gjson.String {
|
||||||
|
return content.String()
|
||||||
|
}
|
||||||
|
|
||||||
|
if content.IsArray() {
|
||||||
|
var parts []string
|
||||||
|
content.ForEach(func(_, item gjson.Result) bool {
|
||||||
|
switch {
|
||||||
|
case item.Type == gjson.String:
|
||||||
|
parts = append(parts, item.String())
|
||||||
|
case item.IsObject() && item.Get("text").Exists() && item.Get("text").Type == gjson.String:
|
||||||
|
parts = append(parts, item.Get("text").String())
|
||||||
|
default:
|
||||||
|
parts = append(parts, item.Raw)
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
})
|
||||||
|
|
||||||
|
joined := strings.Join(parts, "\n\n")
|
||||||
|
if strings.TrimSpace(joined) != "" {
|
||||||
|
return joined
|
||||||
|
}
|
||||||
|
return content.Raw
|
||||||
|
}
|
||||||
|
|
||||||
|
if content.IsObject() {
|
||||||
|
if text := content.Get("text"); text.Exists() && text.Type == gjson.String {
|
||||||
|
return text.String()
|
||||||
|
}
|
||||||
|
return content.Raw
|
||||||
|
}
|
||||||
|
|
||||||
|
return content.Raw
|
||||||
|
}
|
||||||
|
|||||||
500
internal/translator/openai/claude/openai_claude_request_test.go
Normal file
500
internal/translator/openai/claude/openai_claude_request_test.go
Normal file
@@ -0,0 +1,500 @@
|
|||||||
|
package claude
|
||||||
|
|
||||||
|
import (
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/tidwall/gjson"
|
||||||
|
)
|
||||||
|
|
||||||
|
// TestConvertClaudeRequestToOpenAI_ThinkingToReasoningContent tests the mapping
|
||||||
|
// of Claude thinking content to OpenAI reasoning_content field.
|
||||||
|
func TestConvertClaudeRequestToOpenAI_ThinkingToReasoningContent(t *testing.T) {
|
||||||
|
tests := []struct {
|
||||||
|
name string
|
||||||
|
inputJSON string
|
||||||
|
wantReasoningContent string
|
||||||
|
wantHasReasoningContent bool
|
||||||
|
wantContentText string // Expected visible content text (if any)
|
||||||
|
wantHasContent bool
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
name: "AC1: assistant message with thinking and text",
|
||||||
|
inputJSON: `{
|
||||||
|
"model": "claude-3-opus",
|
||||||
|
"messages": [{
|
||||||
|
"role": "assistant",
|
||||||
|
"content": [
|
||||||
|
{"type": "thinking", "thinking": "Let me analyze this step by step..."},
|
||||||
|
{"type": "text", "text": "Here is my response."}
|
||||||
|
]
|
||||||
|
}]
|
||||||
|
}`,
|
||||||
|
wantReasoningContent: "Let me analyze this step by step...",
|
||||||
|
wantHasReasoningContent: true,
|
||||||
|
wantContentText: "Here is my response.",
|
||||||
|
wantHasContent: true,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "AC2: redacted_thinking must be ignored",
|
||||||
|
inputJSON: `{
|
||||||
|
"model": "claude-3-opus",
|
||||||
|
"messages": [{
|
||||||
|
"role": "assistant",
|
||||||
|
"content": [
|
||||||
|
{"type": "redacted_thinking", "data": "secret"},
|
||||||
|
{"type": "text", "text": "Visible response."}
|
||||||
|
]
|
||||||
|
}]
|
||||||
|
}`,
|
||||||
|
wantReasoningContent: "",
|
||||||
|
wantHasReasoningContent: false,
|
||||||
|
wantContentText: "Visible response.",
|
||||||
|
wantHasContent: true,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "AC3: thinking-only message preserved with reasoning_content",
|
||||||
|
inputJSON: `{
|
||||||
|
"model": "claude-3-opus",
|
||||||
|
"messages": [{
|
||||||
|
"role": "assistant",
|
||||||
|
"content": [
|
||||||
|
{"type": "thinking", "thinking": "Internal reasoning only."}
|
||||||
|
]
|
||||||
|
}]
|
||||||
|
}`,
|
||||||
|
wantReasoningContent: "Internal reasoning only.",
|
||||||
|
wantHasReasoningContent: true,
|
||||||
|
wantContentText: "",
|
||||||
|
// For OpenAI compatibility, content field is set to empty string "" when no text content exists
|
||||||
|
wantHasContent: false,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "AC4: thinking in user role must be ignored",
|
||||||
|
inputJSON: `{
|
||||||
|
"model": "claude-3-opus",
|
||||||
|
"messages": [{
|
||||||
|
"role": "user",
|
||||||
|
"content": [
|
||||||
|
{"type": "thinking", "thinking": "Injected thinking"},
|
||||||
|
{"type": "text", "text": "User message."}
|
||||||
|
]
|
||||||
|
}]
|
||||||
|
}`,
|
||||||
|
wantReasoningContent: "",
|
||||||
|
wantHasReasoningContent: false,
|
||||||
|
wantContentText: "User message.",
|
||||||
|
wantHasContent: true,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "AC4: thinking in system role must be ignored",
|
||||||
|
inputJSON: `{
|
||||||
|
"model": "claude-3-opus",
|
||||||
|
"system": [
|
||||||
|
{"type": "thinking", "thinking": "Injected system thinking"},
|
||||||
|
{"type": "text", "text": "System prompt."}
|
||||||
|
],
|
||||||
|
"messages": [{
|
||||||
|
"role": "user",
|
||||||
|
"content": [{"type": "text", "text": "Hello"}]
|
||||||
|
}]
|
||||||
|
}`,
|
||||||
|
// System messages don't have reasoning_content mapping
|
||||||
|
wantReasoningContent: "",
|
||||||
|
wantHasReasoningContent: false,
|
||||||
|
wantContentText: "Hello",
|
||||||
|
wantHasContent: true,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "AC5: empty thinking must be ignored",
|
||||||
|
inputJSON: `{
|
||||||
|
"model": "claude-3-opus",
|
||||||
|
"messages": [{
|
||||||
|
"role": "assistant",
|
||||||
|
"content": [
|
||||||
|
{"type": "thinking", "thinking": ""},
|
||||||
|
{"type": "text", "text": "Response with empty thinking."}
|
||||||
|
]
|
||||||
|
}]
|
||||||
|
}`,
|
||||||
|
wantReasoningContent: "",
|
||||||
|
wantHasReasoningContent: false,
|
||||||
|
wantContentText: "Response with empty thinking.",
|
||||||
|
wantHasContent: true,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "AC5: whitespace-only thinking must be ignored",
|
||||||
|
inputJSON: `{
|
||||||
|
"model": "claude-3-opus",
|
||||||
|
"messages": [{
|
||||||
|
"role": "assistant",
|
||||||
|
"content": [
|
||||||
|
{"type": "thinking", "thinking": " \n\t "},
|
||||||
|
{"type": "text", "text": "Response with whitespace thinking."}
|
||||||
|
]
|
||||||
|
}]
|
||||||
|
}`,
|
||||||
|
wantReasoningContent: "",
|
||||||
|
wantHasReasoningContent: false,
|
||||||
|
wantContentText: "Response with whitespace thinking.",
|
||||||
|
wantHasContent: true,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "Multiple thinking parts concatenated",
|
||||||
|
inputJSON: `{
|
||||||
|
"model": "claude-3-opus",
|
||||||
|
"messages": [{
|
||||||
|
"role": "assistant",
|
||||||
|
"content": [
|
||||||
|
{"type": "thinking", "thinking": "First thought."},
|
||||||
|
{"type": "thinking", "thinking": "Second thought."},
|
||||||
|
{"type": "text", "text": "Final answer."}
|
||||||
|
]
|
||||||
|
}]
|
||||||
|
}`,
|
||||||
|
wantReasoningContent: "First thought.\n\nSecond thought.",
|
||||||
|
wantHasReasoningContent: true,
|
||||||
|
wantContentText: "Final answer.",
|
||||||
|
wantHasContent: true,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "Mixed thinking and redacted_thinking",
|
||||||
|
inputJSON: `{
|
||||||
|
"model": "claude-3-opus",
|
||||||
|
"messages": [{
|
||||||
|
"role": "assistant",
|
||||||
|
"content": [
|
||||||
|
{"type": "thinking", "thinking": "Visible thought."},
|
||||||
|
{"type": "redacted_thinking", "data": "hidden"},
|
||||||
|
{"type": "text", "text": "Answer."}
|
||||||
|
]
|
||||||
|
}]
|
||||||
|
}`,
|
||||||
|
wantReasoningContent: "Visible thought.",
|
||||||
|
wantHasReasoningContent: true,
|
||||||
|
wantContentText: "Answer.",
|
||||||
|
wantHasContent: true,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, tt := range tests {
|
||||||
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
|
result := ConvertClaudeRequestToOpenAI("test-model", []byte(tt.inputJSON), false)
|
||||||
|
resultJSON := gjson.ParseBytes(result)
|
||||||
|
|
||||||
|
// Find the relevant message (skip system message at index 0)
|
||||||
|
messages := resultJSON.Get("messages").Array()
|
||||||
|
if len(messages) < 2 {
|
||||||
|
if tt.wantHasReasoningContent || tt.wantHasContent {
|
||||||
|
t.Fatalf("Expected at least 2 messages (system + user/assistant), got %d", len(messages))
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check the last non-system message
|
||||||
|
var targetMsg gjson.Result
|
||||||
|
for i := len(messages) - 1; i >= 0; i-- {
|
||||||
|
if messages[i].Get("role").String() != "system" {
|
||||||
|
targetMsg = messages[i]
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check reasoning_content
|
||||||
|
gotReasoningContent := targetMsg.Get("reasoning_content").String()
|
||||||
|
gotHasReasoningContent := targetMsg.Get("reasoning_content").Exists()
|
||||||
|
|
||||||
|
if gotHasReasoningContent != tt.wantHasReasoningContent {
|
||||||
|
t.Errorf("reasoning_content existence = %v, want %v", gotHasReasoningContent, tt.wantHasReasoningContent)
|
||||||
|
}
|
||||||
|
|
||||||
|
if gotReasoningContent != tt.wantReasoningContent {
|
||||||
|
t.Errorf("reasoning_content = %q, want %q", gotReasoningContent, tt.wantReasoningContent)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check content
|
||||||
|
content := targetMsg.Get("content")
|
||||||
|
// content has meaningful content if it's a non-empty array, or a non-empty string
|
||||||
|
var gotHasContent bool
|
||||||
|
switch {
|
||||||
|
case content.IsArray():
|
||||||
|
gotHasContent = len(content.Array()) > 0
|
||||||
|
case content.Type == gjson.String:
|
||||||
|
gotHasContent = content.String() != ""
|
||||||
|
default:
|
||||||
|
gotHasContent = false
|
||||||
|
}
|
||||||
|
|
||||||
|
if gotHasContent != tt.wantHasContent {
|
||||||
|
t.Errorf("content existence = %v, want %v", gotHasContent, tt.wantHasContent)
|
||||||
|
}
|
||||||
|
|
||||||
|
if tt.wantHasContent && tt.wantContentText != "" {
|
||||||
|
// Find text content
|
||||||
|
var foundText string
|
||||||
|
content.ForEach(func(_, v gjson.Result) bool {
|
||||||
|
if v.Get("type").String() == "text" {
|
||||||
|
foundText = v.Get("text").String()
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
})
|
||||||
|
if foundText != tt.wantContentText {
|
||||||
|
t.Errorf("content text = %q, want %q", foundText, tt.wantContentText)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// TestConvertClaudeRequestToOpenAI_ThinkingOnlyMessagePreserved tests AC3:
|
||||||
|
// that a message with only thinking content is preserved (not dropped).
|
||||||
|
func TestConvertClaudeRequestToOpenAI_ThinkingOnlyMessagePreserved(t *testing.T) {
|
||||||
|
inputJSON := `{
|
||||||
|
"model": "claude-3-opus",
|
||||||
|
"messages": [
|
||||||
|
{
|
||||||
|
"role": "user",
|
||||||
|
"content": [{"type": "text", "text": "What is 2+2?"}]
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"role": "assistant",
|
||||||
|
"content": [{"type": "thinking", "thinking": "Let me calculate: 2+2=4"}]
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"role": "user",
|
||||||
|
"content": [{"type": "text", "text": "Thanks"}]
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}`
|
||||||
|
|
||||||
|
result := ConvertClaudeRequestToOpenAI("test-model", []byte(inputJSON), false)
|
||||||
|
resultJSON := gjson.ParseBytes(result)
|
||||||
|
|
||||||
|
messages := resultJSON.Get("messages").Array()
|
||||||
|
|
||||||
|
// Should have: system (auto-added) + user + assistant (thinking-only) + user = 4 messages
|
||||||
|
if len(messages) != 4 {
|
||||||
|
t.Fatalf("Expected 4 messages, got %d. Messages: %v", len(messages), resultJSON.Get("messages").Raw)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check the assistant message (index 2) has reasoning_content
|
||||||
|
assistantMsg := messages[2]
|
||||||
|
if assistantMsg.Get("role").String() != "assistant" {
|
||||||
|
t.Errorf("Expected message[2] to be assistant, got %s", assistantMsg.Get("role").String())
|
||||||
|
}
|
||||||
|
|
||||||
|
if !assistantMsg.Get("reasoning_content").Exists() {
|
||||||
|
t.Error("Expected assistant message to have reasoning_content")
|
||||||
|
}
|
||||||
|
|
||||||
|
if assistantMsg.Get("reasoning_content").String() != "Let me calculate: 2+2=4" {
|
||||||
|
t.Errorf("Unexpected reasoning_content: %s", assistantMsg.Get("reasoning_content").String())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestConvertClaudeRequestToOpenAI_ToolResultOrderAndContent(t *testing.T) {
|
||||||
|
inputJSON := `{
|
||||||
|
"model": "claude-3-opus",
|
||||||
|
"messages": [
|
||||||
|
{
|
||||||
|
"role": "assistant",
|
||||||
|
"content": [
|
||||||
|
{"type": "tool_use", "id": "call_1", "name": "do_work", "input": {"a": 1}}
|
||||||
|
]
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"role": "user",
|
||||||
|
"content": [
|
||||||
|
{"type": "text", "text": "before"},
|
||||||
|
{"type": "tool_result", "tool_use_id": "call_1", "content": [{"type":"text","text":"tool ok"}]},
|
||||||
|
{"type": "text", "text": "after"}
|
||||||
|
]
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}`
|
||||||
|
|
||||||
|
result := ConvertClaudeRequestToOpenAI("test-model", []byte(inputJSON), false)
|
||||||
|
resultJSON := gjson.ParseBytes(result)
|
||||||
|
messages := resultJSON.Get("messages").Array()
|
||||||
|
|
||||||
|
// OpenAI requires: tool messages MUST immediately follow assistant(tool_calls).
|
||||||
|
// Correct order: system + assistant(tool_calls) + tool(result) + user(before+after)
|
||||||
|
if len(messages) != 4 {
|
||||||
|
t.Fatalf("Expected 4 messages, got %d. Messages: %s", len(messages), resultJSON.Get("messages").Raw)
|
||||||
|
}
|
||||||
|
|
||||||
|
if messages[0].Get("role").String() != "system" {
|
||||||
|
t.Fatalf("Expected messages[0] to be system, got %s", messages[0].Get("role").String())
|
||||||
|
}
|
||||||
|
|
||||||
|
if messages[1].Get("role").String() != "assistant" || !messages[1].Get("tool_calls").Exists() {
|
||||||
|
t.Fatalf("Expected messages[1] to be assistant tool_calls, got %s: %s", messages[1].Get("role").String(), messages[1].Raw)
|
||||||
|
}
|
||||||
|
|
||||||
|
// tool message MUST immediately follow assistant(tool_calls) per OpenAI spec
|
||||||
|
if messages[2].Get("role").String() != "tool" {
|
||||||
|
t.Fatalf("Expected messages[2] to be tool (must follow tool_calls), got %s", messages[2].Get("role").String())
|
||||||
|
}
|
||||||
|
if got := messages[2].Get("tool_call_id").String(); got != "call_1" {
|
||||||
|
t.Fatalf("Expected tool_call_id %q, got %q", "call_1", got)
|
||||||
|
}
|
||||||
|
if got := messages[2].Get("content").String(); got != "tool ok" {
|
||||||
|
t.Fatalf("Expected tool content %q, got %q", "tool ok", got)
|
||||||
|
}
|
||||||
|
|
||||||
|
// User message comes after tool message
|
||||||
|
if messages[3].Get("role").String() != "user" {
|
||||||
|
t.Fatalf("Expected messages[3] to be user, got %s", messages[3].Get("role").String())
|
||||||
|
}
|
||||||
|
// User message should contain both "before" and "after" text
|
||||||
|
if got := messages[3].Get("content.0.text").String(); got != "before" {
|
||||||
|
t.Fatalf("Expected user text[0] %q, got %q", "before", got)
|
||||||
|
}
|
||||||
|
if got := messages[3].Get("content.1.text").String(); got != "after" {
|
||||||
|
t.Fatalf("Expected user text[1] %q, got %q", "after", got)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestConvertClaudeRequestToOpenAI_ToolResultObjectContent(t *testing.T) {
|
||||||
|
inputJSON := `{
|
||||||
|
"model": "claude-3-opus",
|
||||||
|
"messages": [
|
||||||
|
{
|
||||||
|
"role": "assistant",
|
||||||
|
"content": [
|
||||||
|
{"type": "tool_use", "id": "call_1", "name": "do_work", "input": {"a": 1}}
|
||||||
|
]
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"role": "user",
|
||||||
|
"content": [
|
||||||
|
{"type": "tool_result", "tool_use_id": "call_1", "content": {"foo": "bar"}}
|
||||||
|
]
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}`
|
||||||
|
|
||||||
|
result := ConvertClaudeRequestToOpenAI("test-model", []byte(inputJSON), false)
|
||||||
|
resultJSON := gjson.ParseBytes(result)
|
||||||
|
messages := resultJSON.Get("messages").Array()
|
||||||
|
|
||||||
|
// system + assistant(tool_calls) + tool(result)
|
||||||
|
if len(messages) != 3 {
|
||||||
|
t.Fatalf("Expected 3 messages, got %d. Messages: %s", len(messages), resultJSON.Get("messages").Raw)
|
||||||
|
}
|
||||||
|
|
||||||
|
if messages[2].Get("role").String() != "tool" {
|
||||||
|
t.Fatalf("Expected messages[2] to be tool, got %s", messages[2].Get("role").String())
|
||||||
|
}
|
||||||
|
|
||||||
|
toolContent := messages[2].Get("content").String()
|
||||||
|
parsed := gjson.Parse(toolContent)
|
||||||
|
if parsed.Get("foo").String() != "bar" {
|
||||||
|
t.Fatalf("Expected tool content JSON foo=bar, got %q", toolContent)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestConvertClaudeRequestToOpenAI_AssistantTextToolUseTextOrder(t *testing.T) {
|
||||||
|
inputJSON := `{
|
||||||
|
"model": "claude-3-opus",
|
||||||
|
"messages": [
|
||||||
|
{
|
||||||
|
"role": "assistant",
|
||||||
|
"content": [
|
||||||
|
{"type": "text", "text": "pre"},
|
||||||
|
{"type": "tool_use", "id": "call_1", "name": "do_work", "input": {"a": 1}},
|
||||||
|
{"type": "text", "text": "post"}
|
||||||
|
]
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}`
|
||||||
|
|
||||||
|
result := ConvertClaudeRequestToOpenAI("test-model", []byte(inputJSON), false)
|
||||||
|
resultJSON := gjson.ParseBytes(result)
|
||||||
|
messages := resultJSON.Get("messages").Array()
|
||||||
|
|
||||||
|
// New behavior: content + tool_calls unified in single assistant message
|
||||||
|
// Expect: system + assistant(content[pre,post] + tool_calls)
|
||||||
|
if len(messages) != 2 {
|
||||||
|
t.Fatalf("Expected 2 messages, got %d. Messages: %s", len(messages), resultJSON.Get("messages").Raw)
|
||||||
|
}
|
||||||
|
|
||||||
|
if messages[0].Get("role").String() != "system" {
|
||||||
|
t.Fatalf("Expected messages[0] to be system, got %s", messages[0].Get("role").String())
|
||||||
|
}
|
||||||
|
|
||||||
|
assistantMsg := messages[1]
|
||||||
|
if assistantMsg.Get("role").String() != "assistant" {
|
||||||
|
t.Fatalf("Expected messages[1] to be assistant, got %s", assistantMsg.Get("role").String())
|
||||||
|
}
|
||||||
|
|
||||||
|
// Should have both content and tool_calls in same message
|
||||||
|
if !assistantMsg.Get("tool_calls").Exists() {
|
||||||
|
t.Fatalf("Expected assistant message to have tool_calls")
|
||||||
|
}
|
||||||
|
if got := assistantMsg.Get("tool_calls.0.id").String(); got != "call_1" {
|
||||||
|
t.Fatalf("Expected tool_call id %q, got %q", "call_1", got)
|
||||||
|
}
|
||||||
|
if got := assistantMsg.Get("tool_calls.0.function.name").String(); got != "do_work" {
|
||||||
|
t.Fatalf("Expected tool_call name %q, got %q", "do_work", got)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Content should have both pre and post text
|
||||||
|
if got := assistantMsg.Get("content.0.text").String(); got != "pre" {
|
||||||
|
t.Fatalf("Expected content[0] text %q, got %q", "pre", got)
|
||||||
|
}
|
||||||
|
if got := assistantMsg.Get("content.1.text").String(); got != "post" {
|
||||||
|
t.Fatalf("Expected content[1] text %q, got %q", "post", got)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestConvertClaudeRequestToOpenAI_AssistantThinkingToolUseThinkingSplit(t *testing.T) {
|
||||||
|
inputJSON := `{
|
||||||
|
"model": "claude-3-opus",
|
||||||
|
"messages": [
|
||||||
|
{
|
||||||
|
"role": "assistant",
|
||||||
|
"content": [
|
||||||
|
{"type": "thinking", "thinking": "t1"},
|
||||||
|
{"type": "text", "text": "pre"},
|
||||||
|
{"type": "tool_use", "id": "call_1", "name": "do_work", "input": {"a": 1}},
|
||||||
|
{"type": "thinking", "thinking": "t2"},
|
||||||
|
{"type": "text", "text": "post"}
|
||||||
|
]
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}`
|
||||||
|
|
||||||
|
result := ConvertClaudeRequestToOpenAI("test-model", []byte(inputJSON), false)
|
||||||
|
resultJSON := gjson.ParseBytes(result)
|
||||||
|
messages := resultJSON.Get("messages").Array()
|
||||||
|
|
||||||
|
// New behavior: all content, thinking, and tool_calls unified in single assistant message
|
||||||
|
// Expect: system + assistant(content[pre,post] + tool_calls + reasoning_content[t1+t2])
|
||||||
|
if len(messages) != 2 {
|
||||||
|
t.Fatalf("Expected 2 messages, got %d. Messages: %s", len(messages), resultJSON.Get("messages").Raw)
|
||||||
|
}
|
||||||
|
|
||||||
|
assistantMsg := messages[1]
|
||||||
|
if assistantMsg.Get("role").String() != "assistant" {
|
||||||
|
t.Fatalf("Expected messages[1] to be assistant, got %s", assistantMsg.Get("role").String())
|
||||||
|
}
|
||||||
|
|
||||||
|
// Should have content with both pre and post
|
||||||
|
if got := assistantMsg.Get("content.0.text").String(); got != "pre" {
|
||||||
|
t.Fatalf("Expected content[0] text %q, got %q", "pre", got)
|
||||||
|
}
|
||||||
|
if got := assistantMsg.Get("content.1.text").String(); got != "post" {
|
||||||
|
t.Fatalf("Expected content[1] text %q, got %q", "post", got)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Should have tool_calls
|
||||||
|
if !assistantMsg.Get("tool_calls").Exists() {
|
||||||
|
t.Fatalf("Expected assistant message to have tool_calls")
|
||||||
|
}
|
||||||
|
|
||||||
|
// Should have combined reasoning_content from both thinking blocks
|
||||||
|
if got := assistantMsg.Get("reasoning_content").String(); got != "t1\n\nt2" {
|
||||||
|
t.Fatalf("Expected reasoning_content %q, got %q", "t1\n\nt2", got)
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -480,15 +480,15 @@ func collectOpenAIReasoningTexts(node gjson.Result) []string {
|
|||||||
|
|
||||||
switch node.Type {
|
switch node.Type {
|
||||||
case gjson.String:
|
case gjson.String:
|
||||||
if text := strings.TrimSpace(node.String()); text != "" {
|
if text := node.String(); text != "" {
|
||||||
texts = append(texts, text)
|
texts = append(texts, text)
|
||||||
}
|
}
|
||||||
case gjson.JSON:
|
case gjson.JSON:
|
||||||
if text := node.Get("text"); text.Exists() {
|
if text := node.Get("text"); text.Exists() {
|
||||||
if trimmed := strings.TrimSpace(text.String()); trimmed != "" {
|
if textStr := text.String(); textStr != "" {
|
||||||
texts = append(texts, trimmed)
|
texts = append(texts, textStr)
|
||||||
}
|
}
|
||||||
} else if raw := strings.TrimSpace(node.Raw); raw != "" && !strings.HasPrefix(raw, "{") && !strings.HasPrefix(raw, "[") {
|
} else if raw := node.Raw; raw != "" && !strings.HasPrefix(raw, "{") && !strings.HasPrefix(raw, "[") {
|
||||||
texts = append(texts, raw)
|
texts = append(texts, raw)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -390,6 +390,11 @@ func addEmptySchemaPlaceholder(jsonStr string) string {
|
|||||||
|
|
||||||
// If schema has properties but none are required, add a minimal placeholder.
|
// If schema has properties but none are required, add a minimal placeholder.
|
||||||
if propsVal.IsObject() && !hasRequiredProperties {
|
if propsVal.IsObject() && !hasRequiredProperties {
|
||||||
|
// DO NOT add placeholder if it's a top-level schema (parentPath is empty)
|
||||||
|
// or if we've already added a placeholder reason above.
|
||||||
|
if parentPath == "" {
|
||||||
|
continue
|
||||||
|
}
|
||||||
placeholderPath := joinPath(propsPath, "_")
|
placeholderPath := joinPath(propsPath, "_")
|
||||||
if !gjson.Get(jsonStr, placeholderPath).Exists() {
|
if !gjson.Get(jsonStr, placeholderPath).Exists() {
|
||||||
jsonStr, _ = sjson.Set(jsonStr, placeholderPath+".type", "boolean")
|
jsonStr, _ = sjson.Set(jsonStr, placeholderPath+".type", "boolean")
|
||||||
|
|||||||
@@ -127,8 +127,10 @@ func TestCleanJSONSchemaForAntigravity_AnyOfFlattening_SmartSelection(t *testing
|
|||||||
"type": "object",
|
"type": "object",
|
||||||
"description": "Accepts: null | object",
|
"description": "Accepts: null | object",
|
||||||
"properties": {
|
"properties": {
|
||||||
|
"_": { "type": "boolean" },
|
||||||
"kind": { "type": "string" }
|
"kind": { "type": "string" }
|
||||||
}
|
},
|
||||||
|
"required": ["_"]
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}`
|
}`
|
||||||
|
|||||||
@@ -251,9 +251,14 @@ func ThinkingBudgetToGemini3Level(model string, budget int) (string, bool) {
|
|||||||
|
|
||||||
// modelsWithDefaultThinking lists models that should have thinking enabled by default
|
// modelsWithDefaultThinking lists models that should have thinking enabled by default
|
||||||
// when no explicit thinkingConfig is provided.
|
// when no explicit thinkingConfig is provided.
|
||||||
|
// Note: Gemini 3 models are NOT included here because per Google's official documentation:
|
||||||
|
// - thinkingLevel defaults to "high" (dynamic thinking)
|
||||||
|
// - includeThoughts defaults to false
|
||||||
|
//
|
||||||
|
// We should not override these API defaults; let users explicitly configure if needed.
|
||||||
var modelsWithDefaultThinking = map[string]bool{
|
var modelsWithDefaultThinking = map[string]bool{
|
||||||
"gemini-3-pro-preview": true,
|
// "gemini-3-pro-preview": true,
|
||||||
"gemini-3-pro-image-preview": true,
|
// "gemini-3-pro-image-preview": true,
|
||||||
// "gemini-3-flash-preview": true,
|
// "gemini-3-flash-preview": true,
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -288,37 +293,73 @@ func ApplyDefaultThinkingIfNeeded(model string, body []byte) []byte {
|
|||||||
|
|
||||||
// ApplyGemini3ThinkingLevelFromMetadata applies thinkingLevel from metadata for Gemini 3 models.
|
// ApplyGemini3ThinkingLevelFromMetadata applies thinkingLevel from metadata for Gemini 3 models.
|
||||||
// For standard Gemini API format (generationConfig.thinkingConfig path).
|
// For standard Gemini API format (generationConfig.thinkingConfig path).
|
||||||
// This handles the case where reasoning_effort is specified via model name suffix (e.g., model(minimal)).
|
// This handles the case where reasoning_effort is specified via model name suffix (e.g., model(minimal))
|
||||||
|
// or numeric budget suffix (e.g., model(1000)) which gets converted to a thinkingLevel.
|
||||||
func ApplyGemini3ThinkingLevelFromMetadata(model string, metadata map[string]any, body []byte) []byte {
|
func ApplyGemini3ThinkingLevelFromMetadata(model string, metadata map[string]any, body []byte) []byte {
|
||||||
if !IsGemini3Model(model) {
|
// Use the alias from metadata if available for model type detection
|
||||||
|
lookupModel := ResolveOriginalModel(model, metadata)
|
||||||
|
if !IsGemini3Model(lookupModel) && !IsGemini3Model(model) {
|
||||||
return body
|
return body
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Determine which model to use for validation
|
||||||
|
checkModel := model
|
||||||
|
if IsGemini3Model(lookupModel) {
|
||||||
|
checkModel = lookupModel
|
||||||
|
}
|
||||||
|
|
||||||
|
// First try to get effort string from metadata
|
||||||
effort, ok := ReasoningEffortFromMetadata(metadata)
|
effort, ok := ReasoningEffortFromMetadata(metadata)
|
||||||
if !ok || effort == "" {
|
if ok && effort != "" {
|
||||||
return body
|
if level, valid := ValidateGemini3ThinkingLevel(checkModel, effort); valid {
|
||||||
|
return ApplyGeminiThinkingLevel(body, level, nil)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
// Validate and apply the thinkingLevel
|
|
||||||
if level, valid := ValidateGemini3ThinkingLevel(model, effort); valid {
|
// Fallback: check for numeric budget and convert to thinkingLevel
|
||||||
return ApplyGeminiThinkingLevel(body, level, nil)
|
budget, _, _, matched := ThinkingFromMetadata(metadata)
|
||||||
|
if matched && budget != nil {
|
||||||
|
if level, valid := ThinkingBudgetToGemini3Level(checkModel, *budget); valid {
|
||||||
|
return ApplyGeminiThinkingLevel(body, level, nil)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return body
|
return body
|
||||||
}
|
}
|
||||||
|
|
||||||
// ApplyGemini3ThinkingLevelFromMetadataCLI applies thinkingLevel from metadata for Gemini 3 models.
|
// ApplyGemini3ThinkingLevelFromMetadataCLI applies thinkingLevel from metadata for Gemini 3 models.
|
||||||
// For Gemini CLI API format (request.generationConfig.thinkingConfig path).
|
// For Gemini CLI API format (request.generationConfig.thinkingConfig path).
|
||||||
// This handles the case where reasoning_effort is specified via model name suffix (e.g., model(minimal)).
|
// This handles the case where reasoning_effort is specified via model name suffix (e.g., model(minimal))
|
||||||
|
// or numeric budget suffix (e.g., model(1000)) which gets converted to a thinkingLevel.
|
||||||
func ApplyGemini3ThinkingLevelFromMetadataCLI(model string, metadata map[string]any, body []byte) []byte {
|
func ApplyGemini3ThinkingLevelFromMetadataCLI(model string, metadata map[string]any, body []byte) []byte {
|
||||||
if !IsGemini3Model(model) {
|
// Use the alias from metadata if available for model type detection
|
||||||
|
lookupModel := ResolveOriginalModel(model, metadata)
|
||||||
|
if !IsGemini3Model(lookupModel) && !IsGemini3Model(model) {
|
||||||
return body
|
return body
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Determine which model to use for validation
|
||||||
|
checkModel := model
|
||||||
|
if IsGemini3Model(lookupModel) {
|
||||||
|
checkModel = lookupModel
|
||||||
|
}
|
||||||
|
|
||||||
|
// First try to get effort string from metadata
|
||||||
effort, ok := ReasoningEffortFromMetadata(metadata)
|
effort, ok := ReasoningEffortFromMetadata(metadata)
|
||||||
if !ok || effort == "" {
|
if ok && effort != "" {
|
||||||
return body
|
if level, valid := ValidateGemini3ThinkingLevel(checkModel, effort); valid {
|
||||||
|
return ApplyGeminiCLIThinkingLevel(body, level, nil)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
// Validate and apply the thinkingLevel
|
|
||||||
if level, valid := ValidateGemini3ThinkingLevel(model, effort); valid {
|
// Fallback: check for numeric budget and convert to thinkingLevel
|
||||||
return ApplyGeminiCLIThinkingLevel(body, level, nil)
|
budget, _, _, matched := ThinkingFromMetadata(metadata)
|
||||||
|
if matched && budget != nil {
|
||||||
|
if level, valid := ThinkingBudgetToGemini3Level(checkModel, *budget); valid {
|
||||||
|
return ApplyGeminiCLIThinkingLevel(body, level, nil)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return body
|
return body
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -326,15 +367,17 @@ func ApplyGemini3ThinkingLevelFromMetadataCLI(model string, metadata map[string]
|
|||||||
// For Gemini CLI API format (request.generationConfig.thinkingConfig path).
|
// For Gemini CLI API format (request.generationConfig.thinkingConfig path).
|
||||||
// Returns the modified body if thinkingConfig was added, otherwise returns the original.
|
// Returns the modified body if thinkingConfig was added, otherwise returns the original.
|
||||||
// For Gemini 3 models, uses thinkingLevel instead of thinkingBudget per Google's documentation.
|
// For Gemini 3 models, uses thinkingLevel instead of thinkingBudget per Google's documentation.
|
||||||
func ApplyDefaultThinkingIfNeededCLI(model string, body []byte) []byte {
|
func ApplyDefaultThinkingIfNeededCLI(model string, metadata map[string]any, body []byte) []byte {
|
||||||
if !ModelHasDefaultThinking(model) {
|
// Use the alias from metadata if available for model property lookup
|
||||||
|
lookupModel := ResolveOriginalModel(model, metadata)
|
||||||
|
if !ModelHasDefaultThinking(lookupModel) && !ModelHasDefaultThinking(model) {
|
||||||
return body
|
return body
|
||||||
}
|
}
|
||||||
if gjson.GetBytes(body, "request.generationConfig.thinkingConfig").Exists() {
|
if gjson.GetBytes(body, "request.generationConfig.thinkingConfig").Exists() {
|
||||||
return body
|
return body
|
||||||
}
|
}
|
||||||
// Gemini 3 models use thinkingLevel instead of thinkingBudget
|
// Gemini 3 models use thinkingLevel instead of thinkingBudget
|
||||||
if IsGemini3Model(model) {
|
if IsGemini3Model(lookupModel) || IsGemini3Model(model) {
|
||||||
// Don't set a default - let the API use its dynamic default ("high")
|
// Don't set a default - let the API use its dynamic default ("high")
|
||||||
// Only set includeThoughts
|
// Only set includeThoughts
|
||||||
updated, _ := sjson.SetBytes(body, "request.generationConfig.thinkingConfig.includeThoughts", true)
|
updated, _ := sjson.SetBytes(body, "request.generationConfig.thinkingConfig.includeThoughts", true)
|
||||||
|
|||||||
56
internal/util/sanitize_test.go
Normal file
56
internal/util/sanitize_test.go
Normal file
@@ -0,0 +1,56 @@
|
|||||||
|
package util
|
||||||
|
|
||||||
|
import (
|
||||||
|
"testing"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestSanitizeFunctionName(t *testing.T) {
|
||||||
|
tests := []struct {
|
||||||
|
name string
|
||||||
|
input string
|
||||||
|
expected string
|
||||||
|
}{
|
||||||
|
{"Normal", "valid_name", "valid_name"},
|
||||||
|
{"With Dots", "name.with.dots", "name.with.dots"},
|
||||||
|
{"With Colons", "name:with:colons", "name:with:colons"},
|
||||||
|
{"With Dashes", "name-with-dashes", "name-with-dashes"},
|
||||||
|
{"Mixed Allowed", "name.with_dots:colons-dashes", "name.with_dots:colons-dashes"},
|
||||||
|
{"Invalid Characters", "name!with@invalid#chars", "name_with_invalid_chars"},
|
||||||
|
{"Spaces", "name with spaces", "name_with_spaces"},
|
||||||
|
{"Non-ASCII", "name_with_你好_chars", "name_with____chars"},
|
||||||
|
{"Starts with digit", "123name", "_123name"},
|
||||||
|
{"Starts with dot", ".name", "_.name"},
|
||||||
|
{"Starts with colon", ":name", "_:name"},
|
||||||
|
{"Starts with dash", "-name", "_-name"},
|
||||||
|
{"Starts with invalid char", "!name", "_name"},
|
||||||
|
{"Exactly 64 chars", "this_is_a_very_long_name_that_exactly_reaches_sixty_four_charact", "this_is_a_very_long_name_that_exactly_reaches_sixty_four_charact"},
|
||||||
|
{"Too long (65 chars)", "this_is_a_very_long_name_that_exactly_reaches_sixty_four_charactX", "this_is_a_very_long_name_that_exactly_reaches_sixty_four_charact"},
|
||||||
|
{"Very long", "this_is_a_very_long_name_that_exceeds_the_sixty_four_character_limit_for_function_names", "this_is_a_very_long_name_that_exceeds_the_sixty_four_character_l"},
|
||||||
|
{"Starts with digit (64 chars total)", "1234567890123456789012345678901234567890123456789012345678901234", "_123456789012345678901234567890123456789012345678901234567890123"},
|
||||||
|
{"Starts with invalid char (64 chars total)", "!234567890123456789012345678901234567890123456789012345678901234", "_234567890123456789012345678901234567890123456789012345678901234"},
|
||||||
|
{"Empty", "", ""},
|
||||||
|
{"Single character invalid", "@", "_"},
|
||||||
|
{"Single character valid", "a", "a"},
|
||||||
|
{"Single character digit", "1", "_1"},
|
||||||
|
{"Single character underscore", "_", "_"},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, tt := range tests {
|
||||||
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
|
got := SanitizeFunctionName(tt.input)
|
||||||
|
if got != tt.expected {
|
||||||
|
t.Errorf("SanitizeFunctionName(%q) = %v, want %v", tt.input, got, tt.expected)
|
||||||
|
}
|
||||||
|
// Verify Gemini compliance
|
||||||
|
if len(got) > 64 {
|
||||||
|
t.Errorf("SanitizeFunctionName(%q) result too long: %d", tt.input, len(got))
|
||||||
|
}
|
||||||
|
if len(got) > 0 {
|
||||||
|
first := got[0]
|
||||||
|
if !((first >= 'a' && first <= 'z') || (first >= 'A' && first <= 'Z') || first == '_') {
|
||||||
|
t.Errorf("SanitizeFunctionName(%q) result starts with invalid char: %c", tt.input, first)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -12,9 +12,18 @@ func ModelSupportsThinking(model string) bool {
|
|||||||
if model == "" {
|
if model == "" {
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
// First check the global dynamic registry
|
||||||
if info := registry.GetGlobalRegistry().GetModelInfo(model); info != nil {
|
if info := registry.GetGlobalRegistry().GetModelInfo(model); info != nil {
|
||||||
return info.Thinking != nil
|
return info.Thinking != nil
|
||||||
}
|
}
|
||||||
|
// Fallback: check static model definitions
|
||||||
|
if info := registry.LookupStaticModelInfo(model); info != nil {
|
||||||
|
return info.Thinking != nil
|
||||||
|
}
|
||||||
|
// Fallback: check Antigravity static config
|
||||||
|
if cfg := registry.GetAntigravityModelConfig()[model]; cfg != nil {
|
||||||
|
return cfg.Thinking != nil
|
||||||
|
}
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -63,11 +72,19 @@ func thinkingRangeFromRegistry(model string) (found bool, min int, max int, zero
|
|||||||
if model == "" {
|
if model == "" {
|
||||||
return false, 0, 0, false, false
|
return false, 0, 0, false, false
|
||||||
}
|
}
|
||||||
info := registry.GetGlobalRegistry().GetModelInfo(model)
|
// First check global dynamic registry
|
||||||
if info == nil || info.Thinking == nil {
|
if info := registry.GetGlobalRegistry().GetModelInfo(model); info != nil && info.Thinking != nil {
|
||||||
return false, 0, 0, false, false
|
return true, info.Thinking.Min, info.Thinking.Max, info.Thinking.ZeroAllowed, info.Thinking.DynamicAllowed
|
||||||
}
|
}
|
||||||
return true, info.Thinking.Min, info.Thinking.Max, info.Thinking.ZeroAllowed, info.Thinking.DynamicAllowed
|
// Fallback: check static model definitions
|
||||||
|
if info := registry.LookupStaticModelInfo(model); info != nil && info.Thinking != nil {
|
||||||
|
return true, info.Thinking.Min, info.Thinking.Max, info.Thinking.ZeroAllowed, info.Thinking.DynamicAllowed
|
||||||
|
}
|
||||||
|
// Fallback: check Antigravity static config
|
||||||
|
if cfg := registry.GetAntigravityModelConfig()[model]; cfg != nil && cfg.Thinking != nil {
|
||||||
|
return true, cfg.Thinking.Min, cfg.Thinking.Max, cfg.Thinking.ZeroAllowed, cfg.Thinking.DynamicAllowed
|
||||||
|
}
|
||||||
|
return false, 0, 0, false, false
|
||||||
}
|
}
|
||||||
|
|
||||||
// GetModelThinkingLevels returns the discrete reasoning effort levels for the model.
|
// GetModelThinkingLevels returns the discrete reasoning effort levels for the model.
|
||||||
|
|||||||
@@ -4,16 +4,56 @@
|
|||||||
package util
|
package util
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io/fs"
|
|
||||||
"os"
|
"os"
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
|
"regexp"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
"github.com/router-for-me/CLIProxyAPI/v6/internal/config"
|
"github.com/router-for-me/CLIProxyAPI/v6/internal/config"
|
||||||
log "github.com/sirupsen/logrus"
|
log "github.com/sirupsen/logrus"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
var functionNameSanitizer = regexp.MustCompile(`[^a-zA-Z0-9_.:-]`)
|
||||||
|
|
||||||
|
// SanitizeFunctionName ensures a function name matches the requirements for Gemini/Vertex AI.
|
||||||
|
// It replaces invalid characters with underscores, ensures it starts with a letter or underscore,
|
||||||
|
// and truncates it to 64 characters if necessary.
|
||||||
|
// Regex Rule: [^a-zA-Z0-9_.:-] replaced with _.
|
||||||
|
func SanitizeFunctionName(name string) string {
|
||||||
|
if name == "" {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
|
||||||
|
// Replace invalid characters with underscore
|
||||||
|
sanitized := functionNameSanitizer.ReplaceAllString(name, "_")
|
||||||
|
|
||||||
|
// Ensure it starts with a letter or underscore
|
||||||
|
// Re-reading requirements: Must start with a letter or an underscore.
|
||||||
|
if len(sanitized) > 0 {
|
||||||
|
first := sanitized[0]
|
||||||
|
if !((first >= 'a' && first <= 'z') || (first >= 'A' && first <= 'Z') || first == '_') {
|
||||||
|
// If it starts with an allowed character but not allowed at the beginning (digit, dot, colon, dash),
|
||||||
|
// we must prepend an underscore.
|
||||||
|
|
||||||
|
// To stay within the 64-character limit while prepending, we must truncate first.
|
||||||
|
if len(sanitized) >= 64 {
|
||||||
|
sanitized = sanitized[:63]
|
||||||
|
}
|
||||||
|
sanitized = "_" + sanitized
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
sanitized = "_"
|
||||||
|
}
|
||||||
|
|
||||||
|
// Truncate to 64 characters
|
||||||
|
if len(sanitized) > 64 {
|
||||||
|
sanitized = sanitized[:64]
|
||||||
|
}
|
||||||
|
return sanitized
|
||||||
|
}
|
||||||
|
|
||||||
// SetLogLevel configures the logrus log level based on the configuration.
|
// SetLogLevel configures the logrus log level based on the configuration.
|
||||||
// It sets the log level to DebugLevel if debug mode is enabled, otherwise to InfoLevel.
|
// It sets the log level to DebugLevel if debug mode is enabled, otherwise to InfoLevel.
|
||||||
func SetLogLevel(cfg *config.Config) {
|
func SetLogLevel(cfg *config.Config) {
|
||||||
@@ -53,36 +93,23 @@ func ResolveAuthDir(authDir string) (string, error) {
|
|||||||
return filepath.Clean(authDir), nil
|
return filepath.Clean(authDir), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// CountAuthFiles returns the number of JSON auth files located under the provided directory.
|
// CountAuthFiles returns the number of auth records available through the provided Store.
|
||||||
// The function resolves leading tildes to the user's home directory and performs a case-insensitive
|
// For filesystem-backed stores, this reflects the number of JSON auth files under the configured directory.
|
||||||
// match on the ".json" suffix so that files saved with uppercase extensions are also counted.
|
func CountAuthFiles[T any](ctx context.Context, store interface {
|
||||||
func CountAuthFiles(authDir string) int {
|
List(context.Context) ([]T, error)
|
||||||
dir, err := ResolveAuthDir(authDir)
|
}) int {
|
||||||
|
if store == nil {
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
if ctx == nil {
|
||||||
|
ctx = context.Background()
|
||||||
|
}
|
||||||
|
entries, err := store.List(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Debugf("countAuthFiles: failed to resolve auth directory: %v", err)
|
log.Debugf("countAuthFiles: failed to list auth records: %v", err)
|
||||||
return 0
|
return 0
|
||||||
}
|
}
|
||||||
if dir == "" {
|
return len(entries)
|
||||||
return 0
|
|
||||||
}
|
|
||||||
count := 0
|
|
||||||
walkErr := filepath.WalkDir(dir, func(path string, d fs.DirEntry, err error) error {
|
|
||||||
if err != nil {
|
|
||||||
log.Debugf("countAuthFiles: error accessing %s: %v", path, err)
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
if d.IsDir() {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
if strings.HasSuffix(strings.ToLower(d.Name()), ".json") {
|
|
||||||
count++
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
})
|
|
||||||
if walkErr != nil {
|
|
||||||
log.Debugf("countAuthFiles: walk error: %v", walkErr)
|
|
||||||
}
|
|
||||||
return count
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// WritablePath returns the cleaned WRITABLE_PATH environment variable when it is set.
|
// WritablePath returns the cleaned WRITABLE_PATH environment variable when it is set.
|
||||||
|
|||||||
@@ -81,7 +81,9 @@ func (m *Manager) applyOAuthModelMapping(auth *Auth, requestedModel string, meta
|
|||||||
out[k] = v
|
out[k] = v
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
out[util.ModelMappingOriginalModelMetadataKey] = upstreamModel
|
// Store the requested alias (e.g., "gp") so downstream can use it to look up
|
||||||
|
// model metadata from the global registry where it was registered under this alias.
|
||||||
|
out[util.ModelMappingOriginalModelMetadataKey] = requestedModel
|
||||||
return upstreamModel, out
|
return upstreamModel, out
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -5,6 +5,9 @@ import "github.com/router-for-me/CLIProxyAPI/v6/internal/registry"
|
|||||||
// ModelInfo re-exports the registry model info structure.
|
// ModelInfo re-exports the registry model info structure.
|
||||||
type ModelInfo = registry.ModelInfo
|
type ModelInfo = registry.ModelInfo
|
||||||
|
|
||||||
|
// ModelRegistryHook re-exports the registry hook interface for external integrations.
|
||||||
|
type ModelRegistryHook = registry.ModelRegistryHook
|
||||||
|
|
||||||
// ModelRegistry describes registry operations consumed by external callers.
|
// ModelRegistry describes registry operations consumed by external callers.
|
||||||
type ModelRegistry interface {
|
type ModelRegistry interface {
|
||||||
RegisterClient(clientID, clientProvider string, models []*ModelInfo)
|
RegisterClient(clientID, clientProvider string, models []*ModelInfo)
|
||||||
@@ -13,9 +16,15 @@ type ModelRegistry interface {
|
|||||||
ClearModelQuotaExceeded(clientID, modelID string)
|
ClearModelQuotaExceeded(clientID, modelID string)
|
||||||
ClientSupportsModel(clientID, modelID string) bool
|
ClientSupportsModel(clientID, modelID string) bool
|
||||||
GetAvailableModels(handlerType string) []map[string]any
|
GetAvailableModels(handlerType string) []map[string]any
|
||||||
|
GetAvailableModelsByProvider(provider string) []*ModelInfo
|
||||||
}
|
}
|
||||||
|
|
||||||
// GlobalModelRegistry returns the shared registry instance.
|
// GlobalModelRegistry returns the shared registry instance.
|
||||||
func GlobalModelRegistry() ModelRegistry {
|
func GlobalModelRegistry() ModelRegistry {
|
||||||
return registry.GetGlobalRegistry()
|
return registry.GetGlobalRegistry()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// SetGlobalModelRegistryHook registers an optional hook on the shared global registry instance.
|
||||||
|
func SetGlobalModelRegistryHook(hook ModelRegistryHook) {
|
||||||
|
registry.GetGlobalRegistry().SetHook(hook)
|
||||||
|
}
|
||||||
|
|||||||
211
test/model_alias_thinking_suffix_test.go
Normal file
211
test/model_alias_thinking_suffix_test.go
Normal file
@@ -0,0 +1,211 @@
|
|||||||
|
package test
|
||||||
|
|
||||||
|
import (
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/router-for-me/CLIProxyAPI/v6/internal/runtime/executor"
|
||||||
|
"github.com/router-for-me/CLIProxyAPI/v6/internal/util"
|
||||||
|
"github.com/tidwall/gjson"
|
||||||
|
)
|
||||||
|
|
||||||
|
// TestModelAliasThinkingSuffix tests the 32 test cases defined in docs/thinking_suffix_test_cases.md
|
||||||
|
// These tests verify the thinking suffix parsing and application logic across different providers.
|
||||||
|
func TestModelAliasThinkingSuffix(t *testing.T) {
|
||||||
|
tests := []struct {
|
||||||
|
id int
|
||||||
|
name string
|
||||||
|
provider string
|
||||||
|
requestModel string
|
||||||
|
suffixType string
|
||||||
|
expectedField string // "thinkingBudget", "thinkingLevel", "budget_tokens", "reasoning_effort", "enable_thinking"
|
||||||
|
expectedValue any
|
||||||
|
upstreamModel string // The upstream model after alias resolution
|
||||||
|
isAlias bool
|
||||||
|
}{
|
||||||
|
// === 1. Antigravity Provider ===
|
||||||
|
// 1.1 Budget-only models (Gemini 2.5)
|
||||||
|
{1, "antigravity_original_numeric", "antigravity", "gemini-2.5-computer-use-preview-10-2025(1000)", "numeric", "thinkingBudget", 1000, "gemini-2.5-computer-use-preview-10-2025", false},
|
||||||
|
{2, "antigravity_alias_numeric", "antigravity", "gp(1000)", "numeric", "thinkingBudget", 1000, "gemini-2.5-computer-use-preview-10-2025", true},
|
||||||
|
// 1.2 Budget+Levels models (Gemini 3)
|
||||||
|
{3, "antigravity_original_numeric_to_level", "antigravity", "gemini-3-flash-preview(1000)", "numeric", "thinkingLevel", "low", "gemini-3-flash-preview", false},
|
||||||
|
{4, "antigravity_original_level", "antigravity", "gemini-3-flash-preview(low)", "level", "thinkingLevel", "low", "gemini-3-flash-preview", false},
|
||||||
|
{5, "antigravity_alias_numeric_to_level", "antigravity", "gf(1000)", "numeric", "thinkingLevel", "low", "gemini-3-flash-preview", true},
|
||||||
|
{6, "antigravity_alias_level", "antigravity", "gf(low)", "level", "thinkingLevel", "low", "gemini-3-flash-preview", true},
|
||||||
|
|
||||||
|
// === 2. Gemini CLI Provider ===
|
||||||
|
// 2.1 Budget-only models
|
||||||
|
{7, "gemini_cli_original_numeric", "gemini-cli", "gemini-2.5-pro(8192)", "numeric", "thinkingBudget", 8192, "gemini-2.5-pro", false},
|
||||||
|
{8, "gemini_cli_alias_numeric", "gemini-cli", "g25p(8192)", "numeric", "thinkingBudget", 8192, "gemini-2.5-pro", true},
|
||||||
|
// 2.2 Budget+Levels models
|
||||||
|
{9, "gemini_cli_original_numeric_to_level", "gemini-cli", "gemini-3-flash-preview(1000)", "numeric", "thinkingLevel", "low", "gemini-3-flash-preview", false},
|
||||||
|
{10, "gemini_cli_original_level", "gemini-cli", "gemini-3-flash-preview(low)", "level", "thinkingLevel", "low", "gemini-3-flash-preview", false},
|
||||||
|
{11, "gemini_cli_alias_numeric_to_level", "gemini-cli", "gf(1000)", "numeric", "thinkingLevel", "low", "gemini-3-flash-preview", true},
|
||||||
|
{12, "gemini_cli_alias_level", "gemini-cli", "gf(low)", "level", "thinkingLevel", "low", "gemini-3-flash-preview", true},
|
||||||
|
|
||||||
|
// === 3. Vertex Provider ===
|
||||||
|
// 3.1 Budget-only models
|
||||||
|
{13, "vertex_original_numeric", "vertex", "gemini-2.5-pro(16384)", "numeric", "thinkingBudget", 16384, "gemini-2.5-pro", false},
|
||||||
|
{14, "vertex_alias_numeric", "vertex", "vg25p(16384)", "numeric", "thinkingBudget", 16384, "gemini-2.5-pro", true},
|
||||||
|
// 3.2 Budget+Levels models
|
||||||
|
{15, "vertex_original_numeric_to_level", "vertex", "gemini-3-flash-preview(1000)", "numeric", "thinkingLevel", "low", "gemini-3-flash-preview", false},
|
||||||
|
{16, "vertex_original_level", "vertex", "gemini-3-flash-preview(low)", "level", "thinkingLevel", "low", "gemini-3-flash-preview", false},
|
||||||
|
{17, "vertex_alias_numeric_to_level", "vertex", "vgf(1000)", "numeric", "thinkingLevel", "low", "gemini-3-flash-preview", true},
|
||||||
|
{18, "vertex_alias_level", "vertex", "vgf(low)", "level", "thinkingLevel", "low", "gemini-3-flash-preview", true},
|
||||||
|
|
||||||
|
// === 4. AI Studio Provider ===
|
||||||
|
// 4.1 Budget-only models
|
||||||
|
{19, "aistudio_original_numeric", "aistudio", "gemini-2.5-pro(12000)", "numeric", "thinkingBudget", 12000, "gemini-2.5-pro", false},
|
||||||
|
{20, "aistudio_alias_numeric", "aistudio", "ag25p(12000)", "numeric", "thinkingBudget", 12000, "gemini-2.5-pro", true},
|
||||||
|
// 4.2 Budget+Levels models
|
||||||
|
{21, "aistudio_original_numeric_to_level", "aistudio", "gemini-3-flash-preview(1000)", "numeric", "thinkingLevel", "low", "gemini-3-flash-preview", false},
|
||||||
|
{22, "aistudio_original_level", "aistudio", "gemini-3-flash-preview(low)", "level", "thinkingLevel", "low", "gemini-3-flash-preview", false},
|
||||||
|
{23, "aistudio_alias_numeric_to_level", "aistudio", "agf(1000)", "numeric", "thinkingLevel", "low", "gemini-3-flash-preview", true},
|
||||||
|
{24, "aistudio_alias_level", "aistudio", "agf(low)", "level", "thinkingLevel", "low", "gemini-3-flash-preview", true},
|
||||||
|
|
||||||
|
// === 5. Claude Provider ===
|
||||||
|
{25, "claude_original_numeric", "claude", "claude-sonnet-4-5-20250929(16384)", "numeric", "budget_tokens", 16384, "claude-sonnet-4-5-20250929", false},
|
||||||
|
{26, "claude_alias_numeric", "claude", "cs45(16384)", "numeric", "budget_tokens", 16384, "claude-sonnet-4-5-20250929", true},
|
||||||
|
|
||||||
|
// === 6. Codex Provider ===
|
||||||
|
{27, "codex_original_level", "codex", "gpt-5(high)", "level", "reasoning_effort", "high", "gpt-5", false},
|
||||||
|
{28, "codex_alias_level", "codex", "g5(high)", "level", "reasoning_effort", "high", "gpt-5", true},
|
||||||
|
|
||||||
|
// === 7. Qwen Provider ===
|
||||||
|
{29, "qwen_original_level", "qwen", "qwen3-coder-plus(high)", "level", "enable_thinking", true, "qwen3-coder-plus", false},
|
||||||
|
{30, "qwen_alias_level", "qwen", "qcp(high)", "level", "enable_thinking", true, "qwen3-coder-plus", true},
|
||||||
|
|
||||||
|
// === 8. iFlow Provider ===
|
||||||
|
{31, "iflow_original_level", "iflow", "glm-4.7(high)", "level", "reasoning_effort", "high", "glm-4.7", false},
|
||||||
|
{32, "iflow_alias_level", "iflow", "glm(high)", "level", "reasoning_effort", "high", "glm-4.7", true},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, tt := range tests {
|
||||||
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
|
// Step 1: Parse model suffix (simulates SDK layer normalization)
|
||||||
|
// For "gp(1000)" -> requestedModel="gp", metadata={thinking_budget: 1000}
|
||||||
|
requestedModel, metadata := util.NormalizeThinkingModel(tt.requestModel)
|
||||||
|
|
||||||
|
// Verify suffix was parsed
|
||||||
|
if metadata == nil && (tt.suffixType == "numeric" || tt.suffixType == "level") {
|
||||||
|
t.Errorf("Case #%d: NormalizeThinkingModel(%q) metadata is nil", tt.id, tt.requestModel)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Step 2: Simulate OAuth model mapping
|
||||||
|
// Real flow: applyOAuthModelMapping stores requestedModel (the alias) in metadata
|
||||||
|
if tt.isAlias {
|
||||||
|
if metadata == nil {
|
||||||
|
metadata = make(map[string]any)
|
||||||
|
}
|
||||||
|
metadata[util.ModelMappingOriginalModelMetadataKey] = requestedModel
|
||||||
|
}
|
||||||
|
|
||||||
|
// Step 3: Verify metadata extraction
|
||||||
|
switch tt.suffixType {
|
||||||
|
case "numeric":
|
||||||
|
budget, _, _, matched := util.ThinkingFromMetadata(metadata)
|
||||||
|
if !matched {
|
||||||
|
t.Errorf("Case #%d: ThinkingFromMetadata did not match", tt.id)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if budget == nil {
|
||||||
|
t.Errorf("Case #%d: expected budget in metadata", tt.id)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
// For thinkingBudget/budget_tokens, verify the parsed budget value
|
||||||
|
if tt.expectedField == "thinkingBudget" || tt.expectedField == "budget_tokens" {
|
||||||
|
expectedBudget := tt.expectedValue.(int)
|
||||||
|
if *budget != expectedBudget {
|
||||||
|
t.Errorf("Case #%d: budget = %d, want %d", tt.id, *budget, expectedBudget)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// For thinkingLevel (Gemini 3), verify conversion from budget to level
|
||||||
|
if tt.expectedField == "thinkingLevel" {
|
||||||
|
level, ok := util.ThinkingBudgetToGemini3Level(tt.upstreamModel, *budget)
|
||||||
|
if !ok {
|
||||||
|
t.Errorf("Case #%d: ThinkingBudgetToGemini3Level failed", tt.id)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
expectedLevel := tt.expectedValue.(string)
|
||||||
|
if level != expectedLevel {
|
||||||
|
t.Errorf("Case #%d: converted level = %q, want %q", tt.id, level, expectedLevel)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
case "level":
|
||||||
|
_, _, effort, matched := util.ThinkingFromMetadata(metadata)
|
||||||
|
if !matched {
|
||||||
|
t.Errorf("Case #%d: ThinkingFromMetadata did not match", tt.id)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if effort == nil {
|
||||||
|
t.Errorf("Case #%d: expected effort in metadata", tt.id)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if tt.expectedField == "thinkingLevel" || tt.expectedField == "reasoning_effort" {
|
||||||
|
expectedEffort := tt.expectedValue.(string)
|
||||||
|
if *effort != expectedEffort {
|
||||||
|
t.Errorf("Case #%d: effort = %q, want %q", tt.id, *effort, expectedEffort)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Step 4: Test Gemini-specific thinkingLevel conversion for Gemini 3 models
|
||||||
|
if tt.expectedField == "thinkingLevel" && util.IsGemini3Model(tt.upstreamModel) {
|
||||||
|
body := []byte(`{"request":{"contents":[]}}`)
|
||||||
|
|
||||||
|
// Build metadata simulating real OAuth flow:
|
||||||
|
// - requestedModel (alias like "gf") is stored in model_mapping_original_model
|
||||||
|
// - upstreamModel is passed as the model parameter
|
||||||
|
testMetadata := make(map[string]any)
|
||||||
|
if tt.isAlias {
|
||||||
|
// Real flow: applyOAuthModelMapping stores requestedModel (the alias)
|
||||||
|
testMetadata[util.ModelMappingOriginalModelMetadataKey] = requestedModel
|
||||||
|
}
|
||||||
|
// Copy parsed metadata (thinking_budget, reasoning_effort, etc.)
|
||||||
|
for k, v := range metadata {
|
||||||
|
testMetadata[k] = v
|
||||||
|
}
|
||||||
|
|
||||||
|
result := util.ApplyGemini3ThinkingLevelFromMetadataCLI(tt.upstreamModel, testMetadata, body)
|
||||||
|
levelVal := gjson.GetBytes(result, "request.generationConfig.thinkingConfig.thinkingLevel")
|
||||||
|
|
||||||
|
expectedLevel := tt.expectedValue.(string)
|
||||||
|
if !levelVal.Exists() {
|
||||||
|
t.Errorf("Case #%d: expected thinkingLevel in result", tt.id)
|
||||||
|
} else if levelVal.String() != expectedLevel {
|
||||||
|
t.Errorf("Case #%d: thinkingLevel = %q, want %q", tt.id, levelVal.String(), expectedLevel)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Step 5: Test Gemini 2.5 thinkingBudget application using real ApplyThinkingMetadataCLI flow
|
||||||
|
if tt.expectedField == "thinkingBudget" && util.IsGemini25Model(tt.upstreamModel) {
|
||||||
|
body := []byte(`{"request":{"contents":[]}}`)
|
||||||
|
|
||||||
|
// Build metadata simulating real OAuth flow:
|
||||||
|
// - requestedModel (alias like "gp") is stored in model_mapping_original_model
|
||||||
|
// - upstreamModel is passed as the model parameter
|
||||||
|
testMetadata := make(map[string]any)
|
||||||
|
if tt.isAlias {
|
||||||
|
// Real flow: applyOAuthModelMapping stores requestedModel (the alias)
|
||||||
|
testMetadata[util.ModelMappingOriginalModelMetadataKey] = requestedModel
|
||||||
|
}
|
||||||
|
// Copy parsed metadata (thinking_budget, reasoning_effort, etc.)
|
||||||
|
for k, v := range metadata {
|
||||||
|
testMetadata[k] = v
|
||||||
|
}
|
||||||
|
|
||||||
|
// Use the exported ApplyThinkingMetadataCLI which includes the fallback logic
|
||||||
|
result := executor.ApplyThinkingMetadataCLI(body, testMetadata, tt.upstreamModel)
|
||||||
|
budgetVal := gjson.GetBytes(result, "request.generationConfig.thinkingConfig.thinkingBudget")
|
||||||
|
|
||||||
|
expectedBudget := tt.expectedValue.(int)
|
||||||
|
if !budgetVal.Exists() {
|
||||||
|
t.Errorf("Case #%d: expected thinkingBudget in result", tt.id)
|
||||||
|
} else if int(budgetVal.Int()) != expectedBudget {
|
||||||
|
t.Errorf("Case #%d: thinkingBudget = %d, want %d", tt.id, int(budgetVal.Int()), expectedBudget)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
Reference in New Issue
Block a user