feat: gemini 3.0 pro preview added (as default gemini pro model)

refactor: code cleanup
This commit is contained in:
Fahad
2025-11-18 20:28:27 +04:00
parent 749bc73079
commit 25fd72fbd3
19 changed files with 61 additions and 43 deletions

View File

@@ -59,7 +59,7 @@ class TestAutoModeProviderSelection:
balanced = ModelProviderRegistry.get_preferred_fallback_model(ToolModelCategory.BALANCED)
# Should select appropriate Gemini models
assert extended_reasoning in ["gemini-2.5-pro", "pro"]
assert extended_reasoning in ["gemini-3-pro-preview", "gemini-2.5-pro", "pro"]
assert fast_response in ["gemini-2.5-flash", "flash"]
assert balanced in ["gemini-2.5-flash", "flash"]
@@ -139,7 +139,7 @@ class TestAutoModeProviderSelection:
fast_response = ModelProviderRegistry.get_preferred_fallback_model(ToolModelCategory.FAST_RESPONSE)
# Should prefer Gemini now (based on new provider priority: Gemini before OpenAI)
assert extended_reasoning == "gemini-2.5-pro" # Gemini has higher priority now
assert extended_reasoning == "gemini-3-pro-preview" # Gemini 3 Pro Preview has higher priority now
# Should prefer Gemini for fast response
assert fast_response == "gemini-2.5-flash" # Gemini has higher priority now
@@ -317,7 +317,7 @@ class TestAutoModeProviderSelection:
# Test that providers resolve aliases correctly
test_cases = [
("flash", ProviderType.GOOGLE, "gemini-2.5-flash"),
("pro", ProviderType.GOOGLE, "gemini-2.5-pro"),
("pro", ProviderType.GOOGLE, "gemini-3-pro-preview"), # "pro" now resolves to gemini-3-pro-preview
("mini", ProviderType.OPENAI, "gpt-5-mini"), # "mini" now resolves to gpt-5-mini
("o3mini", ProviderType.OPENAI, "o3-mini"),
("grok", ProviderType.XAI, "grok-4"),