Grok-4 support
This commit is contained in:
@@ -202,7 +202,7 @@ class ModelProvider(ABC):
|
||||
prompt: str,
|
||||
model_name: str,
|
||||
system_prompt: Optional[str] = None,
|
||||
temperature: float = 0.7,
|
||||
temperature: float = 0.3,
|
||||
max_output_tokens: Optional[int] = None,
|
||||
**kwargs,
|
||||
) -> ModelResponse:
|
||||
|
||||
@@ -236,7 +236,7 @@ class CustomProvider(OpenAICompatibleProvider):
|
||||
prompt: str,
|
||||
model_name: str,
|
||||
system_prompt: Optional[str] = None,
|
||||
temperature: float = 0.7,
|
||||
temperature: float = 0.3,
|
||||
max_output_tokens: Optional[int] = None,
|
||||
**kwargs,
|
||||
) -> ModelResponse:
|
||||
|
||||
@@ -375,7 +375,7 @@ class DIALModelProvider(OpenAICompatibleProvider):
|
||||
prompt: str,
|
||||
model_name: str,
|
||||
system_prompt: Optional[str] = None,
|
||||
temperature: float = 0.7,
|
||||
temperature: float = 0.3,
|
||||
max_output_tokens: Optional[int] = None,
|
||||
images: Optional[list[str]] = None,
|
||||
**kwargs,
|
||||
|
||||
@@ -155,7 +155,7 @@ class GeminiModelProvider(ModelProvider):
|
||||
prompt: str,
|
||||
model_name: str,
|
||||
system_prompt: Optional[str] = None,
|
||||
temperature: float = 0.7,
|
||||
temperature: float = 0.3,
|
||||
max_output_tokens: Optional[int] = None,
|
||||
thinking_mode: str = "medium",
|
||||
images: Optional[list[str]] = None,
|
||||
|
||||
@@ -389,7 +389,7 @@ class OpenAICompatibleProvider(ModelProvider):
|
||||
prompt: str,
|
||||
model_name: str,
|
||||
system_prompt: Optional[str] = None,
|
||||
temperature: float = 0.7,
|
||||
temperature: float = 0.3,
|
||||
max_output_tokens: Optional[int] = None,
|
||||
images: Optional[list[str]] = None,
|
||||
**kwargs,
|
||||
|
||||
@@ -221,7 +221,7 @@ class OpenAIModelProvider(OpenAICompatibleProvider):
|
||||
prompt: str,
|
||||
model_name: str,
|
||||
system_prompt: Optional[str] = None,
|
||||
temperature: float = 0.7,
|
||||
temperature: float = 0.3,
|
||||
max_output_tokens: Optional[int] = None,
|
||||
**kwargs,
|
||||
) -> ModelResponse:
|
||||
|
||||
@@ -158,7 +158,7 @@ class OpenRouterProvider(OpenAICompatibleProvider):
|
||||
prompt: str,
|
||||
model_name: str,
|
||||
system_prompt: Optional[str] = None,
|
||||
temperature: float = 0.7,
|
||||
temperature: float = 0.3,
|
||||
max_output_tokens: Optional[int] = None,
|
||||
**kwargs,
|
||||
) -> ModelResponse:
|
||||
|
||||
@@ -24,6 +24,24 @@ class XAIModelProvider(OpenAICompatibleProvider):
|
||||
|
||||
# Model configurations using ModelCapabilities objects
|
||||
SUPPORTED_MODELS = {
|
||||
"grok-4": ModelCapabilities(
|
||||
provider=ProviderType.XAI,
|
||||
model_name="grok-4",
|
||||
friendly_name="X.AI (Grok 4)",
|
||||
context_window=256_000, # 256K tokens
|
||||
max_output_tokens=256_000, # 256K tokens max output
|
||||
supports_extended_thinking=True, # Grok-4 supports reasoning mode
|
||||
supports_system_prompts=True,
|
||||
supports_streaming=True,
|
||||
supports_function_calling=True, # Function calling supported
|
||||
supports_json_mode=True, # Structured outputs supported
|
||||
supports_images=True, # Multimodal capabilities
|
||||
max_image_size_mb=20.0, # Standard image size limit
|
||||
supports_temperature=True,
|
||||
temperature_constraint=create_temperature_constraint("range"),
|
||||
description="GROK-4 (256K context) - Frontier multimodal reasoning model with advanced capabilities",
|
||||
aliases=["grok", "grok4", "grok-4"],
|
||||
),
|
||||
"grok-3": ModelCapabilities(
|
||||
provider=ProviderType.XAI,
|
||||
model_name="grok-3",
|
||||
@@ -40,7 +58,7 @@ class XAIModelProvider(OpenAICompatibleProvider):
|
||||
supports_temperature=True,
|
||||
temperature_constraint=create_temperature_constraint("range"),
|
||||
description="GROK-3 (131K context) - Advanced reasoning model from X.AI, excellent for complex analysis",
|
||||
aliases=["grok", "grok3"],
|
||||
aliases=["grok3"],
|
||||
),
|
||||
"grok-3-fast": ModelCapabilities(
|
||||
provider=ProviderType.XAI,
|
||||
@@ -113,7 +131,7 @@ class XAIModelProvider(OpenAICompatibleProvider):
|
||||
prompt: str,
|
||||
model_name: str,
|
||||
system_prompt: Optional[str] = None,
|
||||
temperature: float = 0.7,
|
||||
temperature: float = 0.3,
|
||||
max_output_tokens: Optional[int] = None,
|
||||
**kwargs,
|
||||
) -> ModelResponse:
|
||||
@@ -133,8 +151,10 @@ class XAIModelProvider(OpenAICompatibleProvider):
|
||||
|
||||
def supports_thinking_mode(self, model_name: str) -> bool:
|
||||
"""Check if the model supports extended thinking mode."""
|
||||
# Currently GROK models do not support extended thinking
|
||||
# This may change with future GROK model releases
|
||||
resolved_name = self._resolve_model_name(model_name)
|
||||
capabilities = self.SUPPORTED_MODELS.get(resolved_name)
|
||||
if capabilities:
|
||||
return capabilities.supports_extended_thinking
|
||||
return False
|
||||
|
||||
def get_preferred_model(self, category: "ToolModelCategory", allowed_models: list[str]) -> Optional[str]:
|
||||
@@ -153,22 +173,28 @@ class XAIModelProvider(OpenAICompatibleProvider):
|
||||
return None
|
||||
|
||||
if category == ToolModelCategory.EXTENDED_REASONING:
|
||||
# Prefer GROK-3 for reasoning
|
||||
if "grok-3" in allowed_models:
|
||||
# Prefer GROK-4 for advanced reasoning with thinking mode
|
||||
if "grok-4" in allowed_models:
|
||||
return "grok-4"
|
||||
elif "grok-3" in allowed_models:
|
||||
return "grok-3"
|
||||
# Fall back to any available model
|
||||
return allowed_models[0]
|
||||
|
||||
elif category == ToolModelCategory.FAST_RESPONSE:
|
||||
# Prefer GROK-3-Fast for speed
|
||||
# Prefer GROK-3-Fast for speed, then GROK-4
|
||||
if "grok-3-fast" in allowed_models:
|
||||
return "grok-3-fast"
|
||||
elif "grok-4" in allowed_models:
|
||||
return "grok-4"
|
||||
# Fall back to any available model
|
||||
return allowed_models[0]
|
||||
|
||||
else: # BALANCED or default
|
||||
# Prefer standard GROK-3 for balanced use
|
||||
if "grok-3" in allowed_models:
|
||||
# Prefer GROK-4 for balanced use (best overall capabilities)
|
||||
if "grok-4" in allowed_models:
|
||||
return "grok-4"
|
||||
elif "grok-3" in allowed_models:
|
||||
return "grok-3"
|
||||
elif "grok-3-fast" in allowed_models:
|
||||
return "grok-3-fast"
|
||||
|
||||
@@ -43,8 +43,8 @@ class XAIModelsTest(BaseSimulatorTest):
|
||||
# Setup test files for later use
|
||||
self.setup_test_files()
|
||||
|
||||
# Test 1: 'grok' alias (should map to grok-3)
|
||||
self.logger.info(" 1: Testing 'grok' alias (should map to grok-3)")
|
||||
# Test 1: 'grok' alias (should map to grok-4)
|
||||
self.logger.info(" 1: Testing 'grok' alias (should map to grok-4)")
|
||||
|
||||
response1, continuation_id = self.call_mcp_tool(
|
||||
"chat",
|
||||
|
||||
@@ -108,9 +108,9 @@ class TestAutoModeComprehensive:
|
||||
"OPENROUTER_API_KEY": None,
|
||||
},
|
||||
{
|
||||
"EXTENDED_REASONING": "grok-3", # GROK-3 for reasoning
|
||||
"EXTENDED_REASONING": "grok-4", # GROK-4 for reasoning (now preferred)
|
||||
"FAST_RESPONSE": "grok-3-fast", # GROK-3-fast for speed
|
||||
"BALANCED": "grok-3", # GROK-3 as balanced
|
||||
"BALANCED": "grok-4", # GROK-4 as balanced (now preferred)
|
||||
},
|
||||
),
|
||||
# Both Gemini and OpenAI available - Google comes first in priority
|
||||
|
||||
@@ -320,7 +320,7 @@ class TestAutoModeProviderSelection:
|
||||
("pro", ProviderType.GOOGLE, "gemini-2.5-pro"),
|
||||
("mini", ProviderType.OPENAI, "gpt-5-mini"), # "mini" now resolves to gpt-5-mini
|
||||
("o3mini", ProviderType.OPENAI, "o3-mini"),
|
||||
("grok", ProviderType.XAI, "grok-3"),
|
||||
("grok", ProviderType.XAI, "grok-4"),
|
||||
("grokfast", ProviderType.XAI, "grok-3-fast"),
|
||||
]
|
||||
|
||||
|
||||
@@ -76,19 +76,21 @@ class TestSupportedModelsAliases:
|
||||
assert isinstance(config.aliases, list), f"{model_name} aliases must be a list"
|
||||
|
||||
# Test specific aliases
|
||||
assert "grok" in provider.SUPPORTED_MODELS["grok-3"].aliases
|
||||
assert "grok" in provider.SUPPORTED_MODELS["grok-4"].aliases
|
||||
assert "grok4" in provider.SUPPORTED_MODELS["grok-4"].aliases
|
||||
assert "grok3" in provider.SUPPORTED_MODELS["grok-3"].aliases
|
||||
assert "grok3fast" in provider.SUPPORTED_MODELS["grok-3-fast"].aliases
|
||||
assert "grokfast" in provider.SUPPORTED_MODELS["grok-3-fast"].aliases
|
||||
|
||||
# Test alias resolution
|
||||
assert provider._resolve_model_name("grok") == "grok-3"
|
||||
assert provider._resolve_model_name("grok") == "grok-4"
|
||||
assert provider._resolve_model_name("grok4") == "grok-4"
|
||||
assert provider._resolve_model_name("grok3") == "grok-3"
|
||||
assert provider._resolve_model_name("grok3fast") == "grok-3-fast"
|
||||
assert provider._resolve_model_name("grokfast") == "grok-3-fast"
|
||||
|
||||
# Test case insensitive resolution
|
||||
assert provider._resolve_model_name("Grok") == "grok-3"
|
||||
assert provider._resolve_model_name("Grok") == "grok-4"
|
||||
assert provider._resolve_model_name("GROKFAST") == "grok-3-fast"
|
||||
|
||||
def test_dial_provider_aliases(self):
|
||||
|
||||
@@ -62,7 +62,8 @@ class TestXAIProvider:
|
||||
provider = XAIModelProvider("test-key")
|
||||
|
||||
# Test shorthand resolution
|
||||
assert provider._resolve_model_name("grok") == "grok-3"
|
||||
assert provider._resolve_model_name("grok") == "grok-4"
|
||||
assert provider._resolve_model_name("grok4") == "grok-4"
|
||||
assert provider._resolve_model_name("grok3") == "grok-3"
|
||||
assert provider._resolve_model_name("grokfast") == "grok-3-fast"
|
||||
assert provider._resolve_model_name("grok3fast") == "grok-3-fast"
|
||||
@@ -106,8 +107,8 @@ class TestXAIProvider:
|
||||
provider = XAIModelProvider("test-key")
|
||||
|
||||
capabilities = provider.get_capabilities("grok")
|
||||
assert capabilities.model_name == "grok-3" # Should resolve to full name
|
||||
assert capabilities.context_window == 131_072
|
||||
assert capabilities.model_name == "grok-4" # Should resolve to full name
|
||||
assert capabilities.context_window == 256_000
|
||||
|
||||
capabilities_fast = provider.get_capabilities("grokfast")
|
||||
assert capabilities_fast.model_name == "grok-3-fast" # Should resolve to full name
|
||||
@@ -119,13 +120,15 @@ class TestXAIProvider:
|
||||
with pytest.raises(ValueError, match="Unsupported X.AI model"):
|
||||
provider.get_capabilities("invalid-model")
|
||||
|
||||
def test_no_thinking_mode_support(self):
|
||||
"""Test that X.AI models don't support thinking mode."""
|
||||
def test_thinking_mode_support(self):
|
||||
"""Test X.AI model thinking mode support - grok-4 supports it, earlier models don't."""
|
||||
provider = XAIModelProvider("test-key")
|
||||
|
||||
assert not provider.supports_thinking_mode("grok-3")
|
||||
assert not provider.supports_thinking_mode("grok-3-fast")
|
||||
assert not provider.supports_thinking_mode("grok")
|
||||
assert provider.supports_thinking_mode("grok-4") # grok-4 supports thinking mode
|
||||
assert provider.supports_thinking_mode("grok") # resolves to grok-4
|
||||
assert provider.supports_thinking_mode("grok4") # resolves to grok-4
|
||||
assert not provider.supports_thinking_mode("grokfast")
|
||||
|
||||
def test_provider_type(self):
|
||||
@@ -145,7 +148,10 @@ class TestXAIProvider:
|
||||
|
||||
# grok-3 should be allowed
|
||||
assert provider.validate_model_name("grok-3") is True
|
||||
assert provider.validate_model_name("grok") is True # Shorthand for grok-3
|
||||
assert provider.validate_model_name("grok3") is True # Shorthand for grok-3
|
||||
|
||||
# grok should be blocked (resolves to grok-4 which is not allowed)
|
||||
assert provider.validate_model_name("grok") is False
|
||||
|
||||
# grok-3-fast should be blocked by restrictions
|
||||
assert provider.validate_model_name("grok-3-fast") is False
|
||||
@@ -161,7 +167,7 @@ class TestXAIProvider:
|
||||
|
||||
provider = XAIModelProvider("test-key")
|
||||
|
||||
# Shorthand "grok" should be allowed (resolves to grok-3)
|
||||
# Shorthand "grok" should be allowed (resolves to grok-4)
|
||||
assert provider.validate_model_name("grok") is True
|
||||
|
||||
# Full name "grok-3" should NOT be allowed (only shorthand "grok" is in restriction list)
|
||||
@@ -219,6 +225,7 @@ class TestXAIProvider:
|
||||
provider = XAIModelProvider("test-key")
|
||||
|
||||
# Check that all expected base models are present
|
||||
assert "grok-4" in provider.SUPPORTED_MODELS
|
||||
assert "grok-3" in provider.SUPPORTED_MODELS
|
||||
assert "grok-3-fast" in provider.SUPPORTED_MODELS
|
||||
|
||||
@@ -234,8 +241,12 @@ class TestXAIProvider:
|
||||
assert grok3_config.supports_extended_thinking is False
|
||||
|
||||
# Check aliases are correctly structured
|
||||
assert "grok" in grok3_config.aliases
|
||||
assert "grok3" in grok3_config.aliases
|
||||
assert "grok3" in grok3_config.aliases # grok3 resolves to grok-3
|
||||
|
||||
# Check grok-4 aliases
|
||||
grok4_config = provider.SUPPORTED_MODELS["grok-4"]
|
||||
assert "grok" in grok4_config.aliases # grok resolves to grok-4
|
||||
assert "grok4" in grok4_config.aliases
|
||||
|
||||
grok3fast_config = provider.SUPPORTED_MODELS["grok-3-fast"]
|
||||
assert "grok3fast" in grok3fast_config.aliases
|
||||
@@ -246,7 +257,7 @@ class TestXAIProvider:
|
||||
"""Test that generate_content resolves aliases before making API calls.
|
||||
|
||||
This is the CRITICAL test that ensures aliases like 'grok' get resolved
|
||||
to 'grok-3' before being sent to X.AI API.
|
||||
to 'grok-4' before being sent to X.AI API.
|
||||
"""
|
||||
# Set up mock OpenAI client
|
||||
mock_client = MagicMock()
|
||||
@@ -271,15 +282,15 @@ class TestXAIProvider:
|
||||
|
||||
# Call generate_content with alias 'grok'
|
||||
result = provider.generate_content(
|
||||
prompt="Test prompt", model_name="grok", temperature=0.7 # This should be resolved to "grok-3"
|
||||
prompt="Test prompt", model_name="grok", temperature=0.7 # This should be resolved to "grok-4"
|
||||
)
|
||||
|
||||
# Verify the API was called with the RESOLVED model name
|
||||
mock_client.chat.completions.create.assert_called_once()
|
||||
call_kwargs = mock_client.chat.completions.create.call_args[1]
|
||||
|
||||
# CRITICAL ASSERTION: The API should receive "grok-3", not "grok"
|
||||
assert call_kwargs["model"] == "grok-3", f"Expected 'grok-3' but API received '{call_kwargs['model']}'"
|
||||
# CRITICAL ASSERTION: The API should receive "grok-4", not "grok"
|
||||
assert call_kwargs["model"] == "grok-4", f"Expected 'grok-4' but API received '{call_kwargs['model']}'"
|
||||
|
||||
# Verify other parameters
|
||||
assert call_kwargs["temperature"] == 0.7
|
||||
@@ -289,7 +300,7 @@ class TestXAIProvider:
|
||||
|
||||
# Verify response
|
||||
assert result.content == "Test response"
|
||||
assert result.model_name == "grok-3" # Should be the resolved name
|
||||
assert result.model_name == "grok-4" # Should be the resolved name
|
||||
|
||||
@patch("providers.openai_compatible.OpenAI")
|
||||
def test_generate_content_other_aliases(self, mock_openai_class):
|
||||
|
||||
Reference in New Issue
Block a user