- Add MAX_OUTPUT_TOKENS constant set to 32,768 (Gemini 2.5 Pro's limit) - Update all tools and chat handler to use MAX_OUTPUT_TOKENS - Add comprehensive tests for output token configuration - Update README with configuration details and system prompt docs This fixes the issue where Gemini responses were being cut off at 8192 tokens, causing Claude to repeatedly ask for the same analysis. Fixes #1 🤖 Generated with [Claude Code](https://claude.ai/code) Co-Authored-By: Claude <noreply@anthropic.com>
44 lines
1.5 KiB
Python
44 lines
1.5 KiB
Python
"""
|
|
Tests for configuration
|
|
"""
|
|
|
|
from config import (DEFAULT_MODEL, MAX_CONTEXT_TOKENS, MAX_OUTPUT_TOKENS,
|
|
TEMPERATURE_ANALYTICAL, TEMPERATURE_BALANCED,
|
|
TEMPERATURE_CREATIVE, TOOL_TRIGGERS, __author__,
|
|
__updated__, __version__)
|
|
|
|
|
|
class TestConfig:
|
|
"""Test configuration values"""
|
|
|
|
def test_version_info(self):
|
|
"""Test version information"""
|
|
assert __version__ == "2.4.0"
|
|
assert __author__ == "Fahad Gilani"
|
|
assert __updated__ == "2025-06-08"
|
|
|
|
def test_model_config(self):
|
|
"""Test model configuration"""
|
|
assert DEFAULT_MODEL == "gemini-2.5-pro-preview-06-05"
|
|
assert MAX_CONTEXT_TOKENS == 1_000_000
|
|
assert MAX_OUTPUT_TOKENS == 32_768
|
|
|
|
def test_temperature_defaults(self):
|
|
"""Test temperature constants"""
|
|
assert TEMPERATURE_ANALYTICAL == 0.2
|
|
assert TEMPERATURE_BALANCED == 0.5
|
|
assert TEMPERATURE_CREATIVE == 0.7
|
|
|
|
def test_tool_triggers(self):
|
|
"""Test tool trigger phrases"""
|
|
assert "think_deeper" in TOOL_TRIGGERS
|
|
assert "review_code" in TOOL_TRIGGERS
|
|
assert "debug_issue" in TOOL_TRIGGERS
|
|
assert "analyze" in TOOL_TRIGGERS
|
|
|
|
# Check some specific triggers
|
|
assert "ultrathink" in TOOL_TRIGGERS["think_deeper"]
|
|
assert "extended thinking" in TOOL_TRIGGERS["think_deeper"]
|
|
assert "find bugs" in TOOL_TRIGGERS["review_code"]
|
|
assert "root cause" in TOOL_TRIGGERS["debug_issue"]
|