doc: updated docs

This commit is contained in:
Fahad
2025-11-18 20:38:35 +04:00
parent d3de61f878
commit 52c4563733
7 changed files with 26 additions and 24 deletions

View File

@@ -33,7 +33,7 @@ Regardless of your default configuration, you can specify models per request:
| Model | Provider | Context | Strengths | Auto Mode Usage |
|-------|----------|---------|-----------|------------------|
| **`pro`** (Gemini 2.5 Pro) | Google | 1M tokens | Extended thinking (up to 32K tokens), deep analysis | Complex architecture, security reviews, deep debugging |
| **`pro`** (Gemini 3.0 Pro) | Google | 1M tokens | Extended thinking (up to 32K tokens), deep analysis | Complex architecture, security reviews, deep debugging |
| **`flash`** (Gemini 2.5 Flash) | Google | 1M tokens | Ultra-fast responses with thinking | Quick checks, formatting, simple analysis |
| **`flash-2.0`** (Gemini 2.0 Flash) | Google | 1M tokens | Latest fast model with audio/video support | Quick analysis with multimodal input |
| **`flashlite`** (Gemini 2.0 Flash Lite) | Google | 1M tokens | Lightweight text-only model | Fast text processing without vision |
@@ -58,7 +58,7 @@ cloud models (expensive/powerful) AND local models (free/private) in the same co
**Model Capabilities:**
- **Gemini Models**: Support thinking modes (minimal to max), web search, 1M context
- **Pro 2.5**: Deep analysis with max 32K thinking tokens
- **Pro 3.0**: Deep analysis with max 32K thinking tokens
- **Flash 2.5**: Ultra-fast with thinking support (24K thinking tokens)
- **Flash 2.0**: Latest fast model with audio/video input (24K thinking tokens)
- **Flash Lite 2.0**: Text-only lightweight model (no thinking support)
@@ -107,7 +107,7 @@ OPENAI_ALLOWED_MODELS=o3,o4-mini
### Thinking Modes & Token Budgets
These only apply to models that support customizing token usage for extended thinking, such as Gemini 2.5 Pro.
These only apply to models that support customizing token usage for extended thinking, such as Gemini 3.0 Pro.
| Mode | Token Budget | Use Case | Cost Impact |
|------|-------------|----------|-------------|
@@ -155,7 +155,7 @@ These only apply to models that support customizing token usage for extended thi
# Complex debugging, letting claude pick the best model
"Use zen to debug this race condition with max thinking mode"
# Architecture analysis with Gemini 2.5 Pro
# Architecture analysis with Gemini 3.0 Pro
"Analyze the entire src/ directory architecture with high thinking using pro"
```
@@ -346,7 +346,7 @@ To help choose the right tool for your needs:
The Zen MCP server supports vision-capable models for analyzing images, diagrams, screenshots, and visual content. Vision support works seamlessly with all tools and conversation threading.
**Supported Models:**
- **Gemini 2.5 Pro & Flash**: Excellent for diagrams, architecture analysis, UI mockups (up to 20MB total)
- **Gemini 3.0 Pro & Flash**: Excellent for diagrams, architecture analysis, UI mockups (up to 20MB total)
- **OpenAI O3/O4 series**: Strong for visual debugging, error screenshots (up to 20MB total)
- **Claude models via OpenRouter**: Good for code screenshots, visual analysis (up to 5MB total)
- **Custom models**: Support varies by model, with 40MB maximum enforced for abuse prevention