docs: update advanced usage and configuration to include new GPT-5.1 models and enhance tool parameters
This commit is contained in:
@@ -4,16 +4,33 @@ This guide covers advanced features, configuration options, and workflows for po
|
||||
|
||||
## Table of Contents
|
||||
|
||||
- [Model Configuration](#model-configuration)
|
||||
- [Model Usage Restrictions](#model-usage-restrictions)
|
||||
- [Thinking Modes](#thinking-modes)
|
||||
- [Tool Parameters](#tool-parameters)
|
||||
- [Context Revival: AI Memory Beyond Context Limits](#context-revival-ai-memory-beyond-context-limits)
|
||||
- [Collaborative Workflows](#collaborative-workflows)
|
||||
- [Working with Large Prompts](#working-with-large-prompts)
|
||||
- [Vision Support](#vision-support)
|
||||
- [Web Search Integration](#web-search-integration)
|
||||
- [System Prompts](#system-prompts)
|
||||
- [Advanced Usage Guide](#advanced-usage-guide)
|
||||
- [Table of Contents](#table-of-contents)
|
||||
- [Model Configuration](#model-configuration)
|
||||
- [Model Usage Restrictions](#model-usage-restrictions)
|
||||
- [Thinking Modes](#thinking-modes)
|
||||
- [Thinking Modes \& Token Budgets](#thinking-modes--token-budgets)
|
||||
- [How to Use Thinking Modes](#how-to-use-thinking-modes)
|
||||
- [Optimizing Token Usage \& Costs](#optimizing-token-usage--costs)
|
||||
- [Tool Parameters](#tool-parameters)
|
||||
- [File-Processing Tools](#file-processing-tools)
|
||||
- [Context Revival: AI Memory Beyond Context Limits](#context-revival-ai-memory-beyond-context-limits)
|
||||
- [**The Breakthrough**](#the-breakthrough)
|
||||
- [Key Benefits](#key-benefits)
|
||||
- [Quick Example](#quick-example)
|
||||
- [Collaborative Workflows](#collaborative-workflows)
|
||||
- [Design → Review → Implement](#design--review--implement)
|
||||
- [Code → Review → Fix](#code--review--fix)
|
||||
- [Debug → Analyze → Solution → Precommit Check → Publish](#debug--analyze--solution--precommit-check--publish)
|
||||
- [Refactor → Review → Implement → Test](#refactor--review--implement--test)
|
||||
- [Tool Selection Guidance](#tool-selection-guidance)
|
||||
- [Vision Support](#vision-support)
|
||||
- [Working with Large Prompts](#working-with-large-prompts)
|
||||
- [Web Search Integration](#web-search-integration)
|
||||
- [System Prompts](#system-prompts)
|
||||
- [Prompt Architecture](#prompt-architecture)
|
||||
- [Specialized Expertise](#specialized-expertise)
|
||||
- [Customization](#customization)
|
||||
|
||||
## Model Configuration
|
||||
|
||||
@@ -41,6 +58,9 @@ Regardless of your default configuration, you can specify models per request:
|
||||
| **`o3-mini`** | OpenAI | 200K tokens | Balanced speed/quality | Moderate complexity tasks |
|
||||
| **`o4-mini`** | OpenAI | 200K tokens | Latest reasoning model | Optimized for shorter contexts |
|
||||
| **`gpt4.1`** | OpenAI | 1M tokens | Latest GPT-4 with extended context | Large codebase analysis, comprehensive reviews |
|
||||
| **`gpt5.1`** (GPT-5.1) | OpenAI | 400K tokens | Flagship reasoning model with configurable thinking effort | Complex problems, balanced agent/coding flows |
|
||||
| **`gpt5.1-codex`** (GPT-5.1 Codex) | OpenAI | 400K tokens | Agentic coding specialization (Responses API) | Advanced coding tasks, structured code generation |
|
||||
| **`gpt5.1-codex-mini`** (GPT-5.1 Codex mini) | OpenAI | 400K tokens | Cost-efficient Codex variant with streaming | Balanced coding tasks, cost-conscious development |
|
||||
| **`gpt5`** (GPT-5) | OpenAI | 400K tokens | Advanced model with reasoning support | Complex problems requiring advanced reasoning |
|
||||
| **`gpt5-mini`** (GPT-5 Mini) | OpenAI | 400K tokens | Efficient variant with reasoning | Balanced performance and capability |
|
||||
| **`gpt5-nano`** (GPT-5 Nano) | OpenAI | 400K tokens | Fastest, cheapest GPT-5 variant | Summarization and classification tasks |
|
||||
@@ -61,6 +81,10 @@ cloud models (expensive/powerful) AND local models (free/private) in the same co
|
||||
- **Flash Lite 2.0**: Text-only lightweight model (no thinking support)
|
||||
- **O3/O4 Models**: Excellent reasoning, systematic analysis, 200K context
|
||||
- **GPT-4.1**: Extended context window (1M tokens), general capabilities
|
||||
- **GPT-5.1 Series**: Latest flagship reasoning models, 400K context
|
||||
- **GPT-5.1**: Flagship model with configurable thinking effort and vision
|
||||
- **GPT-5.1 Codex**: Agentic coding specialization (Responses API, non-streaming)
|
||||
- **GPT-5.1 Codex mini**: Cost-efficient Codex variant with streaming support
|
||||
- **GPT-5 Series**: Advanced reasoning models, 400K context
|
||||
- **GPT-5**: Full-featured with reasoning support and vision
|
||||
- **GPT-5 Mini**: Balanced efficiency and capability
|
||||
@@ -161,7 +185,7 @@ All tools that work with files support **both individual files and entire direct
|
||||
**`analyze`** - Analyze files or directories
|
||||
- `files`: List of file paths or directories (required)
|
||||
- `question`: What to analyze (required)
|
||||
- `model`: auto|pro|flash|flash-2.0|flashlite|o3|o3-mini|o4-mini|gpt4.1|gpt5|gpt5-mini|gpt5-nano (default: server default)
|
||||
- `model`: auto|pro|flash|flash-2.0|flashlite|o3|o3-mini|o4-mini|gpt4.1|gpt5.1|gpt5.1-codex|gpt5.1-codex-mini|gpt5|gpt5-mini|gpt5-nano (default: server default)
|
||||
- `analysis_type`: architecture|performance|security|quality|general
|
||||
- `output_format`: summary|detailed|actionable
|
||||
- `thinking_mode`: minimal|low|medium|high|max (default: medium, Gemini only)
|
||||
@@ -176,7 +200,7 @@ All tools that work with files support **both individual files and entire direct
|
||||
|
||||
**`codereview`** - Review code files or directories
|
||||
- `files`: List of file paths or directories (required)
|
||||
- `model`: auto|pro|flash|flash-2.0|flashlite|o3|o3-mini|o4-mini|gpt4.1|gpt5|gpt5-mini|gpt5-nano (default: server default)
|
||||
- `model`: auto|pro|flash|flash-2.0|flashlite|o3|o3-mini|o4-mini|gpt4.1|gpt5.1|gpt5.1-codex|gpt5.1-codex-mini|gpt5|gpt5-mini|gpt5-nano (default: server default)
|
||||
- `review_type`: full|security|performance|quick
|
||||
- `focus_on`: Specific aspects to focus on
|
||||
- `standards`: Coding standards to enforce
|
||||
@@ -192,7 +216,7 @@ All tools that work with files support **both individual files and entire direct
|
||||
|
||||
**`debug`** - Debug with file context
|
||||
- `error_description`: Description of the issue (required)
|
||||
- `model`: auto|pro|flash|flash-2.0|flashlite|o3|o3-mini|o4-mini|gpt4.1|gpt5|gpt5-mini|gpt5-nano (default: server default)
|
||||
- `model`: auto|pro|flash|flash-2.0|flashlite|o3|o3-mini|o4-mini|gpt4.1|gpt5.1|gpt5.1-codex|gpt5.1-codex-mini|gpt5|gpt5-mini|gpt5-nano (default: server default)
|
||||
- `error_context`: Stack trace or logs
|
||||
- `files`: Files or directories related to the issue
|
||||
- `runtime_info`: Environment details
|
||||
@@ -208,7 +232,7 @@ All tools that work with files support **both individual files and entire direct
|
||||
|
||||
**`thinkdeep`** - Extended analysis with file context
|
||||
- `current_analysis`: Your current thinking (required)
|
||||
- `model`: auto|pro|flash|flash-2.0|flashlite|o3|o3-mini|o4-mini|gpt4.1|gpt5|gpt5-mini|gpt5-nano (default: server default)
|
||||
- `model`: auto|pro|flash|flash-2.0|flashlite|o3|o3-mini|o4-mini|gpt4.1|gpt5.1|gpt5.1-codex|gpt5.1-codex-mini|gpt5|gpt5-mini|gpt5-nano (default: server default)
|
||||
- `problem_context`: Additional context
|
||||
- `focus_areas`: Specific aspects to focus on
|
||||
- `files`: Files or directories for context
|
||||
@@ -224,7 +248,7 @@ All tools that work with files support **both individual files and entire direct
|
||||
**`testgen`** - Comprehensive test generation with edge case coverage
|
||||
- `files`: Code files or directories to generate tests for (required)
|
||||
- `prompt`: Description of what to test, testing objectives, and scope (required)
|
||||
- `model`: auto|pro|flash|flash-2.0|flashlite|o3|o3-mini|o4-mini|gpt4.1|gpt5|gpt5-mini|gpt5-nano (default: server default)
|
||||
- `model`: auto|pro|flash|flash-2.0|flashlite|o3|o3-mini|o4-mini|gpt4.1|gpt5.1|gpt5.1-codex|gpt5.1-codex-mini|gpt5|gpt5-mini|gpt5-nano (default: server default)
|
||||
- `test_examples`: Optional existing test files as style/pattern reference
|
||||
- `thinking_mode`: minimal|low|medium|high|max (default: medium, Gemini only)
|
||||
|
||||
@@ -239,7 +263,7 @@ All tools that work with files support **both individual files and entire direct
|
||||
- `files`: Code files or directories to analyze for refactoring opportunities (required)
|
||||
- `prompt`: Description of refactoring goals, context, and specific areas of focus (required)
|
||||
- `refactor_type`: codesmells|decompose|modernize|organization (required)
|
||||
- `model`: auto|pro|flash|flash-2.0|flashlite|o3|o3-mini|o4-mini|gpt4.1|gpt5|gpt5-mini|gpt5-nano (default: server default)
|
||||
- `model`: auto|pro|flash|flash-2.0|flashlite|o3|o3-mini|o4-mini|gpt4.1|gpt5.1|gpt5.1-codex|gpt5.1-codex-mini|gpt5|gpt5-mini|gpt5-nano (default: server default)
|
||||
- `focus_areas`: Specific areas to focus on (e.g., 'performance', 'readability', 'maintainability', 'security')
|
||||
- `style_guide_examples`: Optional existing code files to use as style/pattern reference
|
||||
- `thinking_mode`: minimal|low|medium|high|max (default: medium, Gemini only)
|
||||
|
||||
Reference in New Issue
Block a user