The gemini CLI was returning plain text instead of JSON, causing clink to fail with:
'Failed to decode Gemini CLI JSON output: Expecting value: line 1 column 1 (char 0)'
Root cause: gemini.json was missing the '-o json' flag in additional_args.
This fix adds '-o json' to ensure JSON output format, aligning with the codex pattern (which uses --json).
Tested with gemini CLI v0.8.2
Stay in codex, plan review and fix complicated bugs, then ask it to spawn claude code and implement the plan.
This uses your current subscription instead of API tokens.
Stay in codex, plan review and fix complicated bugs, then ask it to spawn claude code and implement the plan.
This uses your current subscription instead of API tokens.
The previous commit (f98046c) added shutil.which() to resolve executables,
which broke two tests that only mocked subprocess execution. This commit
adds shutil.which() mocking to both test files to restore test compatibility.
Co-authored-by: Claude <noreply@anthropic.com>
model definitions now support a new `allow_code_generation` flag, only to be used with higher reasoning models such as GPT-5-Pro and-Gemini 2.5-Pro
When `true`, the `chat` tool can now request the external model to generate a full implementation / update / instructions etc and then share the implementation with the calling agent.
This effectively allows us to utilize more powerful models such as GPT-5-Pro to generate code for us or entire implementations (which are either API-only or part of the $200 Pro plan from within the ChatGPT app)
Adds configuration for the new GPT-5 Pro model released by OpenAI.
Specifications from official docs:
- 400K context window
- 272K max output tokens (largest of GPT-5 family)
- Highest reasoning capability (5/5)
- Reasoning token support
- Vision input support (text+image input, text output only)
- No temperature support (fixed, like other reasoning models)
- Available via Responses API only (use_openai_response_api: true)
- Default reasoning effort: high
Accessible via alias 'gpt5pro'.