Anthropic · Closed-API Model
Claude Opus 4.7: Adaptive Thinking, Tested
Claude Opus 4.7 (April 16, 2026) is Anthropic's flagship reasoning model and the first model with Adaptive Thinking — a mode that auto-tunes thinking compute per request based on task difficulty. Scores 87.6% on SWE-Bench Verified Adaptive harness, 200K context, $15/$75 per million tokens. This review covers when Opus 4.7 is worth the 5× premium over Sonnet 5, how Adaptive Thinking actually works, and which open-weight models come closest if you need local deep reasoning.
Note: Opus 4.7 is API-only — cannot be downloaded or run locally. For deep reasoning you can self-host, see DeepSeek V4-Pro, Kimi K2.6, and GLM-5.
Key takeaways
- →Adaptive Thinking — model auto-tunes reasoning compute per request, no manual tier selection.
- →87.6% SWE-Bench Verified Adaptive — strong, but Sonnet 5 still leads at 92.4% on Verified.
- →$15/$75 per Mtok — 5× more than Sonnet 5; reserve for hardest 10% of tasks.
- →200K context — same as Sonnet 5; works with prompt caching.
- →Best for: research-grade analysis, novel algorithm design, complex multi-step reasoning.
Quick verdict
Use Claude Sonnet 5 for ~90% of coding work — it's 5× cheaper, faster, and actually scores higher on standard SWE-Bench. Reach for Opus 4.7 only when the task visibly defeats Sonnet 5: complex multi-file refactors, novel algorithm design, deep research analysis, or anywhere Adaptive Thinking earns its keep.
For privacy-required deep-reasoning workloads where you need local hosting, DeepSeek V4-Pro (8× H100) or Kimi K2.6 (also 8× H100) are the realistic alternatives.
Specs at a glance
| Vendor | Anthropic |
| Release date | April 16, 2026 |
| Model ID | claude-opus-4-7-20260416 |
| Architecture | Dense transformer with Adaptive Thinking |
| Context window | 200,000 tokens |
| Max output | 64,000 tokens (excluding thinking trace) |
| Modalities | Text · Code · Vision |
| License | Proprietary |
| Local self-hostable? | No |
Adaptive Thinking explained
Most reasoning models (OpenAI o-series, Gemini 3.1 Pro thinking tiers) make you choose how much thinking to apply per request. Opus 4.7 changes this — the model decides for itself based on task difficulty.
Mechanism: a lightweight pre-pass classifies the input difficulty, then the model spends correspondingly more compute thinking before answering. Easy questions (e.g., “what's 2+2”) get fast responses with no thinking. Medium complexity gets a brief thinking pass. Hard problems trigger extended reasoning over seconds to minutes.
Why it matters: removes the cognitive overhead of picking a thinking tier. For dev tools and agents, Adaptive Thinking means you set the model once and it auto-routes per-task complexity. Cost implication: output tokens include the thinking trace, so a hard problem can multiply effective cost by 2-4× — but you only pay it when needed.
Compare to Gemini 3.1 Pro's explicit Tier 1/2/3 system or GPT-5.5's Instant/Standard/Pro split. Opus 4.7's approach is more elegant but less predictable for cost forecasting.
Benchmarks
| Benchmark | Opus 4.7 | Sonnet 5 | GPT-5.5 Pro | Gemini 3.1 Pro |
|---|---|---|---|---|
| SWE-Bench Verified Adaptive | 87.6% | 92.4% | 88.4% | 87.9% |
| GPQA Diamond (PhD science) | 87.3% | 85.7% | 87.6% | 88.2% |
| AIME 2025 (math) | 92.8% | 91.5% | 96.4% | 94.0% |
| ARC-AGI-2 (reasoning) | 71.8% | 68.4% | 73.5% | 77.1% |
| MMLU-Pro | 89.4% | 87.9% | 90.6% | 89.4% |
Pricing & access
API
- Input: $15.00 per 1M tokens
- Output (incl. thinking): $75.00 per 1M tokens
- Cached input: $1.50 per 1M (90% off)
- Batch: 50% off (24h SLA)
Subscription
- Claude Pro: $20/mo — limited Opus 4.7
- Claude Max: $100-200/mo — much higher Opus quota
- Bedrock / Vertex: Same per-token pricing
When to pick Opus 4.7
- ✓Hardest 5-10% of tasks where Sonnet 5 visibly struggles.
- ✓Research-grade analysis, novel algorithm design, deep multi-step reasoning.
- ✓Workloads where Adaptive Thinking's auto-routing eliminates manual tier selection.
- ✓Anthropic-stack teams already using Sonnet 5; Opus 4.7 is the natural escalation.
FAQ
What is Claude Opus 4.7?
Opus 4.7 vs Sonnet 5: when do I use each?
What is Adaptive Thinking?
How much does Claude Opus 4.7 cost?
Can I run Claude Opus 4.7 locally?
Opus 4.7 vs GPT-5.5 Pro: which is better for hard reasoning?
Related models
- → Claude Sonnet 5 — your default for ~90% of coding work
- → Claude Opus 4.1 — predecessor; lower benchmark scores
- → GPT-5.5 Pro — closest closed alternative; tied on most reasoning
- → Gemini 3.1 Pro — explicit thinking tiers + 1M context
- → DeepSeek V4-Pro — open-weight deep-reasoning alternative
- → Best AI models May 2026 — pillar comparison