1 vs 1 benchmarks won
Anthropic Claude Opus 4.5 | Google Gemini 3.1 Pro | |
|---|---|---|
| Overview | ||
| Company | Anthropic | |
| Release date | Nov 24 2025 | Feb 19 2026 |
| Model type | — | — |
| Open source | No | No |
| Specifications | ||
Parameters | — | — |
Context window | — | — |
| Benchmarks | ||
Science reasoning GPQA Diamond | 87% | 94.3% |
Software engineering SWE-Bench Verified | 80.9% | 80.6% |
Multimodal understanding MMMU | — | — |
| Timeline | ||
| Release gap | Claude Opus 4.5 shipped 87 days before Gemini 3.1 Pro | |
Claude Opus 4.5 and Gemini 3.1 Pro are evenly matched across the benchmarks they both publish. Claude Opus 4.5 shipped 87 days before Gemini 3.1 Pro, so benchmark comparisons should account for the intervening progress.
Published specifications for these two models are limited — see each model page for the latest details.
On GPQA Diamond, Gemini 3.1 Pro scores 94.3%, 7.3 points above Claude Opus 4.5 at 87%. On SWE-Bench Verified, Claude Opus 4.5 scores 80.9%, 0.3 points above Gemini 3.1 Pro at 80.6%.