1 vs 1 benchmarks won
Anthropic Claude Opus 4.7 | Google Gemini 3.1 Pro | |
|---|---|---|
| Overview | ||
| Company | Anthropic | |
| Release date | Apr 16 2026 | Feb 19 2026 |
| Model type | — | — |
| Open source | No | No |
| Specifications | ||
Parameters | — | — |
Context window | — | — |
| Benchmarks | ||
Science reasoning GPQA Diamond | 94.2% | 94.3% |
Software engineering SWE-Bench Verified | 87.6% | 80.6% |
Multimodal understanding MMMU | — | — |
| Timeline | ||
| Release gap | Gemini 3.1 Pro shipped 56 days before Claude Opus 4.7 | |
Claude Opus 4.7 and Gemini 3.1 Pro are evenly matched across the benchmarks they both publish. Gemini 3.1 Pro shipped 56 days before Claude Opus 4.7, so benchmark comparisons should account for the intervening progress.
Published specifications for these two models are limited — see each model page for the latest details.
On GPQA Diamond, Gemini 3.1 Pro scores 94.3%, 0.1 points above Claude Opus 4.7 at 94.2%. On SWE-Bench Verified, Claude Opus 4.7 scores 87.6%, 7 points above Gemini 3.1 Pro at 80.6%.