Anthropic's Claude Opus 4.7 closes a ten-week sprint where all four major labs shipped flagship models. Each carved out distinct strengths: Opus 4.7 leads software engineering tests, GPT-5.4 dominates computer use tasks, while Gemini 3.1 Pro wins on cost and speed. The convergence on general reasoning scores masks growing specialization across specific domains.
https://www.implicator.ai/opus-4-7-jumps-11-points-on-coding-gemini-3-1-pro-still-wins-on-price/

Claude Opus 4.7 Beats GPT-5.4 and Gemini on Coding Tests
Anthropic released Claude Opus 4.7 Thursday, closing a ten-week race in which every frontier lab shipped a new flagship. Opus 4.7 wins coding and tool use. GPT-5.4 wins computer use. Gemini 3.1 Pro wins price, speed, and multimodal breadth. Four flagships, split four ways.







