
Sign in to join the discussion.
It depends, almost entirely, on who is asking, what scaffolding they are running, and - as of last week - whether they have access to a model that most practitioners will never touch. That last clause is new, and it changes the terms of the entire debate.
Consider where the publicly accessible frontier stands. On SWE-bench Verified, the benchmark most widely cited as the gold standard for agentic coding ability, six production models cluster within 1.3 percentage points of each other: Claude Opus 4.6 (80.8%), Gemini 3.1 Pro (80.6%), MiniMax M2.5 (80.2%), GPT-5.4 (80.0%), Claude Sonnet 4.6 (79.6%).[1] At that level of compression, the benchmark is no longer selecting for model quality. It is selecting for frontier-viability - a binary property, not a ranking.
Then, on April 7, 2026, Anthropic published a system card for Claude Mythos Preview, a model it has decided not to release to the general public.[2] Mythos scores 93.9% on SWE-bench Verified - a 13.1-point leap over the cluster above, the highest score ever recorded on the benchmark. It is not available via Claude.ai or the public API. The leaderboard the industry has been debating for months is, in a meaningful sense, already obsolete.
The benchmark racket is now running at two levels simultaneously. At the first level: six models tied within noise, each lab declaring victory. At the second: a model that has genuinely broken out of the cluster is being withheld, its scores published in a 244-page system card that most practitioners will not read. The industry is arguing about the wrong leaderboard.
Benchmark convergence at the frontier is not an accident. It is the predictable outcome of a system in which evaluation sets are published, training pipelines are massive, and every lab competes on the same handful of tests. The ARC Prize's ARC-AGI-2 benchmark was designed specifically to resist this dynamic - it tests novel rule-induction under conditions meant to defeat memorization. Yet when Gemini 3.1 Pro launched in February 2026 with a score of 77.1%, independently verified by the ARC Prize Foundation, GPT-5.4 arrived three weeks later reporting 74.0% on the standard variant and 83.3% on its extended-reasoning Pro configuration.[3][4] The benchmark designed to be contamination-proof has become a marketing battleground within a single product cycle.
The verification problem is structural. On GPQA Diamond, Artificial Analysis's independent evaluation places Gemini 3.1 Pro at 94.1% and GPT-5.4 at 92.0%.[5] Self-reported lab figures have historically run higher: Gemini's own model card reports 94.3% - close enough to be within noise - while GPT-5.4 Pro's claimed 94.4% has not been independently verified in a like-for-like configuration. The directional incentive for labs to evaluate themselves favorably is structural, regardless of whether each individual gap is large or small. Mythos Preview, now the top-scoring model on GPQA Diamond at 94.6%, is not available for independent verification at all.[2]
OpenAI's decision not to submit GPT-5.4 to the official SWE-bench Verified leaderboard - citing training data contamination concerns - is the most candid acknowledgment yet that the labs themselves no longer fully trust the benchmarks they compete on.[1] That admission deserves more attention than it has received.
SWE-bench Pro makes the harness argument unavoidable. Scale AI runs all submissions under a standardized SWE-Agent scaffold with a 250-turn limit - the closest thing the field has to a controlled experiment. Take a single frontier model and run it twice: once with a minimal SWE-Agent configuration and once with an optimized scaffold. The minimal run scores around 23%. The optimized run: 45% or better.[1] That 22-point delta is larger than the cumulative gap separating any two publicly available frontier models on SWE-bench Verified.
Consider what this implies for resource allocation. The entire competitive conversation about whether to use GPT-5.4 versus Opus 4.6 concerns a performance difference of under one percentage point on the industry's most-cited standard. A team that spent the same effort improving its orchestration layer - context retrieval strategy, retry logic, tool routing - would recover more than twenty points on the same test. The industry has inverted the priority ordering.
Anthropic's Claude Code puts numbers to the gap. Running on Anthropic's own agent scaffolding, it reaches 80.9% on SWE-bench, edging just past raw Opus 4.6's 80.8%.[1] The underlying model weights are identical; what changed is the surrounding engineering: how tools are selected, when the model retries, how context is windowed across long runs. The lesson is that model capability and deployed system performance are not the same quantity - and the gap between them is substantially under the practitioner's control. Mythos Preview's 77.8% on SWE-bench Pro, compared to GPT-5.4's 57.7% on the same standardized scaffold, suggests that when a model's raw capability is genuinely superior, the benchmark registers it clearly.[2] The problem is not that SWE-bench Pro can't detect real differences - it's that the publicly available models are genuinely too close to differentiate.
None of this means the publicly available models are identical. There are real performance gaps on specific axes - they are just not the axes the benchmark headlines tend to emphasize. And Mythos Preview's system card, read carefully, illuminates which gaps are real versus which are measurement artifacts.
On expert judgment of open-ended professional output, the gaps are large and persistent. Artificial Analysis's GDPval-AA evaluation - blind pairwise comparisons by domain professionals across 44 occupational categories - scores GPT-5.4 at 1,675 Elo, Claude Opus 4.6 at 1,616, and Gemini 3.1 Pro at 1,318.[5] That 357-point gap between GPT-5.4 and Gemini 3.1 Pro on professional judgment is much harder to dismiss than a 0.2-point SWE-bench delta. Notably, Claude Sonnet 4.6 sits at 1,649 Elo - above Opus 4.6 - a ranking inversion that the SWE-bench narrative entirely obscures. Mythos Preview has no published GDPval-AA score; it is not available for independent evaluation by design.
The gap between GPT-5.4 and Gemini 3.1 Pro on SWE-bench Verified: 0.2 points. The gap in expert judgment Elo: 357 points. The industry talks about the first number. The second is the one practitioners should care about.
On reasoning under novel conditions, Mythos Preview's USAMO 2026 score of 97.6% - against Opus 4.6's 42.3%, GPT-5.4's 95.2%, and Gemini 3.1 Pro's 74.4% - suggests the public frontier has a genuine ceiling on mathematical reasoning that no harness engineering will raise.[2] For publicly available models, ARC-AGI-2's independently verified data still shows Gemini 3.1 Pro leading GPT-5.4's standard variant - 77.1% against 74.0% - though GPT-5.4 Pro's lab-reported 83.3% clouds the picture when extended reasoning is factored in.[3][4]
On terminal-heavy DevOps workflows, the rankings have shifted materially. Mythos Preview scores 82.0% on Terminal-Bench 2.0 - reaching 92.1% with extended timeouts - while among publicly available models GPT-5.4 holds a real lead at 75.1% against Gemini's 68.5% and Opus 4.6's 65.4%.[2] For teams operating under the public tier, GPT-5.4's Terminal-Bench advantage remains meaningful. For teams with Glasswing access, the picture looks entirely different.
On computer use, Mythos Preview now leads all published scores at 79.6% on OSWorld-Verified, followed by GPT-5.4 at 75.0% and Opus 4.6 at 72.7% - both above the 72.4% human baseline.[2] Gemini 3.1 Pro has no published OSWorld score. Among publicly available models, this is now a performance gradient rather than a binary capability difference - all three leading models have the capability, and GPT-5.4 leads Opus 4.6 by 2.3 points.
Pricing remains the most underreported axis of divergence. Claude Opus 4.6 costs $5.00 per million input tokens and $25.00 per million output; GPT-5.4 runs $2.50/$15.00; Gemini 3.1 Pro sits at $2.00/$12.00.[6][7][8] For agentic pipelines generating high output volumes, the gap between Opus and Gemini compounds dramatically at scale. The open-weight picture is more striking still: MiniMax M2.5 delivers 80.2% SWE-bench Verified at $0.30/$1.20 per million tokens, within 0.6 points of the closed-model leader at roughly one-twentieth the output price.[1] Mythos Preview, for the access-restricted organizations that can reach it, is priced at $25/$125 per million tokens - five times Opus 4.6, and not generally available at any price.[2]
The leaderboard cycle implicitly teaches practitioners to ask the wrong questions. Here is the set of questions that actually predict outcomes.
Where does my orchestration layer stand? Before evaluating any model, assess the maturity of the scaffold around it. A 22-point improvement is available to any team willing to invest in context retrieval, retry logic, and tool routing - no model switch required. That return dwarfs anything available from frontier-to-frontier model shopping among publicly available options.
Does my task have a verifiable ground truth? Frontier models are genuinely interchangeable for tasks scored against a reference answer - code that either runs or doesn't, math that either checks out or doesn't. For tasks requiring domain judgment - drafting a strategic memo, synthesizing conflicting expert opinions, producing financial model narratives - the GDPval-AA Elo gap between GPT-5.4 and Gemini 3.1 Pro is large and practitioner-relevant.
How much of your context window do you actually use? Both Gemini 3.1 Pro and GPT-5.4 nominally support 1 million tokens, but independent long-context retrieval data (MRCR v2 pointwise at 1M tokens) shows Gemini at 26.3%[3] while GPT-5.4's own evals show retrieval accuracy dropping from 97% at 16-32K to roughly 57% at 256-512K tokens.[4] Mythos Preview's GraphWalks BFS score at 256K-1M tokens - 80.0% against Opus 4.6's 38.7% and GPT-5.4's 21.4% - suggests the long-context gap between Mythos and the public tier is the largest single capability differential in the system card.[2] For workloads that genuinely saturate long context, Gemini's degradation profile among public models is shallower; Mythos is in a separate category entirely.
Do you need desktop automation? Among publicly available models, this is now a three-way race rather than a GPT-5.4-only capability. GPT-5.4 leads at 75.0% on OSWorld-Verified, Opus 4.6 sits at 72.7%, and Gemini 3.1 Pro has no published score. For teams with Glasswing access, Mythos at 79.6% leads the field. The decision is now about performance gradient and pricing, not binary capability.
What is your actual cost per completed task? Token-level pricing is a poor proxy for task economics. A model that reaches a correct answer in fewer tokens can be cheaper despite a higher headline rate. The right unit of analysis is task completion cost on your specific workload - not per-million-token list prices.
The frontier model race, as currently conducted and covered, is operating on two tracks that almost never get discussed together. On the public track: six models tied within 1.3 points on SWE-bench Verified, each declared a winner in turn, the press cycle repeating. On the private track: a model that scored 93.9% on the same benchmark is being used to find zero-day vulnerabilities in every major operating system, restricted to 52 organizations, and priced out of reach for all but the best-capitalized teams.
The irony is that Mythos Preview's existence actually strengthens the benchmark-skeptic case rather than undermining it. When a model with genuine, large capability gains arrives, SWE-bench Pro registers it clearly: Mythos at 77.8% versus GPT-5.4 at 57.7% is a 20-point gap that no one disputes. The benchmark works fine when the capability difference is real. The problem is that among publicly available models, the capability differences are not large enough to clear the benchmark's noise floor - and the industry has been treating noise as signal.
Which raises the question the field is not asking: what happens when the noise floor itself rises? The public cluster sitting at 80% on SWE-bench Verified will not stay there. It will inch toward 90%, then toward 95%, compressing further as more models train against the same evaluation sets. The ARC Prize response - design harder benchmarks - has worked once, but it depends on evaluation designers staying permanently ahead of the training pipelines they are trying to test. That is a race with no guaranteed winner. The more troubling trajectory is the one Mythos already illustrates: the genuinely capable models exit the public leaderboard entirely, their scores knowable only through system cards written by the organizations that built them. At that point, independent evaluation becomes structurally impossible, not merely difficult. "Trust the system card" stops being a provisional workaround and becomes the only epistemology available. The field has not reckoned with what that means for accountability, for procurement decisions, or for the broader project of understanding what these systems can actually do.
For practitioners operating on the public tier, the most useful posture remains the same: treat published benchmarks as a coarse filter for frontier-viability, invest the engineering time that would have gone to model selection into harness quality instead, and choose the model whose pricing and specific capability profile fits the actual workload. For the small number of organizations with Glasswing access, the calculus is different - but so is the price tag, and so is the use case.
The leaderboard will not tell practitioners which model to use. It will tell them which labs have the best PR operations. Those are not the same thing.
Morph: Best AI for Coding (2026) - SWE-bench Verified, SWE-bench Pro, Terminal-Bench rankings (March 2026) Inline ↗
Anthropic: Claude Mythos Preview System Card - benchmark scores, Project Glasswing, pricing, safety evaluations (April 7, 2026) Inline ↗
Google DeepMind: Gemini 3.1 Pro Model Card - ARC-AGI-2, MRCR v2, benchmark data (February 2026) Inline ↗
OpenAI: Introducing GPT-5.4 - ARC-AGI-2, OSWorld-Verified, SWE-bench Pro, pricing (March 2026) Inline ↗
Artificial Analysis: GDPval-AA Elo Leaderboard - independent expert judgment evaluation (April 2026) Inline ↗
Google AI: Gemini Developer API Pricing - Gemini 3.1 Pro input/output rates (April 2026) Inline ↗
OpenAI: API Pricing - GPT-5.4 input/output rates (April 2026) Inline ↗
Anthropic: API Pricing - Claude Opus 4.6 input/output rates (April 2026) Inline ↗