AI proved a physics formula humans couldn’t… in 12 hours

Source: AI

GPT-5.2 Pro just conjectured a compact gluon amplitude formula in particle physicsโ€”then proved it in 12 hours. Experts verified it. That’s not autocomplete. That’s pattern-finding at the frontier of science.

And here’s the structural shift: the AI didn’t need to run full simulations to get there. It predicted what mattered, then worked backward.

The industry just spent $650 billion betting you need massive compute to understand models. A physics preprint and a structural prediction technique suggest the opposite might be true.

AI conjectured physics. Humans proved it was right.

The February 2026 preprint credits GPT-5.2 Pro with proposing a formula for gluon scattering amplitudesโ€”something particle physicists couldn’t deriveโ€”then autonomously writing the formal proof. The catch: it only applies in half-collinear regimes. But the method generalizes. “AI suggests; experts prove; the community reviews,” the authors wrote. This isn’t brute-force calculation. It’s conjecture.

The same logic applies to mechanistic interpretability. You don’t need to run every activation to understand which connections matter. Structural propertiesโ€”spectral concentration, downstream path weightโ€”can predict edge importance without touching the model. That’s the efficiency bet mechanistic interpretability researchers are now making, whether they admit it or not.

The $650B compute bet assumed you had to run everything

But V-JEPA 2 hit 65-80% pick-and-place success after 62 hours of robot data in January 2026. Claude Opus 4.5 reached 5 hours of autonomous tasks by late 2025โ€”up from GPT-2’s 3 seconds in 2019. Edge AI markets exploded in 2026 on small-model efficiency, proving structural prediction beats activation analysis for cost-conscious teams.

The pattern: what matters is knowing *which* parts of the network do the work. Not running all of them.

Cost comparison: running full interpretability on GPT-4-scale models takes weeks of compute. Predicting edge importance takes minutes. “The research-to-production gap is closing,” Adaline AI Labs noted in 2026. For whom? Teams that can’t afford OpenAI’s infrastructure budget. And that’s most teams.

The shift to efficient pattern-finding over brute compute extends beyond interpretability. It’s rewriting who gets to do AI research.

The technique worksโ€”until it doesn’t

Honest limitation: the gluon formula only applies in half-collinear regimes. Structural prediction methods work for specific network topologiesโ€”they overlook activation dynamics. MIT Technology Review called mechanistic interpretability a 2026 Breakthrough Technology, noting priorities: “clarifying concepts, setting better benchmarks, scaling techniques.” Translation: the field knows current methods don’t generalize.

Anthropic open-sourced an attribution graph tool in May 2025. Knowledge editing via ROME shows practical applications. But safety-critical systems? You still need full causal analysis. The structural prediction approach won’t replace that.

It’ll make the 80% of interpretability work that doesn’t need perfection radically cheaper. That’s enough to shift who can afford to participate.

The industry bet on scale. A physics preprint proved efficiency. Both workโ€”but only one fits in a university budget, and that determines who gets to develop AI research skills in 2027.

alex morgan
I write about artificial intelligence as it shows up in real life โ€” not in demos or press releases. I focus on how AI changes work, habits, and decision-making once itโ€™s actually used inside tools, teams, and everyday workflows. Most of my reporting looks at second-order effects: what people stop doing, what gets automated quietly, and how responsibility shifts when software starts making decisions for us.