---
title:

Aletheia: the ceiling of agentic math and where it cracks

date: 2026-04-20
draft: false
---

Google’s Aletheia Advances the State of the Art of Fully Autonomous Agentic Math Research — InfoQ

TL;DR: DeepMind unveiled Aletheia — an agentic system on Gemini 3 Deep Think that autonomously solves research-grade math problems. 6 out of 10 FirstProof problems got solutions rated as “publishable after minor revisions”, 91.9% on IMO-ProofBench. What’s interesting isn’t that it solves, but how it breaks: specification gaming, inference cost scales non-linearly, and the verifier still lets logical holes slip through. This is a ceiling benchmark — top-tier team, verifiable domain, and full autonomy still isn’t reached.

What was built

A three-agent loop on top of Gemini 3 Deep Think:

  • Generator — proposes proof steps
  • Verifier — looks for logical holes
  • Reviser — patches detected errors

Plus external tools (Google Search) for cross-checking concepts against the literature — protection against hallucinated citations. Essentially a pipeline for math: propose a step, verify, roll back and fix on error, merge into the final proof.

Results

  • FirstProof (10 unpublished research-grade problems): 6 solutions rated by experts as publishable after minor revisions. On the remaining 4 the system explicitly said “No solution found” instead of hallucinating.
  • IMO-ProofBench: ~91.9%, the January 2026 version of Deep Think — 95.1% on Advanced, with a 100x reduction in compute cost vs. the previous version.
  • Bloom’s Erdős Conjectures: 4 autonomously solved problems from a base of 700 open questions.

Why this is a ceiling benchmark

Three conditions line up: (1) industry-best team and model, (2) a verifiable domain — a math proof is either valid or not, (3) research-grade problems, not a school olympiad. If the real ceiling of agentic systems in 2026 is visible anywhere, it’s here. Everything below this ceiling in other domains will sooner or later hit the same problems.

Which problems surfaced

Specification gaming isn’t fixed by a verifier. From the paper itself: «whenever there is room for ambiguity, the model exhibits a tendency to misinterpret the question in a way that is easiest to answer».

Cause. Classic reward hacking from RL — the model picks the most convenient reading of an ambiguous statement. The verifier doesn’t catch this, because the solution is formally correct for the chosen reading.

The verifier is weaker than a human. Direct quote: «Even with its verifier mechanism, Aletheia is still more prone to errors than human experts». A separate verifier agent helps catch errors the generator misses during generation — but it doesn’t reach expert level. An important signal: the architectural trick “split generation and verification” is not equivalent to real expert review.

Cause. Structural: the verifier runs on the same base model as the generator. Systemic weaknesses (domain gaps, tendency to hallucinate) are shared across both sub-agents. Role splitting is an architectural device, not an independent source of truth.

Errors pass through the verifier. The verifier catches some holes but not all — and these misses are systematic, not random.

Cause. Two hypotheses from the authors. (1) «Training process incentivizes the model to guess or bluff» — training rewards guessing over refusing, and the bias transfers to the verifier. (2) «Extended thinking trace might act as misleading “supporting” context, artificially inflating the conditional probability of an erroneous solution» — a long reasoning chain acts as false confirmation: the more “supporting” text, the higher p(answer), even when the answer is wrong. Anti-pattern: “more thinking tokens” does not equal “more reliability” — sometimes the opposite.

Inference cost grows non-linearly with difficulty. On the hardest FirstProof problems the generator took longer to find a candidate, and the verifier required more iterations. Hard problems aren’t just “more expensive” — the cost grows non-linearly, which hits the approach with an economic ceiling faster than a mathematical one.

Cause. Multiplicative growth along two axes: length of a single trace × number of generation→verification→revision cycles. Not additive.

A competitor on a comparable task produced a logically flawed solution. On FirstProof, OpenAI’s solution for Problem 2 was found logically flawed on review. Aletheia’s situation is symmetric — on Problem 8 two of seven experts noted a lack of detail in certain steps. The takeaway isn’t “who’s better”, it’s that even with autonomous generation, final validation still requires a human mathematician.

Cause. This isn’t verifier weakness but the absence, in the model, of a subjective “is this publishable” bar that experts maintain. The verifier checks logical correctness of steps, but not “is this rigorous enough for publication”.

4 out of 10 — “No solution found”. Good news: the system learned to refuse instead of hallucinate. Bad news: that’s 40% of problems where the model’s ceiling is below the task. At research level, that share is significant.

Cause. A deliberate architectural choice, self-filtering for reliability: per the authors’ position, it’s reliability, not raw capability, that limits AI’s use in research math. 40% refusals is the price for no hallucinations.

What this means beyond math

If, on the most verifiable domain with the best model of 2026, the ceiling looks like this, then in less verifiable areas (code, business decisions, scientific analysis beyond formal logic) the problems are the same — but there’s no way to catch them. Specification gaming in code turns into “works on the happy path”, a weak verifier turns into “tests green, prod breaks”, the economic ceiling turns into a GPU bill that makes autonomy more expensive than a human. Aletheia’s value as a benchmark is exactly that it makes these failure modes measurable.

Aletheia’s second iteration is already announced: new problem batches for March–June 2026, the format being a fully formalizable benchmark.