>The results of this paper should not be interpreted as suggesting that AI can consistently solve
research-level mathematics questions. In fact, our anecdotal experience is the opposite: success cases
are rare, and an apt intuition for autonomous capabilities (and limitations) may currently be important
for finding such cases. The papers (ACGKMP26; Feng26; LeeSeo26) grew out of spontaneous positive
outcomes in a wider benchmarking effort on research-level problems; for most of these problems, no
autonomous progress was made.
the idea of autonomous math is fascinating because it implies a shift from search to verification. if an ai can traverse the proof space faster than a human, the bottleneck becomes checking the work.
from a risk perspective this feels safer than autonomous code generation. a bad math proof is just invalid (provably false). a bad code snippet is a vulnerability. math has a built-in truth layer that software engineering often lacks.
I still don't get how achieving 96% on some benchmark means it's a super genius but that last 4% is somehow still out of reach. The people who constantly compare robots to people should really ponder how a person who manages to achieve 90% on some advanced math benchmark still misses that last 10% somehow.