How could a Superhuman AI mathematician come about?

-
Sanjeev Arora, Princeton
Fine Hall 314

Can AI systems exceed the capabilities of the human experts who provided their training data? The talk will examine the hypothesis of AI selfimprovement, involving mechanisms such as synthetic data generation, reinforcement learning, and toolaugmented reasoning with formal verification loops.

I will also present recent work at Princeton, including the Gödel Prover V2 for Leanbased theorem proving and a new inference pipeline that achieved stateoftheart performance (at the time of evaluation) on IMOProofBench (Advanced) at moderate inference costs ($20$30 per problem). These will illustrate how AI systems are sometimes able to escape cognitive wells”—local optima in a models reasoning capabilities. While providing evidence for the feasibility of selfimprovement, they also highlight important hurdles and open questions.