SOTAVerified

HyperTree Proof Search for Neural Theorem Proving

2022-05-23Unverified0· sign in to hype

Guillaume Lample, Marie-Anne Lachaux, Thibaut Lavril, Xavier Martinet, Amaury Hayat, Gabriel Ebner, Aurélien Rodriguez, Timothée Lacroix

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We propose an online training procedure for a transformer-based automated theorem prover. Our approach leverages a new search algorithm, HyperTree Proof Search (HTPS), inspired by the recent success of AlphaZero. Our model learns from previous proof searches through online training, allowing it to generalize to domains far from the training distribution. We report detailed ablations of our pipeline's main components by studying performance on three environments of increasing complexity. In particular, we show that with HTPS alone, a model trained on annotated proofs manages to prove 65.4% of a held-out set of Metamath theorems, significantly outperforming the previous state of the art of 56.5% by GPT-f. Online training on these unproved theorems increases accuracy to 82.6%. With a similar computational budget, we improve the state of the art on the Lean-based miniF2F-curriculum dataset from 31% to 42% proving accuracy.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
Metamath set.mmEvaristePass@3272.4Unverified
miniF2F-curriculumEvariste-7dPass@6442.5Unverified
miniF2F-curriculumGPT-fPass@6430.6Unverified
miniF2F-curriculumEvaristePass@6432.1Unverified
miniF2F-curriculumEvariste-1dPass@6433.6Unverified
miniF2F-testGPT-fcumulative36.6Unverified
miniF2F-testEvaristecumulative41Unverified
miniF2F-testEvariste-7dcumulative40.6Unverified
miniF2F-testEvariste-1dcumulative38.9Unverified
miniF2F-validGPT-fPass@6447.3Unverified
miniF2F-validEvariste-1dPass@6446.7Unverified
miniF2F-validEvariste-7dPass@6447.5Unverified
miniF2F-validEvaristePass@6458.6Unverified

Reproductions