SOTAVerified

Arithmetic Reasoning

Papers

Showing 150 of 175 papers

TitleStatusHype
Qwen2 Technical ReportCode13
DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language ModelsCode9
Llama 2: Open Foundation and Fine-Tuned Chat ModelsCode8
LLaMA: Open and Efficient Foundation Language ModelsCode7
Sparks of Artificial General Intelligence: Early experiments with GPT-4Code6
Mistral 7BCode6
GPT-4 Technical ReportCode6
Tree of Thoughts: Deliberate Problem Solving with Large Language ModelsCode5
WizardMath: Empowering Mathematical Reasoning for Large Language Models via Reinforced Evol-InstructCode5
ReFT: Representation Finetuning for Language ModelsCode5
OpenMathInstruct-2: Accelerating AI for Math with Massive Open-Source Instruction DataCode4
OpenMathInstruct-1: A 1.8 Million Math Instruction Tuning DatasetCode4
ToRA: A Tool-Integrated Reasoning Agent for Mathematical Problem SolvingCode3
WorkArena++: Towards Compositional Planning and Reasoning-based Common Knowledge Work TasksCode3
LLM-Adapters: An Adapter Family for Parameter-Efficient Fine-Tuning of Large Language ModelsCode3
Llemma: An Open Language Model For MathematicsCode3
Step-DPO: Step-wise Preference Optimization for Long-chain Reasoning of LLMsCode3
Reasoning with Language Model Prompting: A SurveyCode3
PAL: Program-aided Language ModelsCode3
DART-Math: Difficulty-Aware Rejection Tuning for Mathematical Problem-SolvingCode2
MetaMath: Bootstrap Your Own Mathematical Questions for Large Language ModelsCode2
Large Language Models are Zero-Shot ReasonersCode2
Is ChatGPT a General-Purpose Natural Language Processing Task Solver?Code2
CAPO: Cost-Aware Prompt OptimizationCode2
Encouraging Divergent Thinking in Large Language Models through Multi-Agent DebateCode2
An Empirical Study of Data Ability Boundary in LLMs' Math ReasoningCode2
Solving Quantitative Reasoning Problems with Language ModelsCode2
MathCoder: Seamless Code Integration in LLMs for Enhanced Mathematical ReasoningCode2
MuggleMath: Assessing the Impact of Query and Response Augmentation on Math ReasoningCode2
Scaling Relationship on Learning Mathematical Reasoning with Large Language ModelsCode2
Parameter-Efficient Sparsity Crafting from Dense to Mixture-of-Experts for Instruction Tuning on General TasksCode2
Progressive-Hint Prompting Improves Reasoning in Large Language ModelsCode2
Solving Challenging Math Word Problems Using GPT-4 Code Interpreter with Code-based Self-VerificationCode2
Chain of Preference Optimization: Improving Chain-of-Thought Reasoning in LLMsCode2
Breaking the Ceiling of the LLM Community by Treating Token Generation as a Classification for EnsemblingCode2
Boosting Language Models Reasoning with Chain-of-Knowledge PromptingCode1
Math-Shepherd: Verify and Reinforce LLMs Step-by-step without Human AnnotationsCode1
An Investigation of Neuron Activation as a Unified Lens to Explain Chain-of-Thought Eliciting Arithmetic Reasoning of LLMsCode1
Batch Prompting: Efficient Inference with Large Language Model APIsCode1
Language Imbalance Driven Rewarding for Multilingual Self-improvingCode1
Automatic Prompt Augmentation and Selection with Chain-of-Thought from Labeled DataCode1
Large Language Models are Better Reasoners with Self-VerificationCode1
Fed-SB: A Silver Bullet for Extreme Communication Efficiency and Performance in (Private) Federated LoRA Fine-TuningCode1
Automatic Model Selection with Large Language Models for ReasoningCode1
Is the Reversal Curse a Binding Problem? Uncovering Limitations of Transformers from a Basic Generalization FailureCode1
Large Language Models Can Be Easily Distracted by Irrelevant ContextCode1
FedEx-LoRA: Exact Aggregation for Federated and Efficient Fine-Tuning of Foundation ModelsCode1
Hierarchical Prompting Taxonomy: A Universal Evaluation Framework for Large Language Models Aligned with Human Cognitive PrinciplesCode1
Achieving >97% on GSM8K: Deeply Understanding the Problems Makes LLMs Better Solvers for Math Word ProblemsCode1
HALO: Hierarchical Autonomous Logic-Oriented Orchestration for Multi-Agent LLM SystemsCode1
Show:102550
← PrevPage 1 of 4Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Claude 3.5 Sonnet (HPT)Accuracy97.72Unverified
2DUP prompt upon GPT-4Accuracy97.1Unverified
3Qwen2-Math-72B-Instruct (greedy)Accuracy96.7Unverified
4SFT-Mistral-7B (Metamath, OVM, Smart Ensemble)Accuracy96.4Unverified
5OpenMath2-Llama3.1-70B (majority@256)Accuracy96Unverified
6Jiutian-大模型Accuracy95.2Unverified
7DAMOMath-7B(MetaMath, OVM, BS, Ensemble)Accuracy95.1Unverified
8Claude 3 Opus (0-shot chain-of-thought)Accuracy95Unverified
9OpenMath2-Llama3.1-70BAccuracy94.9Unverified
10GPT-4 (Teaching-Inspired)Accuracy94.8Unverified
#ModelMetricClaimedVerifiedStatus
1Text-davinci-002 (175B)(zero-shot-cot)Accuracy78.7Unverified
2Text-davinci-002 (175B) (zero-shot)Accuracy17.7Unverified
#ModelMetricClaimedVerifiedStatus
1Tree of Thoughts (b=5)Success0.74Unverified
#ModelMetricClaimedVerifiedStatus
1GPT-4 (Teaching-Inspired)Accuracy92.2Unverified
#ModelMetricClaimedVerifiedStatus
1GPT-4 (Teaching-Inspired)Accuracy89.2Unverified