SOTAVerified

StrategyQA

StrategyQA aims to measure the ability of models to answer questions that require multi-step implicit reasoning.

Source: BIG-bench

Papers

Showing 2640 of 40 papers

TitleStatusHype
Advancing Process Verification for Large Language Models via Tree-Based Preference Learning0
Improving Attributed Text Generation of Large Language Models via Preference Learning0
Towards Uncertainty-Aware Language Agent0
IAG: Induction-Augmented Generation Framework for Answering Reasoning Questions0
The ART of LLM Refinement: Ask, Refine, and Trust0
Tailoring Self-Rationalizers with Multi-Reward DistillationCode0
Large Language Models Are Also Good Prototypical Commonsense Reasoners0
Answering Unseen Questions With Smaller Language Models Using Rationale Generation and Dense Retrieval0
Teaching Smaller Language Models To Generalise To Unseen Compositional QuestionsCode0
Deduction under Perturbed Evidence: Probing Student Simulation Capabilities of Large Language Models0
Hint of Thought prompting: an explainable and zero-shot approach to reasoning tasks with LLMs0
Self-Evaluation Guided Beam Search for Reasoning0
Distilling Reasoning Capabilities into Smaller Language ModelsCode0
Learning to Decompose: Hypothetical Question Decomposition Based on Comparable Texts0
Better Retrieval May Not Lead to Better Question Answering0
Show:102550
← PrevPage 2 of 2Next →

No leaderboard results yet.