SOTAVerified

StrategyQA

StrategyQA aims to measure the ability of models to answer questions that require multi-step implicit reasoning.

Source: BIG-bench

Papers

Showing 3140 of 40 papers

TitleStatusHype
Question-Analysis Prompting Improves LLM Performance in Reasoning Tasks0
The ART of LLM Refinement: Ask, Refine, and Trust0
Towards Uncertainty-Aware Language Agent0
Unraveling Indirect In-Context Learning Using Influence Functions0
DeLTa: A Decoding Strategy based on Logit Trajectory Prediction Improves Factuality and Reasoning AbilityCode0
Rationale-Aware Answer Verification by Pairwise Self-EvaluationCode0
Distilling Reasoning Capabilities into Smaller Language ModelsCode0
Tailoring Self-Rationalizers with Multi-Reward DistillationCode0
Teaching Smaller Language Models To Generalise To Unseen Compositional QuestionsCode0
Voting or Consensus? Decision-Making in Multi-Agent DebateCode0
Show:102550
← PrevPage 4 of 4Next →

No leaderboard results yet.