SOTAVerified

StrategyQA

StrategyQA aims to measure the ability of models to answer questions that require multi-step implicit reasoning.

Source: BIG-bench

Papers

Showing 1120 of 40 papers

TitleStatusHype
Mutual Reasoning Makes Smaller LLMs Stronger Problem-SolversCode4
Meta-prompting Optimized Retrieval-augmented Generation0
Question-Analysis Prompting Improves LLM Performance in Reasoning Tasks0
Advancing Process Verification for Large Language Models via Tree-Based Preference Learning0
Unchosen Experts Can Contribute Too: Unleashing MoE Models' Power by Self-ContrastCode1
Improving Attributed Text Generation of Large Language Models via Preference Learning0
CR-LT-KGQA: A Knowledge Graph Question Answering Dataset Requiring Commonsense Reasoning and Long-Tail KnowledgeCode1
Distillation Contrastive Decoding: Improving LLMs Reasoning with Contrastive Decoding and DistillationCode1
Towards Uncertainty-Aware Language Agent0
Escape Sky-high Cost: Early-stopping Self-Consistency for Multi-step ReasoningCode1
Show:102550
← PrevPage 2 of 4Next →

No leaderboard results yet.