SOTAVerified

StrategyQA

StrategyQA aims to measure the ability of models to answer questions that require multi-step implicit reasoning.

Source: BIG-bench

Papers

Showing 2130 of 40 papers

TitleStatusHype
Better Retrieval May Not Lead to Better Question Answering0
Deduction under Perturbed Evidence: Probing Student Simulation Capabilities of Large Language Models0
Dialectical Behavior Therapy Approach to LLM Prompting0
Fusing Bidirectional Chains of Thought and Reward Mechanisms A Method for Enhancing Question-Answering Capabilities of Large Language Models for Chinese Intangible Cultural Heritage0
IAG: Induction-Augmented Generation Framework for Answering Reasoning Questions0
Improving Attributed Text Generation of Large Language Models via Preference Learning0
Large Language Models Are Also Good Prototypical Commonsense Reasoners0
Learning to Decompose: Hypothetical Question Decomposition Based on Comparable Texts0
Meta-prompting Optimized Retrieval-augmented Generation0
Proof of Thought : Neurosymbolic Program Synthesis allows Robust and Interpretable Reasoning0
Show:102550
← PrevPage 3 of 4Next →

No leaderboard results yet.