SOTAVerified

HellaSwag

Papers

Showing 2639 of 39 papers

TitleStatusHype
Self-Reasoning Language Models: Unfold Hidden Reasoning Chains with Few Reasoning Catalyst0
When Chosen Wisely, More Data Is What You Need: A Universal Sample-Efficient Strategy For Data Augmentation0
Slimming Down LLMs Without Losing Their Minds0
SmolTulu: Higher Learning Rate to Batch Size Ratios Can Lead to Better Reasoning in SLMs0
Contrastive Decoding Improves Reasoning in Large Language Models0
Who's Harry Potter? Approximate Unlearning in LLMs0
Towards Multilingual LLM Evaluation for European Languages0
Elastic Weight Consolidation for Full-Parameter Continual Pre-Training of Gemma20
HellaSwag-Pro: A Large-Scale Bilingual Benchmark for Evaluating the Robustness of LLMs in Commonsense Reasoning0
GRIN: GRadient-INformed MoE0
More is Less: The Pitfalls of Multi-Model Synthetic Preference Data in DPO Safety Alignment0
Obliviate: Efficient Unmemorization for Protecting Intellectual Property in Large Language Models0
Domain-Adaptive Continued Pre-Training of Small Language Models0
Pre-training Is (Almost) All You Need: An Application to Commonsense Reasoning0
Show:102550
← PrevPage 2 of 2Next →

No leaderboard results yet.