SOTAVerified

Multi-task Language Understanding

The test covers 57 tasks including elementary mathematics, US history, computer science, law, and more. https://arxiv.org/pdf/2009.03300.pdf

Papers

Showing 2130 of 57 papers

TitleStatusHype
Measuring Massive Multitask Language UnderstandingCode2
Solving Quantitative Reasoning Problems with Language ModelsCode2
Routoo: Learning to Route to Large Language Models EffectivelyCode2
ALBERT: A Lite BERT for Self-supervised Learning of Language RepresentationsCode2
Atlas: Few-shot Learning with Retrieval Augmented Language ModelsCode2
UL2: Unifying Language Learning ParadigmsCode1
ArabicMMLU: Assessing Massive Multitask Language Understanding in ArabicCode1
Are Human-generated Demonstrations Necessary for In-context Learning?Code1
Gemini: A Family of Highly Capable Multimodal ModelsCode1
GPT-NeoX-20B: An Open-Source Autoregressive Language ModelCode1
Show:102550
← PrevPage 3 of 6Next →

No leaderboard results yet.