SOTAVerified

Multi-task Language Understanding

The test covers 57 tasks including elementary mathematics, US history, computer science, law, and more. https://arxiv.org/pdf/2009.03300.pdf

Papers

Showing 5157 of 57 papers

TitleStatusHype
Evaluating Large Language Models Trained on CodeCode3
Measuring Massive Multitask Language UnderstandingCode2
Language Models are Few-Shot LearnersCode3
UnifiedQA: Crossing Format Boundaries With a Single QA SystemCode1
ALBERT: A Lite BERT for Self-supervised Learning of Language RepresentationsCode2
RoBERTa: A Robustly Optimized BERT Pretraining ApproachCode1
Language Models are Unsupervised Multitask LearnersCode1
Show:102550
← PrevPage 6 of 6Next →

No leaderboard results yet.