| Training Compute-Optimal Large Language Models | Mar 29, 2022 | AnachronismsAnalogical Similarity | CodeCode Available | 6 |
| When Chosen Wisely, More Data Is What You Need: A Universal Sample-Efficient Strategy For Data Augmentation | Mar 17, 2022 | Data AugmentationHellaSwag | CodeCode Available | 1 |
| Scaling Language Models: Methods, Analysis & Insights from Training Gopher | Dec 8, 2021 | Abstract AlgebraAnachronisms | CodeCode Available | 2 |
| When Chosen Wisely, More Data Is What You Need: A Universal Sample-Efficient Strategy For Data Augmentation | Nov 16, 2021 | Data AugmentationHellaSwag | —Unverified | 0 |
| Comparing Test Sets with Item Response Theory | Jun 1, 2021 | HellaSwagNatural Language Understanding | —Unverified | 0 |
| UNICORN on RAINBOW: A Universal Commonsense Reasoning Model on a New Multitask Benchmark | Mar 24, 2021 | Common Sense ReasoningHellaSwag | CodeCode Available | 1 |
| English Intermediate-Task Training Improves Zero-Shot Cross-Lingual Transfer Too | May 26, 2020 | Cross-Lingual TransferHellaSwag | —Unverified | 0 |
| Pre-training Is (Almost) All You Need: An Application to Commonsense Reasoning | Apr 29, 2020 | AllHellaSwag | —Unverified | 0 |
| HellaSwag: Can a Machine Really Finish Your Sentence? | May 19, 2019 | HellaSwagNatural Language Inference | CodeCode Available | 0 |