SOTAVerified

Sentence Completion

Papers

Showing 1120 of 91 papers

TitleStatusHype
Language Models are Few-Shot LearnersCode3
PaLM: Scaling Language Modeling with PathwaysCode2
Parameter-Efficient Sparsity Crafting from Dense to Mixture-of-Experts for Instruction Tuning on General TasksCode2
DeBERTa: Decoding-enhanced BERT with Disentangled AttentionCode2
Crosslingual Generalization through Multitask FinetuningCode2
Scaling Language Models: Methods, Analysis & Insights from Training GopherCode2
LaMini-LM: A Diverse Herd of Distilled Models from Large-Scale InstructionsCode2
Sheared LLaMA: Accelerating Language Model Pre-training via Structured PruningCode2
The CoT Collection: Improving Zero-shot and Few-shot Learning of Language Models via Chain-of-Thought Fine-TuningCode2
Exploring the Benefits of Training Expert Language Models over Instruction TuningCode1
Show:102550
← PrevPage 2 of 10Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1CompassMTL 567M with TailorAccuracy96.1Unverified
2CompassMTL 567MAccuracy95.6Unverified
3DeBERTa-Large 304M (classification-based)Accuracy95.6Unverified
4GPT-4 (10-shot)Accuracy95.3Unverified
5LLaMA3+MoSLoRAAccuracy95Unverified
6LLaMA-2 13B + MixLoRAAccuracy94.7Unverified
7DeBERTa-Large 304MAccuracy94.7Unverified
8Unicorn 11B (fine-tuned)Accuracy93.9Unverified
9LLaMA-3 8B + MixLoRAAccuracy93.3Unverified
10LLaMA-2 7B + MixLoRAAccuracy93.1Unverified