SOTAVerified

ListOps

Papers

Showing 110 of 22 papers

TitleStatusHype
Context Is Not Comprehension0
Small Models, Smarter Learning: The Power of Joint Task Training0
Investigating Recurrent Transformers with Dynamic HaltCode0
Cached Transformers: Improving Transformers with Differentiable Memory CacheCode1
Opening the Black Box: Analyzing Attention Weights and Hidden States in Pre-trained Language Models for Non-language TasksCode0
Investigating Pre-trained Language Models on Cross-Domain Datasets, a Step Closer to General AI0
Beam Tree Recursive CellsCode0
Sequence Modeling with Multiresolution Convolutional MemoryCode1
DARTFormer: Finding The Best Type Of Attention0
Mega: Moving Average Equipped Gated AttentionCode2
Show:102550
← PrevPage 1 of 3Next →

No leaderboard results yet.