SOTAVerified

ListOps

Papers

Showing 122 of 22 papers

TitleStatusHype
Mega: Moving Average Equipped Gated AttentionCode2
Simplified State Space Layers for Sequence ModelingCode2
Cached Transformers: Improving Transformers with Differentiable Memory CacheCode1
Sequence Modeling with Multiresolution Convolutional MemoryCode1
Training Discrete Deep Generative Models via Gapped Straight-Through EstimatorCode1
Dynamic Token Normalization Improves Vision TransformersCode1
Efficiently Modeling Long Sequences with Structured State SpacesCode1
The Neural Data Router: Adaptive Control Flow in Transformers Improves Systematic GeneralizationCode1
Going Beyond Linear Transformers with Recurrent Fast Weight ProgrammersCode1
Modeling Hierarchical Structures with Continuous Recursive Neural NetworksCode1
Long Range Arena: A Benchmark for Efficient TransformersCode1
ListOps: A Diagnostic Dataset for Latent Tree LearningCode1
Context Is Not Comprehension0
Small Models, Smarter Learning: The Power of Joint Task Training0
Investigating Recurrent Transformers with Dynamic HaltCode0
Opening the Black Box: Analyzing Attention Weights and Hidden States in Pre-trained Language Models for Non-language TasksCode0
Investigating Pre-trained Language Models on Cross-Domain Datasets, a Step Closer to General AI0
Beam Tree Recursive CellsCode0
DARTFormer: Finding The Best Type Of Attention0
ORCHARD: A Benchmark For Measuring Systematic Generalization of Multi-Hierarchical ReasoningCode0
Adaptive Control Flow in Transformers Improves Systematic Generalization0
Ordered MemoryCode0
Show:102550

No leaderboard results yet.