SOTAVerified

ListOps

Papers

Showing 122 of 22 papers

TitleStatusHype
Simplified State Space Layers for Sequence ModelingCode2
Mega: Moving Average Equipped Gated AttentionCode2
Training Discrete Deep Generative Models via Gapped Straight-Through EstimatorCode1
Cached Transformers: Improving Transformers with Differentiable Memory CacheCode1
Dynamic Token Normalization Improves Vision TransformersCode1
Efficiently Modeling Long Sequences with Structured State SpacesCode1
Going Beyond Linear Transformers with Recurrent Fast Weight ProgrammersCode1
ListOps: A Diagnostic Dataset for Latent Tree LearningCode1
Long Range Arena: A Benchmark for Efficient TransformersCode1
Modeling Hierarchical Structures with Continuous Recursive Neural NetworksCode1
Sequence Modeling with Multiresolution Convolutional MemoryCode1
The Neural Data Router: Adaptive Control Flow in Transformers Improves Systematic GeneralizationCode1
Small Models, Smarter Learning: The Power of Joint Task Training0
DARTFormer: Finding The Best Type Of Attention0
Investigating Pre-trained Language Models on Cross-Domain Datasets, a Step Closer to General AI0
Adaptive Control Flow in Transformers Improves Systematic Generalization0
Context Is Not Comprehension0
Investigating Recurrent Transformers with Dynamic HaltCode0
Opening the Black Box: Analyzing Attention Weights and Hidden States in Pre-trained Language Models for Non-language TasksCode0
Beam Tree Recursive CellsCode0
ORCHARD: A Benchmark For Measuring Systematic Generalization of Multi-Hierarchical ReasoningCode0
Ordered MemoryCode0
Show:102550

No leaderboard results yet.