SOTAVerified

Long-range modeling

A new task for testing the long-sequence modeling capabilities and efficiency of language models.

Image credit: SCROLLS: Standardized CompaRison Over Long Language Sequences

Papers

Showing 7180 of 95 papers

TitleStatusHype
Simplified State Space Layers for Sequence ModelingCode2
Investigating Efficiently Extending Transformers for Long Input SummarizationCode3
U-Net vs Transformer: Is U-Net Outdated in Medical Image Registration?Code1
Efficient Long-Text Understanding with Short-Text ModelsCode1
Weakly Supervised Object Localization via Transformer with Implicit Spatial CalibrationCode1
How to Train Your HiPPO: State Space Models with Generalized Orthogonal Basis Projections0
On the Parameterization and Initialization of Diagonal State Space Models0
0/1 Deep Neural Networks via Block Coordinate Descent0
ChordMixer: A Scalable Neural Attention Model for Sequences with Different LengthsCode1
UL2: Unifying Language Learning ParadigmsCode1
Show:102550
← PrevPage 8 of 10Next →

No leaderboard results yet.