SOTAVerified

Long-range modeling

A new task for testing the long-sequence modeling capabilities and efficiency of language models.

Image credit: SCROLLS: Standardized CompaRison Over Long Language Sequences

Papers

Showing 7695 of 95 papers

TitleStatusHype
How to Train Your HiPPO: State Space Models with Generalized Orthogonal Basis ProjectionsCode0
On the Parameterization and Initialization of Diagonal State Space ModelsCode0
0/1 Deep Neural Networks via Block Coordinate Descent0
ChordMixer: A Scalable Neural Attention Model for Sequences with Different LengthsCode1
UL2: Unifying Language Learning ParadigmsCode1
Paramixer: Parameterizing Mixing Links in Sparse Factors Works Better than Dot-Product Self-AttentionCode1
Diagonal State Spaces are as Effective as Structured State SpacesCode0
SCROLLS: Standardized CompaRison Over Long Language SequencesCode1
Classification of Long Sequential Data using Circular Dilated Convolutional Neural NetworksCode1
LongT5: Efficient Text-To-Text Transformer for Long SequencesCode1
Efficiently Modeling Long Sequences with Structured State SpacesCode1
Dyadformer: A Multi-modal Transformer for Long-Range Modeling of Dyadic Interactions0
Long-Range Modeling of Source Code Files with eWASH: Extended Window Access by Syntax Hierarchy0
Sparse Factorization of Large Square MatricesCode0
Image Super-Resolution With Non-Local Sparse AttentionCode1
DSANet: Dynamic Segment Aggregation Network for Video-Level Representation LearningCode1
Gated Relational Graph Attention Networks0
Long Range Arena: A Benchmark for Efficient TransformersCode1
Disentangling and Unifying Graph Convolutions for Skeleton-Based Action RecognitionCode1
V4D:4D Convolutional Neural Networks for Video-level Representation LearningCode1
Show:102550
← PrevPage 4 of 4Next →

No leaderboard results yet.