SOTAVerified

Long-range modeling

A new task for testing the long-sequence modeling capabilities and efficiency of language models.

Image credit: SCROLLS: Standardized CompaRison Over Long Language Sequences

Papers

Showing 7695 of 95 papers

TitleStatusHype
Dimension Mixer: Group Mixing of Input Dimensions for Efficient Function ApproximationCode0
Advancing Regular Language Reasoning in Linear Recurrent Neural NetworksCode0
Improving FHB Screening in Wheat Breeding Using an Efficient Transformer Model0
AICT: An Adaptive Image Compression Transformer0
Focus Your Attention (with Adaptive IIR Filters)0
A General-Purpose Multilingual Document EncoderCode0
RFR-WWANet: Weighted Window Attention-Based Recovery Feature Resolution Network for Unsupervised Image RegistrationCode0
HST-MRF: Heterogeneous Swin Transformer with Multi-Receptive Field for Medical Image Segmentation0
CoLT5: Faster Long-Range Transformers with Conditional Computation0
Token Transformer: Can class token help window-based transformer build better long-range interactions?0
Pose Guided Human Image Synthesis with Partially Decoupled GAN0
CNSNet: A Cleanness-Navigated-Shadow Network for Shadow RemovalCode0
How to Train Your HiPPO: State Space Models with Generalized Orthogonal Basis ProjectionsCode0
On the Parameterization and Initialization of Diagonal State Space ModelsCode0
0/1 Deep Neural Networks via Block Coordinate Descent0
Diagonal State Spaces are as Effective as Structured State SpacesCode0
Dyadformer: A Multi-modal Transformer for Long-Range Modeling of Dyadic Interactions0
Long-Range Modeling of Source Code Files with eWASH: Extended Window Access by Syntax Hierarchy0
Sparse Factorization of Large Square MatricesCode0
Gated Relational Graph Attention Networks0
Show:102550
← PrevPage 4 of 4Next →

No leaderboard results yet.