| Transformers are SSMs: Generalized Models and Efficient Algorithms Through Structured State Space Duality | May 31, 2024 | Language ModelingLanguage Modelling | CodeCode Available | 11 |
| MambaOut: Do We Really Need Mamba for Vision? | May 13, 2024 | image-classificationImage Classification | CodeCode Available | 7 |
| xLSTM 7B: A Recurrent LLM for Fast and Efficient Inference | Mar 17, 2025 | MambaMath | CodeCode Available | 7 |
| VMamba: Visual State Space Model | Jan 18, 2024 | Computational EfficiencyLanguage Modeling | CodeCode Available | 7 |
| MambaVision: A Hybrid Mamba-Transformer Vision Backbone | Jul 10, 2024 | Image ClassificationInstance Segmentation | CodeCode Available | 7 |
| Griffin: Mixing Gated Linear Recurrences with Local Attention for Efficient Language Models | Feb 29, 2024 | Language ModellingMamba | CodeCode Available | 7 |
| Mamba: Linear-Time Sequence Modeling with Selective State Spaces | Dec 1, 2023 | 2D Pose EstimationCommon Sense Reasoning | CodeCode Available | 6 |
| MambaIRv2: Attentive State Space Restoration | Nov 22, 2024 | Computational EfficiencyImage Restoration | CodeCode Available | 5 |
| MambaIR: A Simple Baseline for Image Restoration with State-Space Model | Feb 23, 2024 | Image RestorationImage Super-Resolution | CodeCode Available | 5 |
| Jamba-1.5: Hybrid Transformer-Mamba Models at Scale | Aug 22, 2024 | ChatbotInstruction Following | CodeCode Available | 5 |