| Repeat After Me: Transformers are Better than State Space Models at Copying | Feb 1, 2024 | State Space Models | CodeCode Available | 3 |
| Efficient Fine-tuning of Audio Spectrogram Transformers via Soft Mixture of Adapters | Feb 1, 2024 | Mixture-of-Expertsparameter-efficient fine-tuning | CodeCode Available | 1 |
| Graph-Mamba: Towards Long-Range Graph Sequence Modeling with Selective State Spaces | Feb 1, 2024 | Computational EfficiencyGPU | CodeCode Available | 3 |
| LOCOST: State-Space Models for Long Document Abstractive Summarization | Jan 31, 2024 | Abstractive Text SummarizationBook summarization | CodeCode Available | 1 |
| Particle-MALA and Particle-mGRAD: Gradient-based MCMC methods for high-dimensional state-space models | Jan 26, 2024 | Bayesian InferenceState Space Models | CodeCode Available | 1 |
| Vivim: a Video Vision Mamba for Medical Video Segmentation | Jan 25, 2024 | Lesion SegmentationMamba | CodeCode Available | 2 |
| Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model | Jan 17, 2024 | GPUImage Classification | CodeCode Available | 2 |
| MoE-Mamba: Efficient Selective State Space Models with Mixture of Experts | Jan 8, 2024 | MambaMixture-of-Experts | CodeCode Available | 3 |
| Online Variational Sequential Monte Carlo | Dec 19, 2023 | parameter estimationState Space Models | CodeCode Available | 0 |
| Automatic Rao-Blackwellization for Sequential Monte Carlo with Belief Propagation | Dec 15, 2023 | Bayesian InferenceState Space Models | CodeCode Available | 1 |