SOTAVerified

SepMamba: State-space models for speaker separation using Mamba

2024-10-28Code Available1· sign in to hype

Thor Højhus Avenstrup, Boldizsár Elek, István László Mádi, András Bence Schin, Morten Mørup, Bjørn Sand Jensen, Kenny Falkær Olsen

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Deep learning-based single-channel speaker separation has improved significantly in recent years largely due to the introduction of the transformer-based attention mechanism. However, these improvements come at the expense of intense computational demands, precluding their use in many practical applications. As a computationally efficient alternative with similar modeling capabilities, Mamba was recently introduced. We propose SepMamba, a U-Net-based architecture composed primarily of bidirectional Mamba layers. We find that our approach outperforms similarly-sized prominent models - including transformer-based models - on the WSJ0 2-speaker dataset while enjoying a significant reduction in computational cost, memory usage, and forward pass time. We additionally report strong results for causal variants of SepMamba. Our approach provides a computationally favorable alternative to transformer-based architectures for deep speech separation.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
WSJ0-2mixSepMamba + DM (M)SI-SDRi22.7Unverified
WSJ0-2mixSepMamba + DM (S)SI-SDRi21.2Unverified

Reproductions