SOTAVerified

Channel-Attention Dense U-Net for Multichannel Speech Enhancement

2020-01-30Code Available1· sign in to hype

Bahareh Tolooshams, Ritwik Giri, Andrew H. Song, Umut Isik, Arvindh Krishnaswamy

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Supervised deep learning has gained significant attention for speech enhancement recently. The state-of-the-art deep learning methods perform the task by learning a ratio/binary mask that is applied to the mixture in the time-frequency domain to produce the clean speech. Despite the great performance in the single-channel setting, these frameworks lag in performance in the multichannel setting as the majority of these methods a) fail to exploit the available spatial information fully, and b) still treat the deep architecture as a black box which may not be well-suited for multichannel audio processing. This paper addresses these drawbacks, a) by utilizing complex ratio masking instead of masking on the magnitude of the spectrogram, and more importantly, b) by introducing a channel-attention mechanism inside the deep architecture to mimic beamforming. We propose Channel-Attention Dense U-Net, in which we apply the channel-attention unit recursively on feature maps at every layer of the network, enabling the network to perform non-linear beamforming. We demonstrate the superior performance of the network against the state-of-the-art approaches on the CHiME-3 dataset.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
CHiME-3CA Dense U-Net (Complex)SDR18.64Unverified
CHiME-3Dense U-Net (Complex)SDR18.4Unverified
CHiME-3Dense U-Net (Real)SDR16.86Unverified
CHiME-3U-Net (Real)SDR15.97Unverified
CHiME-3Noisy/unprocessedSDR6.5Unverified

Reproductions