SOTAVerified

Monaural Speech Enhancement with Complex Convolutional Block Attention Module and Joint Time Frequency Losses

2021-02-03Code Available2· sign in to hype

Shengkui Zhao, Trung Hieu Nguyen, Bin Ma

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Deep complex U-Net structure and convolutional recurrent network (CRN) structure achieve state-of-the-art performance for monaural speech enhancement. Both deep complex U-Net and CRN are encoder and decoder structures with skip connections, which heavily rely on the representation power of the complex-valued convolutional layers. In this paper, we propose a complex convolutional block attention module (CCBAM) to boost the representation power of the complex-valued convolutional layers by constructing more informative features. The CCBAM is a lightweight and general module which can be easily integrated into any complex-valued convolutional layers. We integrate CCBAM with the deep complex U-Net and CRN to enhance their performance for speech enhancement. We further propose a mixed loss function to jointly optimize the complex models in both time-frequency (TF) domain and time domain. By integrating CCBAM and the mixed loss, we form a new end-to-end (E2E) complex speech enhancement framework. Ablation experiments and objective evaluations show the superior performance of the proposed approaches (https://github.com/modelscope/ClearerVoice-Studio).

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
Deep Noise Suppression (DNS) ChallengeFRCRNPESQ-WB3.23Unverified
DNS ChallengeDCCRN-MPESQ-NB3.15Unverified
DNS ChallengeDCCRN-MCPESQ-NB3.21Unverified
DNS ChallengeDCCRNPESQ-NB3.04Unverified
VoiceBank + DEMANDD2FormerPESQ (wb)3.43Unverified
WSJ0 + DEMAND + RNNoiseDCUNet-MCPESQ-NB3.44Unverified
WSJ0 + DEMAND + RNNoiseDCCRN-MPESQ-NB3.28Unverified
WSJ0 + DEMAND + RNNoiseDCUNetPESQ-NB3.25Unverified

Reproductions