Shengkui Zhao, Trung Hieu Nguyen, Bin Ma
Deep complex U-Net structure and convolutional recurrent network (CRN) structure achieve state-of-the-art performance for monaural speech enhancement. Both deep complex U-Net and CRN are encoder and decoder structures with skip connections, which heavily rely on the representation power of the complex-valued convolutional layers. In this paper, we propose a complex convolutional block attention module (CCBAM) to boost the representation power of the complex-valued convolutional layers by constructing more informative features. The CCBAM is a lightweight and general module which can be easily integrated into any complex-valued convolutional layers. We integrate CCBAM with the deep complex U-Net and CRN to enhance their performance for speech enhancement. We further propose a mixed loss function to jointly optimize the complex models in both time-frequency (TF) domain and time domain. By integrating CCBAM and the mixed loss, we form a new end-to-end (E2E) complex speech enhancement framework. Ablation experiments and objective evaluations show the superior performance of the proposed approaches (https://github.com/modelscope/ClearerVoice-Studio).
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Speech Enhancement | WSJ0 + DEMAND + RNNoise | PESQ-NB | 3.44 | DCUNet-MC |
| Speech Enhancement | WSJ0 + DEMAND + RNNoise | PESQ-NB | 3.28 | DCCRN-M |
| Speech Enhancement | WSJ0 + DEMAND + RNNoise | PESQ-NB | 3.25 | DCUNet |
| Speech Enhancement | Deep Noise Suppression (DNS) Challenge | PESQ-WB | 3.23 | FRCRN |
| Speech Enhancement | VoiceBank + DEMAND | PESQ (wb) | 3.43 | D2Former |
| Speech Enhancement | VoiceBank + DEMAND | Para. (M) | 0.86 | D2Former |
| Speech Enhancement | DNS Challenge | PESQ-NB | 3.21 | DCCRN-MC |
| Speech Enhancement | DNS Challenge | PESQ-NB | 3.15 | DCCRN-M |
| Speech Enhancement | DNS Challenge | PESQ-NB | 3.04 | DCCRN |