Channel-Attention Dense U-Net For Multichannel Speech Enhancement
Bahareh Tolooshams, Ritwik Giri, Andrew Song, Umut Isik, Arvindh Krishnaswamy
-
SPS
IEEE Members: $11.00
Non-members: $15.00Length: 15:49
Supervised deep learning has gained significant attention for speech enhancement recently. The state-of-the-art deep learning methods perform the task by learning a ratio/binary mask that is applied to the mixture in the time-frequency domain to produce clean speech. Despite the great performance in the single-channel setting, these frameworks lag in performance in the multichannel setting as the majority of them a) fail to fully exploit the available spatial information, and b) still treat the deep architecture as a black box which may not be well-suited for multichannel audio processing. This paper addresses these drawbacks, a) by utilizing complex ratio masking instead of masking on the magnitude of the spectrogram, and more importantly, b) by introducing a channel-attention mechanism inside the deep architecture to mimic beamforming. We propose Channel-Attention Dense U-Net, in which we apply the channel-attention unit recursively on feature maps at every layer of the network, enabling the network to perform non-linear beamforming. We demonstrate the superior performance of the network against the state-of-the-art approaches on the CHiME-3 dataset.