MossFormer: Pushing the Performance Limit of Monaural Speech Separation using Gated Single-head Transformer with Convolution-augmented Joint Self-Attentions
Shengkui Zhao (Alibaba Group); Bin Ma ("Alibaba, Singapore R&D Center")
-
SPS
IEEE Members: $11.00
Non-members: $15.00
Transformer based models have provided significant performance improvements in monaural speech separation. However, there is still a performance gap compared to a recent proposed upper bound due to the inefficient modelling of long-range elemental interactions and local feature patterns. In this work, we propose a gated single-head transformer architecture with convolution-augmented joint self-attentions, named MossFormer. MossFormer employs a joint local and global self-attention architecture for the local chunks and the full sequence. The joint attention enables MossFormer model full-sequence elemental interaction directly. In addition, we employ a powerful attentive gating mechanism with simplified single-head self-attentions and convolutions for the position-wise local pattern modelling. As a consequence, MossFormer significantly outperforms the previous models and achieves the state-of-the-art results on WSJ0-2/3mix and WHAM!/WHAMR! benchmarks.