Skip to main content

Vset: A Multimodal Transformer For Visual Speech Enhancement

Karthik Ramesh, Chao Xing, Wupeng Wang, Dong Wang, Xiao Chen

  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00
    Length: 00:11:45
10 Jun 2021

The transformer architecture has shown great capability in learning long-term dependency and works well in multiple domains. However, transformer has been less considered in audio-visual speech enhancement (AVSE) research, partly due to the convention that treats speech enhancement as a short-time signal processing task. In this paper, we challenge this common belief and show that an audio-visual transformer can significantly improve AVSE performance, by learning the long-term dependency of both intra-modality and inter-modality. We test this new transformer-based AVSE model on the GRID and AVSpeech datasets, and show that it beats several state-of-the-art models by a large margin.

Chairs:
Chandan K A Reddy

Value-Added Bundle(s) Including this Product

More Like This