A Joint Training Framework Of Multi-Look Separator And Speaker Embedding Extractor For Overlapped Speech
Naijun Zheng, Na Li, Bo Wu, Meng Yu, JianWei Yu, Chao Weng, Dan Su, XunYing Liu, Helen Meng
-
SPS
IEEE Members: $11.00
Non-members: $15.00Length: 00:12:07
In multi-talker cases, overlapped speech degrades the speaker verification (SV) performance dramatically. To tackle this challenging problem, speech separation with multi-channel techniques can be adopted to extract each speaker's signals to improve the SV performance. In this paper, a joint training framework of the front-end multi-look speech separator and the back-end speaker embedding extractor is proposed for multi-channel overlapped speech. To better leverage the complementarity between the speech separator and the speaker embedding extractor, several training strategies are proposed to jointly optimize the two modules. Experimental results show that the proposed joint training framework significantly outperforms the individual SV system by around 52% relative EER reduction. Additionally, the robustness of the proposed framework is further evaluated under different conditions.
Chairs:
Takafumi Koshinaka