IMPROVING TRANSFORMER-BASED END-TO-END SPEAKER DIARIZATION BY ASSIGNING AUXILIARY LOSSES TO ATTENTION HEADS
Ye-Rin Jeoung (Hanyang University); Joon-Young Yang (Hanyang University); Jeong-Hwan Choi (Hanyang University); Joon-Hyuk Chang (Hanyang University)
-
SPS
IEEE Members: $11.00
Non-members: $15.00
Transformer-based end-to-end neural speaker diarization (EEND) models utilize the multi-head self-attention (SA) mechanism to enable accurate speaker label prediction in overlapped speech regions. In this study, to enhance the training effectiveness of SA-EEND models, we propose the use of auxiliary losses for the SA heads of the transformer layers. Specifically, we assume that the attention weight matrices of an SA layer are redundant if their patterns are similar to those of the identity matrix. We then explicitly constrain such matrices to exhibit specific speaker activity patterns relevant to voice activity detection or overlapped speech detection tasks. Consequently, we expect the proposed auxiliary losses to guide the transformer layers to exhibit more diverse patterns in the attention weights, thereby reducing the assumed redundancies in the SA heads. The effectiveness of the proposed method is demonstrated using the simulated and CALLHOME datasets for two-speaker diarization tasks, reducing the diarization error rate of the conventional SA-EEND model by 32.58% and 17.11%, respectively.