Jeffreys divergence-based regularization of neural network output distribution applied to speaker recognition
Pierre-Michel Bousquet (Avignon University); Mickael Rouvier (LIA - Avignon University)
-
SPS
IEEE Members: $11.00
Non-members: $15.00
In this paper, we propose a new loss function for speaker recognition with deep neural network, which is based on Jeffreys Divergence. Adding this divergence to the cross-entropy loss function allows to maximize the target value of the output distribution while smoothing the non-target values. This objective function provides highly discriminative features. Beyond this effect, we propose a theoretical justification of its effectiveness and try to understand how this loss function affects the model, in particular the impact on dataset types (in-domain or out-of-domain w.r.t the training corpus). Our experiments show that Jeffreys loss consistently outperforms the state-of-the-art for speaker recognition, especially on out-of-domain data, and helps limit false alarms.