Skip to main content

Jeffreys divergence-based regularization of neural network output distribution applied to speaker recognition

Pierre-Michel Bousquet (Avignon University); Mickael Rouvier (LIA - Avignon University)

  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00
08 Jun 2023

In this paper, we propose a new loss function for speaker recognition with deep neural network, which is based on Jeffreys Divergence. Adding this divergence to the cross-entropy loss function allows to maximize the target value of the output distribution while smoothing the non-target values. This objective function provides highly discriminative features. Beyond this effect, we propose a theoretical justification of its effectiveness and try to understand how this loss function affects the model, in particular the impact on dataset types (in-domain or out-of-domain w.r.t the training corpus). Our experiments show that Jeffreys loss consistently outperforms the state-of-the-art for speaker recognition, especially on out-of-domain data, and helps limit false alarms.

More Like This

  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00