Skip to main content

FEDERATED LEARNING FOR ASR BASED ON WAV2VEC 2.0

Tuan Manh Nguyen (LIA, Avignon University); Salima Mdhaffar (LIA - University of Avignon); Natalia Tomashenko (LIA, University of Avignon); Jean-Francois Bonastre (Université d’Avignon); Yannick Estève (LIA - Avignon University)

  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00
08 Jun 2023

This paper presents a study on the use of federated learning to train an ASR model based on a wav2vec~2.0 model pretrained by self supervision. Carried out on the well-known TED-LIUM~3 dataset, our experiments show that such a model can obtain, with no use of a language model, a word error rate of 10.92% on the official TED-LIUM~3 test set, without sharing any data from the different users. We also analyze the ASR performance for speakers depending to their participation to the federated learning. Since federated learning was first introduced for privacy purposes, we also measure its ability to protect speaker identity. To do that, we exploit an approach to analyze information contains in exchanged models based on a neural network footprint on an indicator dataset. This analysis is made layer-wise and shows which layers in an exchanged wav2vec 2.0-based model bring the speaker identity information.

More Like This

  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00
  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00
  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00