Memory Layers With Multi-Head Attention Mechanisms For Text-Dependent Speaker Verification
Victoria Mingote, Antonio Miguel, Alfonso Ortega, Eduardo Lleida
-
SPS
IEEE Members: $11.00
Non-members: $15.00Length: 00:10:06
In this paper, we explore an approach based on memory layers and multi-head attention mechanisms to improve in an efficient way the performance of text-dependent speaker verification (SV) systems. The most extended SV systems based on Deep Neural Networks (DNN) extract the embedding of the utterance from the average pooling of the temporal dimension after processing. Unlike previous works, we can exploit the phonetic knowledge needed for text-dependent SV systems by combining the temporal attention of multiple parallel heads with the phonetic embeddings extracted from a phonetic classification network, which helps to guide to the attention mechanism with the role of the positional embedding. The addition of a memory layer to a text-dependent SV system was tested on the RSR2015-part II and DeepMine-part I databases, where, in both cases outperformed the baseline result and the reference system based on the same transformer network without the memory layer.
Chairs:
Paola Garcia