LEVERAGING SPARSITY WITH SPIKING RECURRENT NEURAL NETWORKS FOR ENERGY-EFFICIENT KEYWORD SPOTTING
Manon Dampfhoffer (SPINTEC University Grenoble Alpes); Thomas Mesquida (CEA LIST); Emmanuel Hardy (CEA-Leti); Alexandre Valentian (CEA-List); Lorena Anghel (SPINTEC University Grenoble Alpes)
-
SPS
IEEE Members: $11.00
Non-members: $15.00
Bio-inspired Spiking Neural Networks (SNNs) are promising candidates to replace standard Artificial Neural Networks (ANNs) for energy-efficient keyword spotting (KWS) systems. In this work, we compare the trade-off between accuracy and energy-efficiency of a gated recurrent SNN (SpikGRU) with a standard Gated Recurrent Unit (GRU) on the Google Speech Command Dataset (GSCD) v2. We show that, by taking advantage of the sparse spiking activity of the SNN, both accuracy and energy-efficiency can be increased. Leveraging data sparsity by using spiking inputs, such as those produced by spiking audio feature extractors or dynamic sensors, can further improve energy-efficiency. We demonstrate state-of-the-art results for SNNs on GSCD v2 with up to 95.9% accuracy. Moreover, SpikGRU can achieve similar accuracy than GRU while reducing the number of operations by up to 82%.