Training Robust Spiking Neural Networks with ViewPoint Transform and SpatioTemporal Stretching
Haibo Shen (Huazhong University of Science and Technology); Juyu Xiao (Huazhong University of Science and Technology); Yihao Luo (Yichang Testing Technique R&D Institute); Xiang Cao (School of Computer Science and Technology, Huazhong University of Science and Technology); Liangqi Zhang (Huazhong University of Science and Technology); Tianjiang Wang (School of Computer Science and Technology, Huazhong University of Science and Technology)
-
SPS
IEEE Members: $11.00
Non-members: $15.00
Neuromorphic vision sensors (event cameras) simulate biological visual perception systems and have the advantages of high temporal resolution, less data redundancy, low power consumption, and large dynamic range. Since both events and spikes are modeled from neural signals, event cameras are inherently suitable for spiking neural networks (SNNs), which are considered promising models for artificial intelligence (AI) and theoretical neuroscience. However, the unconventional visual signals of these cameras pose a great challenge to the robustness of spiking neural networks. In this paper, we propose a novel data augmentation method, ViewPoint Transform and SpatioTemporal Stretching (VPT-STS). It improves the robustness of SNNs by transforming the rotation centers and angles in the spatiotemporal domain to generate samples from different viewpoints. Furthermore, we introduce the spatiotemporal stretching to avoid potential information loss in viewpoint transformation. Extensive experiments on prevailing neuromorphic datasets demonstrate that VPT-STS is broadly effective on multi-event representations and significantly outperforms pure spatial geometric transformations. Notably, the SNNs model with VPT-STS achieves a state-of-the-art accuracy of 84.4% on the DVS-CIFAR10 dataset.