MULTI-HEAD UNCERTAINTY INFERENCE FOR ADVERSARIAL ATTACK DETECTION
Yuqi Yang (Beijing University of Posts and Telecommunications); Songyun Yang (Beijing University of Posts and Telecommunications); Jiyang Xie (Beijing University of Posts and Telecommunications); Zhongwei Si (Beijing University of Posts and Telecommunications); Kai Guo (BUPT); Ke Zhang (North China Electric Power University); Kongming Liang (Beijing University of Posts and Telecommunications)
-
SPS
IEEE Members: $11.00
Non-members: $15.00
Deep neural networks (DNNs) are sensitive and susceptible to tiny perturbations by adversarial attacks which cause erroneous predictions. Various methods, including adversarial defense and uncertainty inference (UI), have been developed to overcome adversarial attacks in recent years. In this paper, we propose a multi-head uncertainty inference (MH-UI) framework for detecting adversarial attack examples. We adopt a multi-head architecture with multiple prediction heads to obtain predictions from different depths in the DNNs and introduce shallow information for the UI. Using independent heads at different depths, the normalized predictions are assumed to follow the same Dirichlet distribution, and we estimate the distribution parameter of it by moment matching. Cognitive uncertainty brought by the adversarial attacks will be reflected and amplified in the distribution. Experimental results show that the proposed MH-UI framework has good performance in different settings of adversarial attack detection tasks.