DEFENDING AGAINST UNIVERSAL PATCH ATTACKS BY RESTRICTING TOKEN ATTENTION IN VISION TRANSFORMERS
Hongwei Yu (University of Science and Technology Beijing); Jiansheng Chen (University of Science and Technology Beijing); Huimin Ma (University of Science and Technology Beijing); Cheng Yu (Tsinghua University); Xinlong Ding (University of Science and Technology Beijing)
-
SPS
IEEE Members: $11.00
Non-members: $15.00
Previous works reveal that similar to CNNs, vision transformers (ViT) are also vulnerable to universal adversarial patch attacks. In this paper, we empirically reveal and mathematically explain that the shallow tokens in the transformer and the attention of the network can largely influence the classification result. Adversarial patches usually produce large feature norm for the corresponding shallow token vectors which can attract the attention anomalously. Inspired by this, we propose a restriction operation on the attention matrix, which effectively reduces the influence of the patch region. Experiments on ImageNet validate that our proposal can effectively improve ViT's robustness towards white-box universal patch attacks while maintaining satisfactory classification accuracy for clean samples.