Skip to main content

DEFENDING AGAINST UNIVERSAL PATCH ATTACKS BY RESTRICTING TOKEN ATTENTION IN VISION TRANSFORMERS

Hongwei Yu (University of Science and Technology Beijing); Jiansheng Chen (University of Science and Technology Beijing); Huimin Ma (University of Science and Technology Beijing); Cheng Yu (Tsinghua University); Xinlong Ding (University of Science and Technology Beijing)

  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00
06 Jun 2023

Previous works reveal that similar to CNNs, vision transformers (ViT) are also vulnerable to universal adversarial patch attacks. In this paper, we empirically reveal and mathematically explain that the shallow tokens in the transformer and the attention of the network can largely influence the classification result. Adversarial patches usually produce large feature norm for the corresponding shallow token vectors which can attract the attention anomalously. Inspired by this, we propose a restriction operation on the attention matrix, which effectively reduces the influence of the patch region. Experiments on ImageNet validate that our proposal can effectively improve ViT's robustness towards white-box universal patch attacks while maintaining satisfactory classification accuracy for clean samples.

More Like This

  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00
  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00
  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00