Headless Horseman: Adversarial Attacks On Transfer Learning Models
Ahmed Abdelkader, Michael Curry, Liam Fowl, Tom Goldstein, Avi Schwarzschild, Manli Shu, Chen Zhu, Christoph Studer
-
SPS
IEEE Members: $11.00
Non-members: $15.00Length: 12:14
Transfer learning facilitates the training of task-specific classifiers using pre-trained models as feature extractors. We present a family of transferable adversarial attacks against such classifiers, generated without access to the classification head. We call these headless attacks. We first demonstrate successful transfer attacks against a victim network using only its feature extractor. This motivates the introduction of a label-blind adversarial attack. This transfer attack method does not require any information about the class-label space of the victim. Our attack lowers the accuracy of a ResNet18 trained on CIFAR10 by over 40%.