Timezone: »

Perturbing Across the Feature Hierarchy to Improve Standard and Strict Blackbox Attack Transferability
Nathan Inkawhich · Kevin J Liang · Binghui Wang · Matthew Inkawhich · Lawrence Carin · Yiran Chen

Wed Dec 09 09:00 AM -- 11:00 AM (PST) @ Poster Session 3 #918
We consider the blackbox transfer-based targeted adversarial attack threat model in the realm of deep neural network (DNN) image classifiers. Rather than focusing on crossing decision boundaries at the output layer of the source model, our method perturbs representations throughout the extracted feature hierarchy to resemble other classes. We design a flexible attack framework that allows for multi-layer perturbations and demonstrates state-of-the-art targeted transfer performance between ImageNet DNNs. We also show the superiority of our feature space methods under a relaxation of the common assumption that the source and target models are trained on the same dataset and label space, in some instances achieving a $10\times$ increase in targeted success rate relative to other blackbox transfer methods. Finally, we analyze why the proposed methods outperform existing attack strategies and show an extension of the method in the case when limited queries to the blackbox model are allowed.

Author Information

Nathan Inkawhich (Duke University)
Kevin J Liang (Facebook)
Binghui Wang (Duke University)
Matthew Inkawhich (Duke University)
Lawrence Carin (Duke University)
Yiran Chen (Duke University)

More from the Same Authors