Poster

Efficient and Effective Augmentation Strategy for Adversarial Training

Sravanti Addepalli · Samyak Jain · Venkatesh Babu R

Hall J #422

Keywords: [ adversarial training ] [ Adversarial Robustness ] [ Data Augmentation ]

[ Abstract ]
[ Poster [ OpenReview
Thu 1 Dec 2 p.m. PST — 4 p.m. PST
 
Spotlight presentation: Lightning Talks 6A-3
Thu 8 Dec 6 p.m. PST — 6:15 p.m. PST

Abstract:

Adversarial training of Deep Neural Networks is known to be significantly more data-hungry when compared to standard training. Furthermore, complex data augmentations such as AutoAugment, which have led to substantial gains in standard training of image classifiers, have not been successful with Adversarial Training. We first explain this contrasting behavior by viewing augmentation during training as a problem of domain generalization, and further propose Diverse Augmentation-based Joint Adversarial Training (DAJAT) to use data augmentations effectively in adversarial training. We aim to handle the conflicting goals of enhancing the diversity of the training dataset and training with data that is close to the test distribution by using a combination of simple and complex augmentations with separate batch normalization layers during training. We further utilize the popular Jensen-Shannon divergence loss to encourage the \emph{joint} learning of the \emph{diverse augmentations}, thereby allowing simple augmentations to guide the learning of complex ones. Lastly, to improve the computational efficiency of the proposed method, we propose and utilize a two-step defense, Ascending Constraint Adversarial Training (ACAT), that uses an increasing epsilon schedule and weight-space smoothing to prevent gradient masking. The proposed method DAJAT achieves substantially better robustness-accuracy trade-off when compared to existing methods on the RobustBench Leaderboard on ResNet-18 and WideResNet-34-10. The code for implementing DAJAT is available here: https://github.com/val-iisc/DAJAT

Chat is not available.