SOTAVerified

Adversarial Training for Free!

2019-04-29NeurIPS 2019Code Available1· sign in to hype

Ali Shafahi, Mahyar Najibi, Amin Ghiasi, Zheng Xu, John Dickerson, Christoph Studer, Larry S. Davis, Gavin Taylor, Tom Goldstein

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Adversarial training, in which a network is trained on adversarial examples, is one of the few defenses against adversarial attacks that withstands strong attacks. Unfortunately, the high cost of generating strong adversarial examples makes standard adversarial training impractical on large-scale problems like ImageNet. We present an algorithm that eliminates the overhead cost of generating adversarial examples by recycling the gradient information computed when updating model parameters. Our "free" adversarial training algorithm achieves comparable robustness to PGD adversarial training on the CIFAR-10 and CIFAR-100 datasets at negligible additional cost compared to natural training, and can be 7 to 30 times faster than other strong adversarial training methods. Using a single workstation with 4 P100 GPUs and 2 days of runtime, we can train a robust model for the large-scale ImageNet classification task that maintains 40% accuracy against PGD attacks. The code is available at https://github.com/ashafahi/free_adv_train.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
ImageNet (non-targeted PGD, max perturbation=4)ResNet-152 free-m=4Accuracy36Unverified
ImageNet (non-targeted PGD, max perturbation=4)ResNet-101 free-m=4Accuracy34.3Unverified
ImageNet (non-targeted PGD, max perturbation=4)ResNet-50 free-m=4Accuracy31.8Unverified

Reproductions