SOTAVerified

Universalization of any adversarial attack using very few test examples

2020-05-18Code Available0· sign in to hype

Sandesh Kamath, Amit Deshpande, K V Subrahmanyam, Vineeth N Balasubramanian

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Deep learning models are known to be vulnerable not only to input-dependent adversarial attacks but also to input-agnostic or universal adversarial attacks. Dezfooli et al. Dezfooli17,Dezfooli17anal construct universal adversarial attack on a given model by looking at a large number of training data points and the geometry of the decision boundary near them. Subsequent work Khrulkov18 constructs universal attack by looking only at test examples and intermediate layers of the given model. In this paper, we propose a simple universalization technique to take any input-dependent adversarial attack and construct a universal attack by only looking at very few adversarial test examples. We do not require details of the given model and have negligible computational overhead for universalization. We theoretically justify our universalization technique by a spectral property common to many input-dependent adversarial perturbations, e.g., gradients, Fast Gradient Sign Method (FGSM) and DeepFool. Using matrix concentration inequalities and spectral perturbation bounds, we show that the top singular vector of input-dependent adversarial directions on a small test sample gives an effective and simple universal adversarial attack. For VGG16 and VGG19 models trained on ImageNet, our simple universalization of Gradient, FGSM, and DeepFool perturbations using a test sample of 64 images gives fooling rates comparable to state-of-the-art universal attacks Dezfooli17,Khrulkov18 for reasonable norms of perturbation. Code available at https://github.com/ksandeshk/svd-uap .

Tasks

Reproductions