SOTAVerified

On Norm-Agnostic Robustness of Adversarial Training

2019-05-15Unverified0· sign in to hype

Bai Li, Changyou Chen, Wenlin Wang, Lawrence Carin

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Adversarial examples are carefully perturbed in-puts for fooling machine learning models. A well-acknowledged defense method against such examples is adversarial training, where adversarial examples are injected into training data to increase robustness. In this paper, we propose a new attack to unveil an undesired property of the state-of-the-art adversarial training, that is it fails to obtain robustness against perturbations in _2 and _ norms simultaneously. We discuss a possible solution to this issue and its limitations as well.

Tasks

Reproductions