SOTAVerified

Provable Robustness of Adversarial Training for Learning Halfspaces with Noise

2021-04-19Unverified0· sign in to hype

Difan Zou, Spencer Frei, Quanquan Gu

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We analyze the properties of adversarial training for learning adversarially robust halfspaces in the presence of agnostic label noise. Denoting OPT_p,r as the best robust classification error achieved by a halfspace that is robust to perturbations of _p balls of radius r, we show that adversarial training on the standard binary cross-entropy loss yields adversarially robust halfspaces up to (robust) classification error O(OPT_2,r) for p=2, and O(d^1/4 OPT_, r + d^1/2 OPT_,r) when p=. Our results hold for distributions satisfying anti-concentration properties enjoyed by log-concave isotropic distributions among others. We additionally show that if one instead uses a nonconvex sigmoidal loss, adversarial training yields halfspaces with an improved robust classification error of O(OPT_2,r) for p=2, and O(d^1/4OPT_, r) when p=. To the best of our knowledge, this is the first work to show that adversarial training provably yields robust classifiers in the presence of noise.

Tasks

Reproductions