RoBIC: A benchmark suite for assessing classifiers robustness
2021-02-10Code Available0· sign in to hype
Thibault Maho, Benoît Bonnet, Teddy Furon, Erwan Le Merrer
Code Available — Be the first to reproduce this paper.
ReproduceCode
- gitlab.inria.fr/tmaho/robustness_benchmarkOfficialnone★ 0
Abstract
Many defenses have emerged with the development of adversarial attacks. Models must be objectively evaluated accordingly. This paper systematically tackles this concern by proposing a new parameter-free benchmark we coin RoBIC. RoBIC fairly evaluates the robustness of image classifiers using a new half-distortion measure. It gauges the robustness of the network against white and black box attacks, independently of its accuracy. RoBIC is faster than the other available benchmarks. We present the significant differences in the robustness of 16 recent models as assessed by RoBIC.