SOTAVerified

Attention Masks Help Adversarial Attacks to Bypass Safety Detectors

2024-11-07Code Available0· sign in to hype

Yunfan Shi

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Despite recent research advancements in adversarial attack methods, current approaches against XAI monitors are still discoverable and slower. In this paper, we present an adaptive framework for attention mask generation to enable stealthy, explainable and efficient PGD image classification adversarial attack under XAI monitors. Specifically, we utilize mutation XAI mixture and multitask self-supervised X-UNet for attention mask generation to guide PGD attack. Experiments on MNIST (MLP), CIFAR-10 (AlexNet) have shown that our system can outperform benchmark PGD, Sparsefool and SOTA SINIFGSM in balancing among stealth, efficiency and explainability which is crucial for effectively fooling SOTA defense protected classifiers.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
CIFAR-10XU-NetRobust Accuracy1Unverified

Reproductions