SOTAVerified

Adversarial Attack

An Adversarial Attack is a technique to find a perturbation that changes the prediction of a machine learning model. The perturbation can be very small and imperceptible to human eyes.

Source: Recurrent Attention Model with Log-Polar Mapping is Robust against Adversarial Attacks

Papers

Showing 10261050 of 1808 papers

TitleStatusHype
Learning to Defend by Learning to Attack0
Learning to Defense by Learning to Attack0
Learning to Detect Adversarial Examples Based on Class Scores0
Zeroth-Order Stochastic Alternating Direction Method of Multipliers for Nonconvex Nonsmooth Optimization0
Visual Adversarial Attack on Vision-Language Models for Autonomous Driving0
Visual Attack and Defense on Text0
Thundernna: a white box adversarial attack0
Thwarting finite difference adversarial attacks with output randomization0
Time-aware Gradient Attack on Dynamic Network Link Prediction0
Left-right Discrepancy for Adversarial Attack on Stereo Networks0
Less is More: A Stealthy and Efficient Adversarial Attack Method for DRL-based Autonomous Driving Policies0
Less is More: Understanding Word-level Textual Adversarial Attack via n-gram Frequency Descend0
To be Robust and to be Fair: Aligning Fairness with Robustness0
LFAA: Crafting Transferable Targeted Adversarial Examples with Low-Frequency Perturbations0
Patch Synthesis for Property Repair of Deep Neural Networks0
To Each (Textual Sequence) Its Own: Improving Memorized-Data Unlearning in Large Language Models0
Light Lies: Optical Adversarial Attack0
BOSH: An Efficient Meta Algorithm for Decision-based Attacks0
OTAD: An Optimal Transport-Induced Robust Model for Agnostic Adversarial Attack0
Limited Budget Adversarial Attack Against Online Image Stream0
Linear Backpropagation Leads to Faster Convergence0
Linear system security -- detection and correction of adversarial attacks in the noise-free case0
LLMs Can Defend Themselves Against Jailbreaking in a Practical Manner: A Vision Paper0
Towards Accuracy-Fairness Paradox: Adversarial Example-based Data Augmentation for Visual Debiasing0
Local Competition and Stochasticity for Adversarial Robustness in Deep Learning0
Show:102550
← PrevPage 42 of 73Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Xu et al.Attack: PGD2078.68Unverified
23-ensemble of multi-resolution self-ensemblesAttack: AutoAttack78.13Unverified
3TRADES-ANCRA/ResNet18Attack: AutoAttack59.7Unverified
4AdvTraining [madry2018]Attack: PGD2048.44Unverified
5TRADES [zhang2019b]Attack: PGD2045.9Unverified
6XU-NetRobust Accuracy1Unverified
#ModelMetricClaimedVerifiedStatus
13-ensemble of multi-resolution self-ensemblesAttack: AutoAttack51.28Unverified
2multi-resolution self-ensemblesAttack: AutoAttack47.85Unverified