SOTAVerified

Adversarial Attack

An Adversarial Attack is a technique to find a perturbation that changes the prediction of a machine learning model. The perturbation can be very small and imperceptible to human eyes.

Source: Recurrent Attention Model with Log-Polar Mapping is Robust against Adversarial Attacks

Papers

Showing 2650 of 1808 papers

TitleStatusHype
ScoreAdv: Score-based Targeted Generation of Natural Adversarial Examples via Diffusion ModelsCode1
Adversarial Attacks and Detection in Visual Place Recognition for Safer Robot NavigationCode1
Learning Safety Constraints for Large Language ModelsCode1
3D Gaussian Splat VulnerabilitiesCode1
SafeScientist: Toward Risk-Aware Scientific Discoveries by LLM AgentsCode1
Audio Jailbreak Attacks: Exposing Vulnerabilities in SpeechGPT in a White-Box FrameworkCode1
GenoArmory: A Unified Evaluation Framework for Adversarial Attacks on Genomic Foundation ModelsCode1
Fast and Low-Cost Genomic Foundation Models via Outlier RemovalCode1
sudo rm -rf agentic_securityCode1
CyberLLMInstruct: A New Dataset for Analysing Safety of Fine-Tuned LLMs Using Cyber Security DataCode1
Data-free Universal Adversarial Perturbation with Pseudo-semantic PriorCode1
Iron Sharpens Iron: Defending Against Attacks in Machine-Generated Text Detection with Adversarial TrainingCode1
To Think or Not to Think: Exploring the Unthinking Vulnerability in Large Reasoning ModelsCode1
HateBench: Benchmarking Hate Speech Detectors on LLM-Generated Content and Hate CampaignsCode1
Physics-Based Adversarial Attack on Near-Infrared Human Detector for Nighttime Surveillance Camera SystemsCode1
Human-in-the-Loop Generation of Adversarial Texts: A Case Study on Tibetan ScriptCode1
A2RNet: Adversarial Attack Resilient Network for Robust Infrared and Visible Image FusionCode1
Adversarial Vulnerabilities in Large Language Models for Time Series ForecastingCode1
Exploiting the Index Gradients for Optimization-Based Jailbreaking on Large Language ModelsCode1
Hiding Faces in Plain Sight: Defending DeepFakes by Disrupting Face DetectionCode1
Semantic-Aligned Adversarial Evolution Triangle for High-Transferability Vision-Language AttackCode1
Transferable Adversarial Attacks on SAM and Its Downstream ModelsCode1
Malacopula: adversarial automatic speaker verification attacks using a neural-based generalised Hammerstein modelCode1
Ensemble everything everywhere: Multi-scale aggregation for adversarial robustnessCode1
Guardians of Image Quality: Benchmarking Defenses Against Adversarial Attacks on Image Quality MetricsCode1
Show:102550
← PrevPage 2 of 73Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Xu et al.Attack: PGD2078.68Unverified
23-ensemble of multi-resolution self-ensemblesAttack: AutoAttack78.13Unverified
3TRADES-ANCRA/ResNet18Attack: AutoAttack59.7Unverified
4AdvTraining [madry2018]Attack: PGD2048.44Unverified
5TRADES [zhang2019b]Attack: PGD2045.9Unverified
6XU-NetRobust Accuracy1Unverified
#ModelMetricClaimedVerifiedStatus
13-ensemble of multi-resolution self-ensemblesAttack: AutoAttack51.28Unverified
2multi-resolution self-ensemblesAttack: AutoAttack47.85Unverified