SOTAVerified

Adversarial Robustness

Adversarial Robustness evaluates the vulnerabilities of machine learning models under various types of adversarial attacks.

Papers

Showing 101125 of 1746 papers

TitleStatusHype
Comparing the Robustness of Modern No-Reference Image- and Video-Quality Metrics to Adversarial AttacksCode1
Composite Adversarial AttacksCode1
CosPGD: an efficient white-box adversarial attack for pixel-wise prediction tasksCode1
Decision-based Black-box Attack Against Vision Transformers via Patch-wise Adversarial RemovalCode1
Decoupled Kullback-Leibler Divergence LossCode1
Defense Against Adversarial Attacks on No-Reference Image Quality Models with Gradient Norm RegularizationCode1
AdvDrop: Adversarial Attack to DNNs by Dropping InformationCode1
An Embarrassingly Simple Backdoor Attack on Self-supervised LearningCode1
Achieving robustness in classification using optimal transport with hinge regularizationCode1
Adversarial Machine Learning: Bayesian PerspectivesCode1
GenoArmory: A Unified Evaluation Framework for Adversarial Attacks on Genomic Foundation ModelsCode1
Adversarial Robustness of Bottleneck Injected Deep Neural Networks for Task-Oriented CommunicationCode1
Drop-Bottleneck: Learning Discrete Compressed Representation for Noise-Robust ExplorationCode1
Efficient Exact Verification of Binarized Neural NetworksCode1
Adversarial Attack and Defense in Deep RankingCode1
Adversarial Prompt Tuning for Vision-Language ModelsCode1
Adversarial Pruning: A Survey and Benchmark of Pruning Methods for Adversarial RobustnessCode1
Adversarial Reasoning at Jailbreaking TimeCode1
Adversarial Attack on Deep Learning-Based Splice LocalizationCode1
Enhancing adversarial robustness in Natural Language Inference using explanationsCode1
Adversarial Robustification via Text-to-Image Diffusion ModelsCode1
Enhancing Adversarial Robustness via Test-time Transformation EnsemblingCode1
Evaluating the Adversarial Robustness of Adaptive Test-time DefensesCode1
Adversarial Image Color Transformations in Explicit Color Filter SpaceCode1
Adversarial Robustness in Graph Neural Networks: A Hamiltonian ApproachCode1
Show:102550
← PrevPage 5 of 70Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeBERTa (single model)Accuracy0.61Unverified
2ALBERT (single model)Accuracy0.59Unverified
3T5 (single model)Accuracy0.57Unverified
4SMART_RoBERTa (single model)Accuracy0.54Unverified
5FreeLB (single model)Accuracy0.5Unverified
6RoBERTa (single model)Accuracy0.5Unverified
7InfoBERT (single model)Accuracy0.46Unverified
8ELECTRA (single model)Accuracy0.42Unverified
9BERT (single model)Accuracy0.34Unverified
10SMART_BERT (single model)Accuracy0.3Unverified
#ModelMetricClaimedVerifiedStatus
1Mixed classifierAccuracy95.23Unverified
2Stochastic-LWTA/PGD/WideResNet-34-10Accuracy92.26Unverified
3Stochastic-LWTA/PGD/WideResNet-34-5Accuracy91.88Unverified
4GLOT-DRAccuracy84.13Unverified
5TRADES-ANCRA/ResNet18Accuracy81.7Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet-50 (SGD, Cosine)Accuracy77.4Unverified
2ResNet-50 (SGD, Step)Accuracy76.9Unverified
3DeiT-S (AdamW, Cosine)Accuracy76.8Unverified
4ResNet-50 (AdamW, Cosine)Accuracy76.4Unverified
#ModelMetricClaimedVerifiedStatus
1DeiT-S (AdamW, Cosine)Accuracy12.2Unverified
2ResNet-50 (SGD, Cosine)Accuracy3.3Unverified
3ResNet-50 (SGD, Step)Accuracy3.2Unverified
4ResNet-50 (AdamW, Cosine)Accuracy3.1Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet-50 (AdamW, Cosine)mean Corruption Error (mCE)59.3Unverified
2ResNet-50 (SGD, Step)mean Corruption Error (mCE)57.9Unverified
3ResNet-50 (SGD, Cosine)mean Corruption Error (mCE)56.9Unverified
4DeiT-S (AdamW, Cosine)mean Corruption Error (mCE)48Unverified
#ModelMetricClaimedVerifiedStatus
1DeiT-S (AdamW, Cosine)Accuracy13Unverified
2ResNet-50 (SGD, Cosine)Accuracy8.4Unverified
3ResNet-50 (SGD, Step)Accuracy8.3Unverified
4ResNet-50 (AdamW, Cosine)Accuracy8.1Unverified
#ModelMetricClaimedVerifiedStatus
1Mixed ClassifierClean Accuracy85.21Unverified
2ResNet18/MART-ANCRAClean Accuracy60.1Unverified