SOTAVerified

Adversarial Robustness

Adversarial Robustness evaluates the vulnerabilities of machine learning models under various types of adversarial attacks.

Papers

Showing 601625 of 1746 papers

TitleStatusHype
A Study on the Uncertainty of Convolutional Layers in Deep Neural Networks0
Adversarial Robustness of Deep Learning: Theory, Algorithms, and Applications0
Few-Shot Adversarial Low-Rank Fine-Tuning of Vision-Language Models0
Finding a human-like classifier0
Formalizing Generalization and Adversarial Robustness of Neural Networks to Weight Perturbations0
Generalizability of Adversarial Robustness Under Distribution Shifts0
Feature Losses for Adversarial Robustness0
Adaptive Batch Normalization Networks for Adversarial Robustness0
Feature Prioritization and Regularization Improve Standard Accuracy and Adversarial Robustness0
Associative Adversarial Learning Based on Selective Attack0
Non-adversarial Robustness of Deep Learning Methods for Computer Vision0
Assessing the Adversarial Robustness of Monte Carlo and Distillation Methods for Deep Bayesian Neural Network Classification0
Assessing Adversarial Robustness of Large Language Models: An Empirical Study0
Adversarial Robustness May Be at Odds With Simplicity0
A Spectral Perspective towards Understanding and Improving Adversarial Robustness0
Ensemble Adversarial Defense via Integration of Multiple Dispersed Low Curvature Models0
Improving Transformation-based Defenses against Adversarial Examples with First-order Perturbations0
Adversarial examples attack based on random warm restart mechanism and improved Nesterov momentum0
Feature Distillation With Guided Adversarial Contrastive Learning0
Enhancing the "Immunity" of Mixture-of-Experts Networks for Adversarial Defense0
Enhancing the Antidote: Improved Pointwise Certifications against Poisoning Attacks0
A Simple Framework to Enhance the Adversarial Robustness of Deep Learning-based Intrusion Detection System0
Enhancing Quantum Adversarial Robustness by Randomized Encodings0
A Simple Fine-tuning Is All You Need: Towards Robust Deep Learning Via Adversarial Fine-tuning0
Adversarial Robustness is at Odds with Lazy Training0
Show:102550
← PrevPage 25 of 70Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeBERTa (single model)Accuracy0.61Unverified
2ALBERT (single model)Accuracy0.59Unverified
3T5 (single model)Accuracy0.57Unverified
4SMART_RoBERTa (single model)Accuracy0.54Unverified
5FreeLB (single model)Accuracy0.5Unverified
6RoBERTa (single model)Accuracy0.5Unverified
7InfoBERT (single model)Accuracy0.46Unverified
8ELECTRA (single model)Accuracy0.42Unverified
9BERT (single model)Accuracy0.34Unverified
10SMART_BERT (single model)Accuracy0.3Unverified
#ModelMetricClaimedVerifiedStatus
1Mixed classifierAccuracy95.23Unverified
2Stochastic-LWTA/PGD/WideResNet-34-10Accuracy92.26Unverified
3Stochastic-LWTA/PGD/WideResNet-34-5Accuracy91.88Unverified
4GLOT-DRAccuracy84.13Unverified
5TRADES-ANCRA/ResNet18Accuracy81.7Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet-50 (SGD, Cosine)Accuracy77.4Unverified
2ResNet-50 (SGD, Step)Accuracy76.9Unverified
3DeiT-S (AdamW, Cosine)Accuracy76.8Unverified
4ResNet-50 (AdamW, Cosine)Accuracy76.4Unverified
#ModelMetricClaimedVerifiedStatus
1DeiT-S (AdamW, Cosine)Accuracy12.2Unverified
2ResNet-50 (SGD, Cosine)Accuracy3.3Unverified
3ResNet-50 (SGD, Step)Accuracy3.2Unverified
4ResNet-50 (AdamW, Cosine)Accuracy3.1Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet-50 (AdamW, Cosine)mean Corruption Error (mCE)59.3Unverified
2ResNet-50 (SGD, Step)mean Corruption Error (mCE)57.9Unverified
3ResNet-50 (SGD, Cosine)mean Corruption Error (mCE)56.9Unverified
4DeiT-S (AdamW, Cosine)mean Corruption Error (mCE)48Unverified
#ModelMetricClaimedVerifiedStatus
1DeiT-S (AdamW, Cosine)Accuracy13Unverified
2ResNet-50 (SGD, Cosine)Accuracy8.4Unverified
3ResNet-50 (SGD, Step)Accuracy8.3Unverified
4ResNet-50 (AdamW, Cosine)Accuracy8.1Unverified
#ModelMetricClaimedVerifiedStatus
1Mixed ClassifierClean Accuracy85.21Unverified
2ResNet18/MART-ANCRAClean Accuracy60.1Unverified