SOTAVerified

Adversarial Robustness

Adversarial Robustness evaluates the vulnerabilities of machine learning models under various types of adversarial attacks.

Papers

Showing 451475 of 1746 papers

TitleStatusHype
Adversarial Masked Autoencoder Purifier with Defense Transferability0
Boosting Certified Robustness for Time Series Classification with Efficient Self-Ensemble0
Adversarially Robust Few-shot Learning via Parameter Co-distillation of Similarity and Class Concept Learners0
Chaos Theory and Adversarial Robustness0
Characterizing the adversarial vulnerability of speech self-supervised learning0
A Survey and Evaluation of Adversarial Attacks for Object Detection0
Effects of Loss Functions And Target Representations on Adversarial Robustness0
Class-Aware Domain Adaptation for Improving Adversarial Robustness0
Class-Aware Robust Adversarial Training for Object Detection0
Classifier Guidance Enhances Diffusion-based Adversarial Purification by Preserving Predictive Information0
A Framework for Verification of Wasserstein Adversarial Robustness0
Boosting Barely Robust Learners: A New Perspective on Adversarial Robustness0
Dynamical Low-Rank Compression of Neural Networks with Robustness under Adversarial Attacks0
SOAR: Second-Order Adversarial Regularization0
A Fundamental Accuracy--Robustness Trade-off in Regression and Classification0
Adversarial Prompt Distillation for Vision-Language Models0
Boosting Adversarial Robustness From The Perspective of Effective Margin Regularization0
Collective Robustness Certificates: Exploiting Interdependence in Graph Neural Networks0
Adversarially Robust Estimate and Risk Analysis in Linear Regression0
Dynamic Defense Approach for Adversarial Robustness in Deep Neural Networks via Stochastic Ensemble Smoothed Model0
Boosting Adversarial Robustness and Generalization with Structural Prior0
Boosting Accuracy and Robustness of Student Models via Adaptive Adversarial Distillation0
Adversarial Robustness through Local Linearization0
A Closer Look at the Adversarial Robustness of Information Bottleneck Models0
Adversarial Robustness through Dynamic Ensemble Learning0
Show:102550
← PrevPage 19 of 70Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeBERTa (single model)Accuracy0.61Unverified
2ALBERT (single model)Accuracy0.59Unverified
3T5 (single model)Accuracy0.57Unverified
4SMART_RoBERTa (single model)Accuracy0.54Unverified
5FreeLB (single model)Accuracy0.5Unverified
6RoBERTa (single model)Accuracy0.5Unverified
7InfoBERT (single model)Accuracy0.46Unverified
8ELECTRA (single model)Accuracy0.42Unverified
9BERT (single model)Accuracy0.34Unverified
10SMART_BERT (single model)Accuracy0.3Unverified
#ModelMetricClaimedVerifiedStatus
1Mixed classifierAccuracy95.23Unverified
2Stochastic-LWTA/PGD/WideResNet-34-10Accuracy92.26Unverified
3Stochastic-LWTA/PGD/WideResNet-34-5Accuracy91.88Unverified
4GLOT-DRAccuracy84.13Unverified
5TRADES-ANCRA/ResNet18Accuracy81.7Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet-50 (SGD, Cosine)Accuracy77.4Unverified
2ResNet-50 (SGD, Step)Accuracy76.9Unverified
3DeiT-S (AdamW, Cosine)Accuracy76.8Unverified
4ResNet-50 (AdamW, Cosine)Accuracy76.4Unverified
#ModelMetricClaimedVerifiedStatus
1DeiT-S (AdamW, Cosine)Accuracy12.2Unverified
2ResNet-50 (SGD, Cosine)Accuracy3.3Unverified
3ResNet-50 (SGD, Step)Accuracy3.2Unverified
4ResNet-50 (AdamW, Cosine)Accuracy3.1Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet-50 (AdamW, Cosine)mean Corruption Error (mCE)59.3Unverified
2ResNet-50 (SGD, Step)mean Corruption Error (mCE)57.9Unverified
3ResNet-50 (SGD, Cosine)mean Corruption Error (mCE)56.9Unverified
4DeiT-S (AdamW, Cosine)mean Corruption Error (mCE)48Unverified
#ModelMetricClaimedVerifiedStatus
1DeiT-S (AdamW, Cosine)Accuracy13Unverified
2ResNet-50 (SGD, Cosine)Accuracy8.4Unverified
3ResNet-50 (SGD, Step)Accuracy8.3Unverified
4ResNet-50 (AdamW, Cosine)Accuracy8.1Unverified
#ModelMetricClaimedVerifiedStatus
1Mixed ClassifierClean Accuracy85.21Unverified
2ResNet18/MART-ANCRAClean Accuracy60.1Unverified