SOTAVerified

Adversarial Robustness

Adversarial Robustness evaluates the vulnerabilities of machine learning models under various types of adversarial attacks.

Papers

Showing 351375 of 1746 papers

TitleStatusHype
A Frequency Perspective of Adversarial Robustness0
Adversarial Prompt Distillation for Vision-Language Models0
A Framework for Verification of Wasserstein Adversarial Robustness0
A Flat Minima Perspective on Understanding Augmentations and Model Robustness0
AdvFunMatch: When Consistent Teaching Meets Adversarial Robustness0
A Finer Calibration Analysis for Adversarial Robustness0
Affine-Invariant Robust Training0
Adversarial Amendment is the Only Force Capable of Transforming an Enemy into a Friend0
Deadwooding: Robust Global Pruning for Deep Neural Networks0
Adversarial Masked Autoencoder Purifier with Defense Transferability0
DataFreeShield: Defending Adversarial Attacks without Training Data0
AdvFilter: Predictive Perturbation-aware Filtering against Adversarial Attack via Multi-domain Learning0
Adversarial Alignment for LLMs Requires Simpler, Reproducible, and More Measurable Objectives0
Mining Data Impressions from Deep Models as Substitute for the Unavailable Training Data0
Deciphering the Definition of Adversarial Robustness for post-hoc OOD Detectors0
DecodingTrust: A Comprehensive Assessment of Trustworthiness in GPT Models0
Adversary Agnostic Robust Deep Reinforcement Learning0
Cross-Entropy Loss Functions: Theoretical Analysis and Applications0
CSTAR: Towards Compact and STructured Deep Neural Networks with Adversarial Robustness0
Adversarially Robust Video Perception by Seeing Motion0
Adversarial Training via Adaptive Knowledge Amalgamation of an Ensemble of Teachers0
Adversarial alignment: Breaking the trade-off between the strength of an attack and its relevance to human perception0
Curvature Dynamic Black-box Attack: revisiting adversarial robustness via dynamic curvature estimation0
Adversarial Training for Face Recognition Systems using Contrastive Adversarial Learning and Triplet Loss Fine-tuning0
Adversarially-Robust TD Learning with Markovian Data: Finite-Time Rates and Fundamental Limits0
Show:102550
← PrevPage 15 of 70Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeBERTa (single model)Accuracy0.61Unverified
2ALBERT (single model)Accuracy0.59Unverified
3T5 (single model)Accuracy0.57Unverified
4SMART_RoBERTa (single model)Accuracy0.54Unverified
5FreeLB (single model)Accuracy0.5Unverified
6RoBERTa (single model)Accuracy0.5Unverified
7InfoBERT (single model)Accuracy0.46Unverified
8ELECTRA (single model)Accuracy0.42Unverified
9BERT (single model)Accuracy0.34Unverified
10SMART_BERT (single model)Accuracy0.3Unverified
#ModelMetricClaimedVerifiedStatus
1Mixed classifierAccuracy95.23Unverified
2Stochastic-LWTA/PGD/WideResNet-34-10Accuracy92.26Unverified
3Stochastic-LWTA/PGD/WideResNet-34-5Accuracy91.88Unverified
4GLOT-DRAccuracy84.13Unverified
5TRADES-ANCRA/ResNet18Accuracy81.7Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet-50 (SGD, Cosine)Accuracy77.4Unverified
2ResNet-50 (SGD, Step)Accuracy76.9Unverified
3DeiT-S (AdamW, Cosine)Accuracy76.8Unverified
4ResNet-50 (AdamW, Cosine)Accuracy76.4Unverified
#ModelMetricClaimedVerifiedStatus
1DeiT-S (AdamW, Cosine)Accuracy12.2Unverified
2ResNet-50 (SGD, Cosine)Accuracy3.3Unverified
3ResNet-50 (SGD, Step)Accuracy3.2Unverified
4ResNet-50 (AdamW, Cosine)Accuracy3.1Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet-50 (AdamW, Cosine)mean Corruption Error (mCE)59.3Unverified
2ResNet-50 (SGD, Step)mean Corruption Error (mCE)57.9Unverified
3ResNet-50 (SGD, Cosine)mean Corruption Error (mCE)56.9Unverified
4DeiT-S (AdamW, Cosine)mean Corruption Error (mCE)48Unverified
#ModelMetricClaimedVerifiedStatus
1DeiT-S (AdamW, Cosine)Accuracy13Unverified
2ResNet-50 (SGD, Cosine)Accuracy8.4Unverified
3ResNet-50 (SGD, Step)Accuracy8.3Unverified
4ResNet-50 (AdamW, Cosine)Accuracy8.1Unverified
#ModelMetricClaimedVerifiedStatus
1Mixed ClassifierClean Accuracy85.21Unverified
2ResNet18/MART-ANCRAClean Accuracy60.1Unverified