SOTAVerified

Adversarial Robustness

Adversarial Robustness evaluates the vulnerabilities of machine learning models under various types of adversarial attacks.

Papers

Showing 876900 of 1746 papers

TitleStatusHype
Perceptual Deep Neural Networks: Adversarial Robustness through Input Recreation0
Performance and Non-adversarial Robustness of the Segment Anything Model 2 in Surgical Video Segmentation0
Perlin Noise Improve Adversarial Robustness0
Perturbation-Invariant Adversarial Training for Neural Ranking Models: Improving the Effectiveness-Robustness Trade-Off0
Perturbation Type Categorization for Multiple _p Bounded Adversarial Robustness0
Phase-shifted Adversarial Training0
Physical-layer Adversarial Robustness for Deep Learning-based Semantic Communications0
Visually Adversarial Attacks and Defenses in the Physical World: A Survey0
Planting Undetectable Backdoors in Machine Learning Models0
Playing it Safe: Adversarial Robustness with an Abstain Option0
Poisoning Evasion: Symbiotic Adversarial Robustness for Graph Neural Networks0
Certifiably Robust Reinforcement Learning through Model-Based Abstract Interpretation0
Policy Smoothing for Provably Robust Reinforcement Learning0
Power up! Robust Graph Convolutional Network based on Graph Powering0
Practical Convex Formulation of Robust One-hidden-layer Neural Network Training0
Pre-trained Model Guided Mixture Knowledge Distillation for Adversarial Federated Learning0
Principal Eigenvalue Regularization for Improved Worst-Class Certified Robustness of Smoothed Classifiers0
PRISON: Unmasking the Criminal Potential of Large Language Models0
Probabilistic Robustness in Deep Learning: A Concise yet Comprehensive Guide0
Probing the Robustness of Vision-Language Pretrained Models: A Multimodal Adversarial Attack Approach0
Promoting Robustness of Randomized Smoothing: Two Cost-Effective Approaches0
Proper Measure for Adversarial Robustness0
Prototypical Examples in Deep Learning: Metrics, Characteristics, and Utility0
Pro-tuning: Unified Prompt Tuning for Vision Tasks0
Provable Adversarial Robustness for Group Equivariant Tasks: Graphs, Point Clouds, Molecules, and More0
Show:102550
← PrevPage 36 of 70Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeBERTa (single model)Accuracy0.61Unverified
2ALBERT (single model)Accuracy0.59Unverified
3T5 (single model)Accuracy0.57Unverified
4SMART_RoBERTa (single model)Accuracy0.54Unverified
5FreeLB (single model)Accuracy0.5Unverified
6RoBERTa (single model)Accuracy0.5Unverified
7InfoBERT (single model)Accuracy0.46Unverified
8ELECTRA (single model)Accuracy0.42Unverified
9BERT (single model)Accuracy0.34Unverified
10SMART_BERT (single model)Accuracy0.3Unverified
#ModelMetricClaimedVerifiedStatus
1Mixed classifierAccuracy95.23Unverified
2Stochastic-LWTA/PGD/WideResNet-34-10Accuracy92.26Unverified
3Stochastic-LWTA/PGD/WideResNet-34-5Accuracy91.88Unverified
4GLOT-DRAccuracy84.13Unverified
5TRADES-ANCRA/ResNet18Accuracy81.7Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet-50 (SGD, Cosine)Accuracy77.4Unverified
2ResNet-50 (SGD, Step)Accuracy76.9Unverified
3DeiT-S (AdamW, Cosine)Accuracy76.8Unverified
4ResNet-50 (AdamW, Cosine)Accuracy76.4Unverified
#ModelMetricClaimedVerifiedStatus
1DeiT-S (AdamW, Cosine)Accuracy12.2Unverified
2ResNet-50 (SGD, Cosine)Accuracy3.3Unverified
3ResNet-50 (SGD, Step)Accuracy3.2Unverified
4ResNet-50 (AdamW, Cosine)Accuracy3.1Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet-50 (AdamW, Cosine)mean Corruption Error (mCE)59.3Unverified
2ResNet-50 (SGD, Step)mean Corruption Error (mCE)57.9Unverified
3ResNet-50 (SGD, Cosine)mean Corruption Error (mCE)56.9Unverified
4DeiT-S (AdamW, Cosine)mean Corruption Error (mCE)48Unverified
#ModelMetricClaimedVerifiedStatus
1DeiT-S (AdamW, Cosine)Accuracy13Unverified
2ResNet-50 (SGD, Cosine)Accuracy8.4Unverified
3ResNet-50 (SGD, Step)Accuracy8.3Unverified
4ResNet-50 (AdamW, Cosine)Accuracy8.1Unverified
#ModelMetricClaimedVerifiedStatus
1Mixed ClassifierClean Accuracy85.21Unverified
2ResNet18/MART-ANCRAClean Accuracy60.1Unverified