SOTAVerified

Adversarial Robustness

Adversarial Robustness evaluates the vulnerabilities of machine learning models under various types of adversarial attacks.

Papers

Showing 326350 of 1746 papers

TitleStatusHype
Analyzing Adversarial Attacks Against Deep Learning for Intrusion Detection in IoT Networks0
Confronting the Reproducibility Crisis: A Case Study of Challenges in Cybersecurity AI0
An Adversarial Robustness Benchmark for Enterprise Network Intrusion Detection0
A Comparative Analysis of Adversarial Robustness for Quantum and Classical Machine Learning Models0
A Multiclass Boosting Framework for Achieving Fast and Provable Adversarial Robustness0
Adversarial Robustness Across Representation Spaces0
AdvCat: Domain-Agnostic Robustness Assessment for Cybersecurity-Critical Applications with Categorical Inputs0
A More Biologically Plausible Local Learning Rule for ANNs0
A margin-based replacement for cross-entropy loss0
Adversarial Attacks and Defenses for Speech Recognition Systems0
ALMA: Aggregated Lipschitz Maximization Attack on Auto-encoders0
Algorithmic Bias and Data Bias: Understanding the Relation between Distributionally Robust Optimization and Data Curation0
Adversarial Risk and the Dangers of Evaluating Against Weak Attacks0
aiXamine: Simplified LLM Safety and Security0
AI-Compass: A Comprehensive and Effective Multi-module Testing Tool for AI Systems0
CARES: Comprehensive Evaluation of Safety and Adversarial Robustness in Medical LLMs0
Conflict-Aware Adversarial Training0
A Hybrid Defense Strategy for Boosting Adversarial Robustness in Vision-Language Models0
A Holistic Assessment of the Reliability of Machine Learning Systems0
A High Dimensional Statistical Model for Adversarial Training: Geometry and Trade-Offs0
A Closer Look at the Adversarial Robustness of Information Bottleneck Models0
A Geometrical Approach to Evaluate the Adversarial Robustness of Deep Neural Networks0
A Fundamental Accuracy--Robustness Trade-off in Regression and Classification0
A Frequency Perspective of Adversarial Robustness0
Adversarial Prompt Distillation for Vision-Language Models0
Show:102550
← PrevPage 14 of 70Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeBERTa (single model)Accuracy0.61Unverified
2ALBERT (single model)Accuracy0.59Unverified
3T5 (single model)Accuracy0.57Unverified
4SMART_RoBERTa (single model)Accuracy0.54Unverified
5FreeLB (single model)Accuracy0.5Unverified
6RoBERTa (single model)Accuracy0.5Unverified
7InfoBERT (single model)Accuracy0.46Unverified
8ELECTRA (single model)Accuracy0.42Unverified
9BERT (single model)Accuracy0.34Unverified
10SMART_BERT (single model)Accuracy0.3Unverified
#ModelMetricClaimedVerifiedStatus
1Mixed classifierAccuracy95.23Unverified
2Stochastic-LWTA/PGD/WideResNet-34-10Accuracy92.26Unverified
3Stochastic-LWTA/PGD/WideResNet-34-5Accuracy91.88Unverified
4GLOT-DRAccuracy84.13Unverified
5TRADES-ANCRA/ResNet18Accuracy81.7Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet-50 (SGD, Cosine)Accuracy77.4Unverified
2ResNet-50 (SGD, Step)Accuracy76.9Unverified
3DeiT-S (AdamW, Cosine)Accuracy76.8Unverified
4ResNet-50 (AdamW, Cosine)Accuracy76.4Unverified
#ModelMetricClaimedVerifiedStatus
1DeiT-S (AdamW, Cosine)Accuracy12.2Unverified
2ResNet-50 (SGD, Cosine)Accuracy3.3Unverified
3ResNet-50 (SGD, Step)Accuracy3.2Unverified
4ResNet-50 (AdamW, Cosine)Accuracy3.1Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet-50 (AdamW, Cosine)mean Corruption Error (mCE)59.3Unverified
2ResNet-50 (SGD, Step)mean Corruption Error (mCE)57.9Unverified
3ResNet-50 (SGD, Cosine)mean Corruption Error (mCE)56.9Unverified
4DeiT-S (AdamW, Cosine)mean Corruption Error (mCE)48Unverified
#ModelMetricClaimedVerifiedStatus
1DeiT-S (AdamW, Cosine)Accuracy13Unverified
2ResNet-50 (SGD, Cosine)Accuracy8.4Unverified
3ResNet-50 (SGD, Step)Accuracy8.3Unverified
4ResNet-50 (AdamW, Cosine)Accuracy8.1Unverified
#ModelMetricClaimedVerifiedStatus
1Mixed ClassifierClean Accuracy85.21Unverified
2ResNet18/MART-ANCRAClean Accuracy60.1Unverified