SOTAVerified

Adversarial Robustness

Adversarial Robustness evaluates the vulnerabilities of machine learning models under various types of adversarial attacks.

Papers

Showing 10511075 of 1746 papers

TitleStatusHype
MMARD: Improving the Min-Max Optimization Process in Adversarial Robustness Distillation0
MMDT: Decoding the Trustworthiness and Safety of Multimodal Foundation Models0
A More Biologically Plausible Local Learning Rule for ANNs0
Towards Adversarially Robust Deep Image Denoising0
Model-Agnostic Meta-Attack: Towards Reliable Evaluation of Adversarial Robustness0
A margin-based replacement for cross-entropy loss0
Out of Thin Air: Exploring Data-Free Adversarial Robustness Distillation0
Towards Adversarially Robust Text Classifiers by Learning to Reweight Clean Examples0
Model Unlearning via Sparse Autoencoder Subspace Guided Projections0
MOREL: Enhancing Adversarial Robustness through Multi-Objective Representation Learning0
Multi-Agent Reinforcement Learning in Cybersecurity: From Fundamentals to Applications0
Multimodal Large Language Models for Enhanced Traffic Safety: A Comprehensive Review and Future Trends0
ALMA: Aggregated Lipschitz Maximization Attack on Auto-encoders0
Towards Adversarially Robust Vision-Language Models: Insights from Design Choices and Prompt Formatting Techniques0
Multi-Scale Architectures Matter: On the Adversarial Robustness of Flow-based Lossless Compression0
Algorithmic Bias and Data Bias: Understanding the Relation between Distributionally Robust Optimization and Data Curation0
Multi-stage Optimization based Adversarial Training0
aiXamine: Simplified LLM Safety and Security0
Multi-view Representation Learning from Malware to Defend Against Adversarial Variants0
NAP-Tuning: Neural Augmented Prompt Tuning for Adversarially Robust Vision-Language Models0
Narrowing Class-Wise Robustness Gaps in Adversarial Training0
Nash Equilibria and Pitfalls of Adversarial Training in Adversarial Robustness Games0
Nearly Solved? Robust Deepfake Detection Requires More than Visual Forensics0
AI-Compass: A Comprehensive and Effective Multi-module Testing Tool for AI Systems0
Neural Architecture Dilation for Adversarial Robustness0
Show:102550
← PrevPage 43 of 70Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeBERTa (single model)Accuracy0.61Unverified
2ALBERT (single model)Accuracy0.59Unverified
3T5 (single model)Accuracy0.57Unverified
4SMART_RoBERTa (single model)Accuracy0.54Unverified
5FreeLB (single model)Accuracy0.5Unverified
6RoBERTa (single model)Accuracy0.5Unverified
7InfoBERT (single model)Accuracy0.46Unverified
8ELECTRA (single model)Accuracy0.42Unverified
9BERT (single model)Accuracy0.34Unverified
10SMART_BERT (single model)Accuracy0.3Unverified
#ModelMetricClaimedVerifiedStatus
1Mixed classifierAccuracy95.23Unverified
2Stochastic-LWTA/PGD/WideResNet-34-10Accuracy92.26Unverified
3Stochastic-LWTA/PGD/WideResNet-34-5Accuracy91.88Unverified
4GLOT-DRAccuracy84.13Unverified
5TRADES-ANCRA/ResNet18Accuracy81.7Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet-50 (SGD, Cosine)Accuracy77.4Unverified
2ResNet-50 (SGD, Step)Accuracy76.9Unverified
3DeiT-S (AdamW, Cosine)Accuracy76.8Unverified
4ResNet-50 (AdamW, Cosine)Accuracy76.4Unverified
#ModelMetricClaimedVerifiedStatus
1DeiT-S (AdamW, Cosine)Accuracy12.2Unverified
2ResNet-50 (SGD, Cosine)Accuracy3.3Unverified
3ResNet-50 (SGD, Step)Accuracy3.2Unverified
4ResNet-50 (AdamW, Cosine)Accuracy3.1Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet-50 (AdamW, Cosine)mean Corruption Error (mCE)59.3Unverified
2ResNet-50 (SGD, Step)mean Corruption Error (mCE)57.9Unverified
3ResNet-50 (SGD, Cosine)mean Corruption Error (mCE)56.9Unverified
4DeiT-S (AdamW, Cosine)mean Corruption Error (mCE)48Unverified
#ModelMetricClaimedVerifiedStatus
1DeiT-S (AdamW, Cosine)Accuracy13Unverified
2ResNet-50 (SGD, Cosine)Accuracy8.4Unverified
3ResNet-50 (SGD, Step)Accuracy8.3Unverified
4ResNet-50 (AdamW, Cosine)Accuracy8.1Unverified
#ModelMetricClaimedVerifiedStatus
1Mixed ClassifierClean Accuracy85.21Unverified
2ResNet18/MART-ANCRAClean Accuracy60.1Unverified