SOTAVerified

Adversarial Robustness

Adversarial Robustness evaluates the vulnerabilities of machine learning models under various types of adversarial attacks.

Papers

Showing 9511000 of 1746 papers

TitleStatusHype
RobustBlack: Challenging Black-Box Adversarial Attacks on State-of-the-Art Defenses0
Robust Certification for Laplace Learning on Geometric Graphs0
Robust Collective Classification against Structural Attacks0
Robust Dataset Distillation by Matching Adversarial Trajectories0
Adversarial Robustness through Bias Variance Decomposition: A New Perspective for Federated Learning0
Robust Decentralized Learning with Local Updates and Gradient Tracking0
Robust Deep Learning Ensemble against Deception0
Robust Distillation via Untargeted and Targeted Intermediate Adversarial Samples0
RobustEdge: Low Power Adversarial Detection for Cloud-Edge Systems0
Robust Ensemble Model Training via Random Layer Sampling Against Adversarial Attack0
Robust Explainability: A Tutorial on Gradient-Based Attribution Methods for Deep Neural Networks0
Robustified Domain Adaptation0
Robust Information Retrieval0
Robust Linear Regression: Phase-Transitions and Precise Tradeoffs for General Norms0
Robust low-rank training via approximate orthonormal constraints0
RobustMQ: Benchmarking Robustness of Quantized Models0
Robust Multi-Agent Reinforcement Learning Driven by Correlated Equilibrium0
Robustness Against Adversarial Attacks via Learning Confined Adversarial Polytopes0
Robustness Certificates for Implicit Neural Networks: A Mixed Monotone Contractive Approach0
Robustness-Congruent Adversarial Training for Secure Machine Learning Model Updates0
Robustness Implies Privacy in Statistical Estimation0
A Systematic Review of Robustness in Deep Learning for Computer Vision: Mind the gap?0
Robustness May Be at Odds with Fairness: An Empirical Study on Class-wise Accuracy0
Robustness of deep learning classification to adversarial input on GPUs: asynchronous parallel accumulation is a source of vulnerability0
Robustness of Explanation Methods for NLP Models0
Testing robustness of predictions of trained classifiers against naturally occurring perturbations0
Robustness Of Saak Transform Against Adversarial Attacks0
Robustness-preserving Lifelong Learning via Dataset Condensation0
Robust Physical-World Attacks on Face Recognition0
Robust Proxy: Improving Adversarial Robustness by Robust Proxy Learning0
Robust Regularization with Adversarial Labelling of Perturbed Samples0
Robust Subgraph Learning by Monitoring Early Training Representations0
Robust Survival Analysis with Adversarial Regularization0
Robust Tickets Can Transfer Better: Drawing More Transferable Subnetworks in Transfer Learning0
Robust Transferable Feature Extractors: Learning to Defend Pre-Trained Networks Against White Box Adversaries0
Robust Transfer Learning with Pretrained Language Models through Adapters0
Robust Unsupervised Domain Adaptation for 3D Point Cloud Segmentation Under Source Adversarial Attacks0
RoMA: Robust Malware Attribution via Byte-level Adversarial Training with Global Perturbations and Adversarial Consistency Regularization0
RoSearch: Search for Robust Student Architectures When Distilling Pre-trained Language Models0
RUSH: Robust Contrastive Learning via Randomized Smoothing0
SafeGenes: Evaluating the Adversarial Robustness of Genomic Foundation Models0
LLM Safeguard is a Double-Edged Sword: Exploiting False Positives for Denial-of-Service Attacks0
Sample Complexity of Adversarially Robust Linear Classification on Separated Data0
Sample Efficient Detection and Classification of Adversarial Attacks via Self-Supervised Embeddings0
Scalable Lipschitz Estimation for CNNs0
Scalable Neural Learning for Verifiable Consistency with Temporal Specifications0
Scalable Quantitative Verification For Deep Neural Networks0
Scalable Whitebox Attacks on Tree-based Models0
Scoring Black-Box Models for Adversarial Robustness0
Second Order Optimization for Adversarial Robustness and Interpretability0
Show:102550
← PrevPage 20 of 35Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeBERTa (single model)Accuracy0.61Unverified
2ALBERT (single model)Accuracy0.59Unverified
3T5 (single model)Accuracy0.57Unverified
4SMART_RoBERTa (single model)Accuracy0.54Unverified
5FreeLB (single model)Accuracy0.5Unverified
6RoBERTa (single model)Accuracy0.5Unverified
7InfoBERT (single model)Accuracy0.46Unverified
8ELECTRA (single model)Accuracy0.42Unverified
9BERT (single model)Accuracy0.34Unverified
10SMART_BERT (single model)Accuracy0.3Unverified
#ModelMetricClaimedVerifiedStatus
1Mixed classifierAccuracy95.23Unverified
2Stochastic-LWTA/PGD/WideResNet-34-10Accuracy92.26Unverified
3Stochastic-LWTA/PGD/WideResNet-34-5Accuracy91.88Unverified
4GLOT-DRAccuracy84.13Unverified
5TRADES-ANCRA/ResNet18Accuracy81.7Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet-50 (SGD, Cosine)Accuracy77.4Unverified
2ResNet-50 (SGD, Step)Accuracy76.9Unverified
3DeiT-S (AdamW, Cosine)Accuracy76.8Unverified
4ResNet-50 (AdamW, Cosine)Accuracy76.4Unverified
#ModelMetricClaimedVerifiedStatus
1DeiT-S (AdamW, Cosine)Accuracy12.2Unverified
2ResNet-50 (SGD, Cosine)Accuracy3.3Unverified
3ResNet-50 (SGD, Step)Accuracy3.2Unverified
4ResNet-50 (AdamW, Cosine)Accuracy3.1Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet-50 (AdamW, Cosine)mean Corruption Error (mCE)59.3Unverified
2ResNet-50 (SGD, Step)mean Corruption Error (mCE)57.9Unverified
3ResNet-50 (SGD, Cosine)mean Corruption Error (mCE)56.9Unverified
4DeiT-S (AdamW, Cosine)mean Corruption Error (mCE)48Unverified
#ModelMetricClaimedVerifiedStatus
1DeiT-S (AdamW, Cosine)Accuracy13Unverified
2ResNet-50 (SGD, Cosine)Accuracy8.4Unverified
3ResNet-50 (SGD, Step)Accuracy8.3Unverified
4ResNet-50 (AdamW, Cosine)Accuracy8.1Unverified
#ModelMetricClaimedVerifiedStatus
1Mixed ClassifierClean Accuracy85.21Unverified
2ResNet18/MART-ANCRAClean Accuracy60.1Unverified