SOTAVerified

Adversarial Robustness

Adversarial Robustness evaluates the vulnerabilities of machine learning models under various types of adversarial attacks.

Papers

Showing 10011050 of 1746 papers

TitleStatusHype
Topological Signatures of Adversaries in Multimodal Alignments0
Likelihood Landscapes: A Unifying Principle Behind Many Adversarial Defenses0
Generalized No Free Lunch Theorem for Adversarial Robustness0
Limited but consistent gains in adversarial robustness by co-training object recognition models with human EEG0
Lipschitz Constant Meets Condition Number: Learning Robust and Compact Deep Neural Networks0
Lipschitz regularized Deep Neural Networks generalize and are adversarially robust0
Local Competition and Stochasticity for Adversarial Robustness in Deep Learning0
Local Competition and Uncertainty for Adversarial Robustness in Deep Learning0
Logit Calibration and Feature Contrast for Robust Federated Learning on Non-IID Data0
A3E: Aligned and Augmented Adversarial Ensemble for Accurate, Robust and Privacy-Preserving EEG Decoding0
Long-tailed Adversarial Training with Self-Distillation0
Accelerating Adversarial Perturbation by 50% with Semi-backward Propagation0
An empirical study of pretrained representations for few-shot classification0
Toward Adversarial Robustness by Diversity in an Ensemble of Specialized Deep Neural Networks0
LOTS about Attacking Deep Features0
Adversarial Attacks and Defenses for Speech Recognition Systems0
Lower Difficulty and Better Robustness: A Bregman Divergence Perspective for Adversarial Training0
L_p-norm Distortion-Efficient Adversarial Attack0
Towards Reliable Neural Specifications0
An Empirical Study of Accuracy, Fairness, Explainability, Distributional Robustness, and Adversarial Robustness0
Lyapunov Neural ODE State-Feedback Control Policies0
Maintaining Adversarial Robustness in Continuous Learning0
Manifold-aware Training: Increase Adversarial Robustness with Feature Clustering0
Adapters Mixup: Mixing Parameter-Efficient Adapters to Enhance the Adversarial Robustness of Fine-tuned Pre-trained Text Classifiers0
And/or trade-off in artificial neurons: impact on adversarial robustness0
Masks and Mimicry: Strategic Obfuscation and Impersonation Attacks on Authorship Verification0
Universal Adversarial Training with Class-Wise Perturbations0
Measuring Adversarial Datasets0
Measuring Equality in Machine Learning Security Defenses: A Case Study in Speech Recognition0
Measuring the Effect of Causal Disentanglement on the Adversarial Robustness of Neural Network Models0
Classifier-independent Lower-Bounds for Adversarial Robustness0
Analyzing Adversarial Robustness of Deep Neural Networks in Pixel Space: a Semantic Perspective0
Meta Invariance Defense Towards Generalizable Robustness to Unknown Adversarial Attacks0
Towards Achieving Adversarial Robustness Beyond Perceptual Limits0
Adversarial Amendment is the Only Force Capable of Transforming an Enemy into a Friend0
Analyzing Adversarial Attacks Against Deep Learning for Intrusion Detection in IoT Networks0
Adversarial Alignment for LLMs Requires Simpler, Reproducible, and More Measurable Objectives0
Mind the box: l_1-APGD for sparse adversarial attacks on image classifiers0
Min-Max Optimization under Delays0
MIRST-DM: Multi-Instance RST with Drop-Max Layer for Robust Classification of Breast Cancer0
An Adversarial Robustness Benchmark for Enterprise Network Intrusion Detection0
Mitigating Closed-model Adversarial Examples with Bayesian Neural Modeling for Enhanced End-to-End Speech Recognition0
Towards Adversarially Robust and Domain Generalizable Stereo Matching by Rethinking DNN Feature Backbones0
Mitigating Low-Frequency Bias: Feature Recalibration and Frequency Attention Regularization for Adversarial Robustness0
A Multiclass Boosting Framework for Achieving Fast and Provable Adversarial Robustness0
Towards Adversarially Robust Continual Learning0
Mixing Classifiers to Alleviate the Accuracy-Robustness Trade-Off0
Mixture of Robust Experts (MoRE):A Robust Denoising Method towards multiple perturbations0
Universally Robust Graph Neural Networks by Preserving Neighbor Similarity0
D4: Detection of Adversarial Diffusion Deepfakes Using Disjoint Ensembles0
Show:102550
← PrevPage 21 of 35Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeBERTa (single model)Accuracy0.61Unverified
2ALBERT (single model)Accuracy0.59Unverified
3T5 (single model)Accuracy0.57Unverified
4SMART_RoBERTa (single model)Accuracy0.54Unverified
5FreeLB (single model)Accuracy0.5Unverified
6RoBERTa (single model)Accuracy0.5Unverified
7InfoBERT (single model)Accuracy0.46Unverified
8ELECTRA (single model)Accuracy0.42Unverified
9BERT (single model)Accuracy0.34Unverified
10SMART_BERT (single model)Accuracy0.3Unverified
#ModelMetricClaimedVerifiedStatus
1Mixed classifierAccuracy95.23Unverified
2Stochastic-LWTA/PGD/WideResNet-34-10Accuracy92.26Unverified
3Stochastic-LWTA/PGD/WideResNet-34-5Accuracy91.88Unverified
4GLOT-DRAccuracy84.13Unverified
5TRADES-ANCRA/ResNet18Accuracy81.7Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet-50 (SGD, Cosine)Accuracy77.4Unverified
2ResNet-50 (SGD, Step)Accuracy76.9Unverified
3DeiT-S (AdamW, Cosine)Accuracy76.8Unverified
4ResNet-50 (AdamW, Cosine)Accuracy76.4Unverified
#ModelMetricClaimedVerifiedStatus
1DeiT-S (AdamW, Cosine)Accuracy12.2Unverified
2ResNet-50 (SGD, Cosine)Accuracy3.3Unverified
3ResNet-50 (SGD, Step)Accuracy3.2Unverified
4ResNet-50 (AdamW, Cosine)Accuracy3.1Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet-50 (AdamW, Cosine)mean Corruption Error (mCE)59.3Unverified
2ResNet-50 (SGD, Step)mean Corruption Error (mCE)57.9Unverified
3ResNet-50 (SGD, Cosine)mean Corruption Error (mCE)56.9Unverified
4DeiT-S (AdamW, Cosine)mean Corruption Error (mCE)48Unverified
#ModelMetricClaimedVerifiedStatus
1DeiT-S (AdamW, Cosine)Accuracy13Unverified
2ResNet-50 (SGD, Cosine)Accuracy8.4Unverified
3ResNet-50 (SGD, Step)Accuracy8.3Unverified
4ResNet-50 (AdamW, Cosine)Accuracy8.1Unverified
#ModelMetricClaimedVerifiedStatus
1Mixed ClassifierClean Accuracy85.21Unverified
2ResNet18/MART-ANCRAClean Accuracy60.1Unverified