SOTAVerified

Adversarial Robustness

Adversarial Robustness evaluates the vulnerabilities of machine learning models under various types of adversarial attacks.

Papers

Showing 751800 of 1746 papers

TitleStatusHype
A Robust Adversarial Ensemble with Causal (Feature Interaction) Interpretations for Image Classification0
FullLoRA-AT: Efficiently Boosting the Robustness of Pretrained Vision Transformers0
Fully Dynamic Adversarially Robust Correlation Clustering in Polylogarithmic Update Time0
FuncFooler: A Practical Black-box Attack Against Learning-based Binary Code Similarity Detection Methods0
Functional Network: A Novel Framework for Interpretability of Deep Neural Networks0
Function Composition in Trustworthy Machine Learning: Implementation Choices, Insights, and Questions0
Function-Space Regularization for Deep Bayesian Classification0
Function-Space Variational Inference for Deep Bayesian Classification0
Game-Theoretic Defenses for Robust Conformal Prediction Against Adversarial Attacks in Medical Imaging0
Improved Adversarial Robustness by Reducing Open Space Risk via Tent Activations0
GARNET: A Spectral Approach to Robust and Scalable Graph Neural Networks0
Improved Branch and Bound for Neural Network Verification via Lagrangian Decomposition0
Empirical Study of the Decision Region and Robustness in Deep Neural Networks0
Accelerating Adversarial Perturbation by 50% with Semi-backward Propagation0
Biologically inspired sleep algorithm for increased generalization and adversarial robustness in deep neural networks0
General Coded Computing: Adversarial Settings0
Emoti-Attack: Zero-Perturbation Adversarial Attacks on NLP Systems via Emoji Sequences0
Generalizable Deepfake Detection with Phase-Based Motion Analysis0
_1 Adversarial Robustness Certificates: a Randomized Smoothing Approach0
Generalization Error Analysis of Neural networks with Gradient Based Regularization0
Generalization of Neural Combinatorial Solvers Through the Lens of Adversarial Robustness0
Incorporating Hidden Layer representation into Adversarial Attacks and Defences0
Generalized but not Robust? Comparing the Effects of Data Modification Methods on Out-of-Domain Generalization and Adversarial Robustness0
Are Time-Series Foundation Models Deployment-Ready? A Systematic Study of Adversarial Robustness Across Domains0
Generalizing and Improving Jacobian and Hessian Regularization0
Generate and Verify: Semantically Meaningful Formal Analysis of Neural Network Perception Systems0
Improve Adversarial Robustness via Weight Penalization on Classification Layer0
SOAR: Second-Order Adversarial Regularization0
MadNet: Using a MAD Optimization for Defending Against Adversarial Attacks0
Generating Structured Adversarial Attacks Using Frank-Wolfe Method0
Boosting Barely Robust Learners: A New Perspective on Adversarial Robustness0
GenFighter: A Generative and Evolutive Textual Attack Removal0
Eight challenges in developing theory of intelligence0
GenMix: Effective Data Augmentation with Generative Diffusion Model Image Editing0
Training Graph Neural Networks Using Non-Robust Samples0
GHN-Q: Parameter Prediction for Unseen Quantized Convolutional Architectures via Graph Hypernetworks0
Are Perceptually-Aligned Gradients a General Property of Robust Classifiers?0
Global Adversarial Robustness Guarantees for Neural Networks0
Bridged Adversarial Training0
GNN-Ensemble: Towards Random Decision Graph Neural Networks0
Imperceptible Adversarial Attacks on Point Clouds Guided by Point-to-Surface Field0
GradDiv: Adversarial Robustness of Randomized Neural Networks via Gradient Diversity Regularization0
Are models trained on temporally-continuous data streams more adversarially robust?0
Impact of Low-bitwidth Quantization on the Adversarial Robustness for Embedded Neural Networks0
Adversarial Robustness Unhardening via Backdoor Attacks in Federated Learning0
GridMix: Strong regularization through local context mapping0
Grimm: A Plug-and-Play Perturbation Rectifier for Graph Neural Networks Defending against Poisoning Attacks0
Guess First to Enable Better Compression and Adversarial Robustness0
Guidance Through Surrogate: Towards a Generic Diagnostic Attack0
Adversarial Robustness in Deep Learning: Attacks on Fragile Neurons0
Show:102550
← PrevPage 16 of 35Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeBERTa (single model)Accuracy0.61Unverified
2ALBERT (single model)Accuracy0.59Unverified
3T5 (single model)Accuracy0.57Unverified
4SMART_RoBERTa (single model)Accuracy0.54Unverified
5FreeLB (single model)Accuracy0.5Unverified
6RoBERTa (single model)Accuracy0.5Unverified
7InfoBERT (single model)Accuracy0.46Unverified
8ELECTRA (single model)Accuracy0.42Unverified
9BERT (single model)Accuracy0.34Unverified
10SMART_BERT (single model)Accuracy0.3Unverified
#ModelMetricClaimedVerifiedStatus
1Mixed classifierAccuracy95.23Unverified
2Stochastic-LWTA/PGD/WideResNet-34-10Accuracy92.26Unverified
3Stochastic-LWTA/PGD/WideResNet-34-5Accuracy91.88Unverified
4GLOT-DRAccuracy84.13Unverified
5TRADES-ANCRA/ResNet18Accuracy81.7Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet-50 (SGD, Cosine)Accuracy77.4Unverified
2ResNet-50 (SGD, Step)Accuracy76.9Unverified
3DeiT-S (AdamW, Cosine)Accuracy76.8Unverified
4ResNet-50 (AdamW, Cosine)Accuracy76.4Unverified
#ModelMetricClaimedVerifiedStatus
1DeiT-S (AdamW, Cosine)Accuracy12.2Unverified
2ResNet-50 (SGD, Cosine)Accuracy3.3Unverified
3ResNet-50 (SGD, Step)Accuracy3.2Unverified
4ResNet-50 (AdamW, Cosine)Accuracy3.1Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet-50 (AdamW, Cosine)mean Corruption Error (mCE)59.3Unverified
2ResNet-50 (SGD, Step)mean Corruption Error (mCE)57.9Unverified
3ResNet-50 (SGD, Cosine)mean Corruption Error (mCE)56.9Unverified
4DeiT-S (AdamW, Cosine)mean Corruption Error (mCE)48Unverified
#ModelMetricClaimedVerifiedStatus
1DeiT-S (AdamW, Cosine)Accuracy13Unverified
2ResNet-50 (SGD, Cosine)Accuracy8.4Unverified
3ResNet-50 (SGD, Step)Accuracy8.3Unverified
4ResNet-50 (AdamW, Cosine)Accuracy8.1Unverified
#ModelMetricClaimedVerifiedStatus
1Mixed ClassifierClean Accuracy85.21Unverified
2ResNet18/MART-ANCRAClean Accuracy60.1Unverified