SOTAVerified

Adversarial Robustness

Adversarial Robustness evaluates the vulnerabilities of machine learning models under various types of adversarial attacks.

Papers

Showing 15761600 of 1746 papers

TitleStatusHype
Benchmarking Adversarial Robustness0
P-CapsNets: a General Form of Convolutional Neural Networks0
APRICOT: A Dataset of Physical Adversarial Attacks on Object DetectionCode0
What it Thinks is Important is Important: Robustness Transfers through Input GradientsCode0
Feature Losses for Adversarial Robustness0
Exploring the Back Alleys: Analysing The Robustness of Alternative Neural Network Architectures against Adversarial Attacks0
An Empirical Study on the Relation between Network Interpretability and Adversarial RobustnessCode0
Towards Robust Image Classification Using Sequential Attention Models0
Error Correcting Output Codes Improve Probability Estimation and Adversarial Robustness of Deep Neural NetworksCode0
Can Attention Masks Improve Adversarial Robustness?0
An Adaptive View of Adversarial Robustness from Test-time Smoothing DefenseCode0
Playing it Safe: Adversarial Robustness with an Abstain Option0
CAMUS: A Framework to Build Formal Specifications for Deep Perception Systems Using Simulators0
Verifiability and Predictability: Interpreting Utilities of Network Architectures for Point Cloud ProcessingCode0
Adversarial Robustness of Flow-Based Generative Models0
AdvKnn: Adversarial Attacks On K-Nearest Neighbor Classifiers With Approximate GradientsCode0
Finding a human-like classifier0
Preventing Gradient Attenuation in Lipschitz Constrained Convolutional NetworksCode0
MadNet: Using a MAD Optimization for Defending Against Adversarial Attacks0
Fault Tolerance of Neural Networks in Adversarial Settings0
Distribution Density, Tails, and Outliers in Machine Learning: Metrics and Applications0
Certified Adversarial Robustness for Deep Reinforcement Learning0
Label Smoothing and Logit Squeezing: A Replacement for Adversarial Training?0
A Useful Taxonomy for Adversarial Robustness of Neural Networks0
Are Perceptually-Aligned Gradients a General Property of Robust Classifiers?0
Show:102550
← PrevPage 64 of 70Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1DeBERTa (single model)Accuracy0.61Unverified
2ALBERT (single model)Accuracy0.59Unverified
3T5 (single model)Accuracy0.57Unverified
4SMART_RoBERTa (single model)Accuracy0.54Unverified
5FreeLB (single model)Accuracy0.5Unverified
6RoBERTa (single model)Accuracy0.5Unverified
7InfoBERT (single model)Accuracy0.46Unverified
8ELECTRA (single model)Accuracy0.42Unverified
9BERT (single model)Accuracy0.34Unverified
10SMART_BERT (single model)Accuracy0.3Unverified
#ModelMetricClaimedVerifiedStatus
1Mixed classifierAccuracy95.23Unverified
2Stochastic-LWTA/PGD/WideResNet-34-10Accuracy92.26Unverified
3Stochastic-LWTA/PGD/WideResNet-34-5Accuracy91.88Unverified
4GLOT-DRAccuracy84.13Unverified
5TRADES-ANCRA/ResNet18Accuracy81.7Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet-50 (SGD, Cosine)Accuracy77.4Unverified
2ResNet-50 (SGD, Step)Accuracy76.9Unverified
3DeiT-S (AdamW, Cosine)Accuracy76.8Unverified
4ResNet-50 (AdamW, Cosine)Accuracy76.4Unverified
#ModelMetricClaimedVerifiedStatus
1DeiT-S (AdamW, Cosine)Accuracy12.2Unverified
2ResNet-50 (SGD, Cosine)Accuracy3.3Unverified
3ResNet-50 (SGD, Step)Accuracy3.2Unverified
4ResNet-50 (AdamW, Cosine)Accuracy3.1Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet-50 (AdamW, Cosine)mean Corruption Error (mCE)59.3Unverified
2ResNet-50 (SGD, Step)mean Corruption Error (mCE)57.9Unverified
3ResNet-50 (SGD, Cosine)mean Corruption Error (mCE)56.9Unverified
4DeiT-S (AdamW, Cosine)mean Corruption Error (mCE)48Unverified
#ModelMetricClaimedVerifiedStatus
1DeiT-S (AdamW, Cosine)Accuracy13Unverified
2ResNet-50 (SGD, Cosine)Accuracy8.4Unverified
3ResNet-50 (SGD, Step)Accuracy8.3Unverified
4ResNet-50 (AdamW, Cosine)Accuracy8.1Unverified
#ModelMetricClaimedVerifiedStatus
1Mixed ClassifierClean Accuracy85.21Unverified
2ResNet18/MART-ANCRAClean Accuracy60.1Unverified