1 |
Adversarial Robustness Limits via Scaling-Law and Human-Alignment Studies
|
93.68% |
73.71% |
73.71% |
× |
× |
WideResNet-94-16 |
ICML 2024 |
2 |
MeanSparse: Post-Training Robustness Enhancement Through Mean-Centered Feature Sparsification
|
93.60% |
75.28% |
73.10% |
× |
☑ |
MeanSparse WideResNet-94-16 |
arXiv, Jun 2024 |
3 |
Adversarial Robustness Limits via Scaling-Law and Human-Alignment Studies
|
93.11% |
71.59% |
71.59% |
× |
× |
WideResNet-82-8 |
ICML 2024 |
4 |
Robust Principles: Architectural Design Principles for Adversarially Robust CNNs
|
93.27% |
71.07% |
71.07% |
× |
× |
RaWideResNet-70-16 |
BMVC 2023 |
5 |
Better Diffusion Models Further Improve Adversarial Training
|
93.25% |
70.69% |
70.69% |
× |
× |
WideResNet-70-16 |
ICML 2023 |
6 |
MixedNUTS: Training-Free Accuracy-Robustness Balance via Nonlinearly Mixed Classifiers
|
95.19% |
70.08% |
69.71% |
× |
☑ |
ResNet-152 + WideResNet-70-16 |
TMLR, Aug 2024 |
7 |
MeanSparse: Post-Training Robustness Enhancement Through Mean-Centered Feature Sparsification
|
93.24% |
72.08% |
68.94% |
× |
☑ |
MeanSparse RaWideResNet-70-16 |
arXiv, Jun 2024 |
8 |
Improving the Accuracy-Robustness Trade-off of Classifiers via Adaptive Smoothing
|
95.23% |
68.06% |
68.06% |
× |
☑ |
ResNet-152 + WideResNet-70-16 + mixing network |
SIMODS 2024 |
9 |
Decoupled Kullback-Leibler Divergence Loss
|
92.16% |
67.73% |
67.73% |
× |
× |
WideResNet-28-10 |
NeurIPS 2024 |
10 |
Better Diffusion Models Further Improve Adversarial Training
|
92.44% |
67.31% |
67.31% |
× |
× |
WideResNet-28-10 |
ICML 2023 |
11 |
Fixing Data Augmentation to Improve Adversarial Robustness
|
92.23% |
66.58% |
66.56% |
× |
☑ |
WideResNet-70-16 |
arXiv, Mar 2021 |
12 |
Improving Robustness using Generated Data
|
88.74% |
66.11% |
66.10% |
× |
× |
WideResNet-70-16 |
NeurIPS 2021 |
13 |
Uncovering the Limits of Adversarial Training against Norm-Bounded Adversarial Examples
|
91.10% |
65.88% |
65.87% |
× |
☑ |
WideResNet-70-16 |
arXiv, Oct 2020 |
14 |
Revisiting Residual Networks for Adversarial Robustness: An Architectural Perspective
|
91.58% |
65.79% |
65.79% |
× |
☑ |
WideResNet-A4 |
arXiv, Dec. 2022 |
15 |
Fixing Data Augmentation to Improve Adversarial Robustness
|
88.50% |
64.64% |
64.58% |
× |
× |
WideResNet-106-16 |
arXiv, Mar 2021 |
16 |
Stable Neural ODE with Lyapunov-Stable Equilibrium Points for Defending Against Adversarial Attacks
|
93.73% |
71.28% |
64.20% |
☑ |
☑ |
WideResNet-70-16, Neural ODE block |
NeurIPS 2021 |
17 |
Fixing Data Augmentation to Improve Adversarial Robustness
|
88.54% |
64.25% |
64.20% |
× |
× |
WideResNet-70-16 |
arXiv, Mar 2021 |
18 |
Exploring and Exploiting Decision Boundary Dynamics for Adversarial Robustness
|
93.69% |
63.89% |
63.89% |
× |
× |
WideResNet-28-10 |
ICLR 2023 |
19 |
Improving Robustness using Generated Data
|
87.50% |
63.44% |
63.38% |
× |
× |
WideResNet-28-10 |
NeurIPS 2021 |
20 |
Robustness and Accuracy Could Be Reconcilable by (Proper) Definition
|
89.01% |
63.35% |
63.35% |
× |
× |
WideResNet-70-16 |
ICML 2022 |
21 |
Helper-based Adversarial Training: Reducing Excessive Margin to Achieve a Better Accuracy vs. Robustness Trade-off
|
91.47% |
62.83% |
62.83% |
× |
☑ |
WideResNet-34-10 |
OpenReview, Jun 2021 |
22 |
Robust Learning Meets Generative Models: Can Proxy Distributions Improve Adversarial Robustness?
|
87.30% |
62.79% |
62.79% |
× |
× |
ResNest152 |
ICLR 2022 |
23 |
Uncovering the Limits of Adversarial Training against Norm-Bounded Adversarial Examples
|
89.48% |
62.80% |
62.76% |
× |
☑ |
WideResNet-28-10 |
arXiv, Oct 2020 |
24 |
Exploring Architectural Ingredients of Adversarially Robust Deep Neural Networks
|
91.23% |
62.54% |
62.54% |
× |
☑ |
WideResNet-34-R |
NeurIPS 2021 |
25 |
Exploring Architectural Ingredients of Adversarially Robust Deep Neural Networks
|
90.56% |
61.56% |
61.56% |
× |
☑ |
WideResNet-34-R |
NeurIPS 2021 |
26 |
Parameterizing Activation Functions for Adversarial Robustness
|
87.02% |
61.55% |
61.55% |
× |
× |
WideResNet-28-10-PSSiLU |
arXiv, Oct 2021 |
27 |
Robustness and Accuracy Could Be Reconcilable by (Proper) Definition
|
88.61% |
61.04% |
61.04% |
× |
× |
WideResNet-28-10 |
ICML 2022 |
28 |
Helper-based Adversarial Training: Reducing Excessive Margin to Achieve a Better Accuracy vs. Robustness Trade-off
|
88.16% |
60.97% |
60.97% |
× |
× |
WideResNet-28-10 |
OpenReview, Jun 2021 |
29 |
Fixing Data Augmentation to Improve Adversarial Robustness
|
87.33% |
60.75% |
60.73% |
× |
× |
WideResNet-28-10 |
arXiv, Mar 2021 |
30 |
Do Wider Neural Networks Really Help Adversarial Robustness?
|
87.67% |
60.65% |
60.65% |
Unknown |
☑ |
WideResNet-34-15 |
arXiv, Oct 2020 |
31 |
Improving Neural Network Robustness via Persistency of Excitation
|
86.53% |
60.41% |
60.41% |
× |
☑ |
WideResNet-34-15 |
ACC 2022 |
32 |
Robust Learning Meets Generative Models: Can Proxy Distributions Improve Adversarial Robustness?
|
86.68% |
60.27% |
60.27% |
× |
× |
WideResNet-34-10 |
ICLR 2022 |
33 |
Adversarial Weight Perturbation Helps Robust Generalization
|
88.25% |
60.04% |
60.04% |
× |
☑ |
WideResNet-28-10 |
NeurIPS 2020 |
34 |
Improving Neural Network Robustness via Persistency of Excitation
|
89.46% |
59.66% |
59.66% |
× |
☑ |
WideResNet-28-10 |
ACC 2022 |
35 |
Geometry-aware Instance-reweighted Adversarial Training
|
89.36% |
59.64% |
59.64% |
× |
☑ |
WideResNet-28-10 |
ICLR 2021 |
36 |
Unlabeled Data Improves Adversarial Robustness
|
89.69% |
59.53% |
59.53% |
× |
☑ |
WideResNet-28-10 |
NeurIPS 2019 |
37 |
Improving Robustness using Generated Data
|
87.35% |
58.63% |
58.50% |
× |
× |
PreActResNet-18 |
NeurIPS 2021 |
38 |
Data filtering for efficient adversarial training
|
86.10% |
58.09% |
58.09% |
× |
× |
WideResNet-34-20 |
Pattern Recognition 2024 |
39 |
Scaling Adversarial Training to Large Perturbation Bounds
|
85.32% |
58.04% |
58.04% |
× |
× |
WideResNet-34-10 |
ECCV 2022 |
40 |
Efficient and Effective Augmentation Strategy for Adversarial Training
|
88.71% |
57.81% |
57.81% |
× |
× |
WideResNet-34-10 |
NeurIPS 2022 |
41 |
LTD: Low Temperature Distillation for Robust Adversarial Training
|
86.03% |
57.71% |
57.71% |
× |
× |
WideResNet-34-20 |
arXiv, Nov 2021 |
42 |
Helper-based Adversarial Training: Reducing Excessive Margin to Achieve a Better Accuracy vs. Robustness Trade-off
|
89.02% |
57.67% |
57.67% |
× |
☑ |
PreActResNet-18 |
OpenReview, Jun 2021 |
43 |
LAS-AT: Adversarial Training with Learnable Attack Strategy
|
85.66% |
57.61% |
57.61% |
× |
× |
WideResNet-70-16 |
arXiv, Mar 2022 |
44 |
A Light Recipe to Train Robust Vision Transformers
|
91.73% |
57.58% |
57.58% |
× |
☑ |
XCiT-L12 |
arXiv, Sep 2022 |
45 |
Data filtering for efficient adversarial training
|
86.54% |
57.30% |
57.30% |
× |
× |
WideResNet-34-10 |
Pattern Recognition 2024 |
46 |
A Light Recipe to Train Robust Vision Transformers
|
91.30% |
57.27% |
57.27% |
× |
☑ |
XCiT-M12 |
arXiv, Sep 2022 |
47 |
Uncovering the Limits of Adversarial Training against Norm-Bounded Adversarial Examples
|
85.29% |
57.20% |
57.14% |
× |
× |
WideResNet-70-16 |
arXiv, Oct 2020 |
48 |
HYDRA: Pruning Adversarially Robust Neural Networks
|
88.98% |
57.14% |
57.14% |
× |
☑ |
WideResNet-28-10 |
NeurIPS 2020 |
49 |
Decoupled Kullback-Leibler Divergence Loss
|
85.31% |
57.09% |
57.09% |
× |
× |
WideResNet-34-10 |
NeurIPS 2024 |
50 |
Helper-based Adversarial Training: Reducing Excessive Margin to Achieve a Better Accuracy vs. Robustness Trade-off
|
86.86% |
57.09% |
57.09% |
× |
× |
PreActResNet-18 |
OpenReview, Jun 2021 |
51 |
LTD: Low Temperature Distillation for Robust Adversarial Training
|
85.21% |
56.94% |
56.94% |
× |
× |
WideResNet-34-10 |
arXiv, Nov 2021 |
52 |
Uncovering the Limits of Adversarial Training against Norm-Bounded Adversarial Examples
|
85.64% |
56.86% |
56.82% |
× |
× |
WideResNet-34-20 |
arXiv, Oct 2020 |
53 |
Fixing Data Augmentation to Improve Adversarial Robustness
|
83.53% |
56.66% |
56.66% |
× |
× |
PreActResNet-18 |
arXiv, Mar 2021 |
54 |
Improving Adversarial Robustness Requires Revisiting Misclassified Examples
|
87.50% |
56.29% |
56.29% |
× |
☑ |
WideResNet-28-10 |
ICLR 2020 |
55 |
LAS-AT: Adversarial Training with Learnable Attack Strategy
|
84.98% |
56.26% |
56.26% |
× |
× |
WideResNet-34-10 |
arXiv, Mar 2022 |
56 |
Adversarial Weight Perturbation Helps Robust Generalization
|
85.36% |
56.17% |
56.17% |
× |
× |
WideResNet-34-10 |
NeurIPS 2020 |
57 |
A Light Recipe to Train Robust Vision Transformers
|
90.06% |
56.14% |
56.14% |
× |
☑ |
XCiT-S12 |
arXiv, Sep 2022 |
58 |
Are Labels Required for Improving Adversarial Robustness?
|
86.46% |
56.03% |
56.03% |
Unknown |
☑ |
WideResNet-28-10 |
NeurIPS 2019 |
59 |
Robust Learning Meets Generative Models: Can Proxy Distributions Improve Adversarial Robustness?
|
84.59% |
55.54% |
55.54% |
× |
× |
ResNet-18 |
ICLR 2022 |
60 |
Using Pre-Training Can Improve Model Robustness and Uncertainty
|
87.11% |
54.92% |
54.92% |
× |
☑ |
WideResNet-28-10 |
ICML 2019 |
61 |
Bag of Tricks for Adversarial Training
|
86.43% |
54.39% |
54.39% |
Unknown |
× |
WideResNet-34-20 |
ICLR 2021 |
62 |
Boosting Adversarial Training with Hypersphere Embedding
|
85.14% |
53.74% |
53.74% |
× |
× |
WideResNet-34-20 |
NeurIPS 2020 |
63 |
Learnable Boundary Guided Adversarial Training
|
88.70% |
53.57% |
53.57% |
× |
× |
WideResNet-34-20 |
ICCV 2021 |
64 |
Attacks Which Do Not Kill Training Make Adversarial Learning Stronger
|
84.52% |
53.51% |
53.51% |
× |
× |
WideResNet-34-10 |
ICML 2020 |
65 |
Overfitting in adversarially robust deep learning
|
85.34% |
53.42% |
53.42% |
× |
× |
WideResNet-34-20 |
ICML 2020 |
66 |
Self-Adaptive Training: beyond Empirical Risk Minimization
|
83.48% |
53.34% |
53.34% |
Unknown |
× |
WideResNet-34-10 |
NeurIPS 2020 |
67 |
Theoretically Principled Trade-off between Robustness and Accuracy
|
84.92% |
53.08% |
53.08% |
Unknown |
× |
WideResNet-34-10 |
ICML 2019 |
68 |
Learnable Boundary Guided Adversarial Training
|
88.22% |
52.86% |
52.86% |
× |
× |
WideResNet-34-10 |
ICCV 2021 |
69 |
Adversarial Robustness through Local Linearization
|
86.28% |
52.84% |
52.84% |
Unknown |
× |
WideResNet-40-8 |
NeurIPS 2019 |
70 |
Efficient and Effective Augmentation Strategy for Adversarial Training
|
85.71% |
52.48% |
52.48% |
× |
× |
ResNet-18 |
NeurIPS 2022 |
71 |
Adversarial Robustness: From Self-Supervised Pre-Training to Fine-Tuning
|
86.04% |
51.56% |
51.56% |
Unknown |
× |
ResNet-50 |
CVPR 2020 |
72 |
Efficient Robust Training via Backward Smoothing
|
85.32% |
51.12% |
51.12% |
Unknown |
× |
WideResNet-34-10 |
arXiv, Oct 2020 |
73 |
Scaling Adversarial Training to Large Perturbation Bounds
|
80.24% |
51.06% |
51.06% |
× |
× |
ResNet-18 |
ECCV 2022 |
74 |
Improving Adversarial Robustness Through Progressive Hardening
|
86.84% |
50.72% |
50.72% |
Unknown |
× |
WideResNet-34-10 |
arXiv, Mar 2020 |
75 |
Robustness library
|
87.03% |
49.25% |
49.25% |
Unknown |
× |
ResNet-50 |
GitHub, Oct 2019 |
76 |
Harnessing the Vulnerability of Latent Layers in Adversarially Trained Models
|
87.80% |
49.12% |
49.12% |
Unknown |
× |
WideResNet-34-10 |
IJCAI 2019 |
77 |
Metric Learning for Adversarial Robustness
|
86.21% |
47.41% |
47.41% |
Unknown |
× |
WideResNet-34-10 |
NeurIPS 2019 |
78 |
You Only Propagate Once: Accelerating Adversarial Training via Maximal Principle
|
87.20% |
44.83% |
44.83% |
Unknown |
× |
WideResNet-34-10 |
NeurIPS 2019 |
79 |
Towards Deep Learning Models Resistant to Adversarial Attacks
|
87.14% |
44.04% |
44.04% |
Unknown |
× |
WideResNet-34-10 |
ICLR 2018 |
80 |
Understanding and Improving Fast Adversarial Training
|
79.84% |
43.93% |
43.93% |
Unknown |
× |
PreActResNet-18 |
NeurIPS 2020 |
81 |
Rethinking Softmax Cross-Entropy Loss for Adversarial Robustness
|
80.89% |
43.48% |
43.48% |
Unknown |
× |
ResNet-32 |
ICLR 2020 |
82 |
Fast is better than free: Revisiting adversarial training
|
83.34% |
43.21% |
43.21% |
Unknown |
× |
PreActResNet-18 |
ICLR 2020 |
83 |
Adversarial Training for Free!
|
86.11% |
41.47% |
41.47% |
Unknown |
× |
WideResNet-34-10 |
NeurIPS 2019 |
84 |
MMA Training: Direct Input Space Margin Maximization through Adversarial Training
|
84.36% |
41.44% |
41.44% |
Unknown |
× |
WideResNet-28-4 |
ICLR 2020 |
85 |
A Tunable Robust Pruning Framework Through Dynamic Network Rewiring of DNNs
|
87.32% |
40.41% |
40.41% |
× |
× |
ResNet-18 |
ASP-DAC 2021 |
86 |
Controlling Neural Level Sets
|
81.30% |
40.22% |
40.22% |
Unknown |
× |
ResNet-18 |
NeurIPS 2019 |
87 |
Robustness via Curvature Regularization, and Vice Versa
|
83.11% |
38.50% |
38.50% |
Unknown |
× |
ResNet-18 |
CVPR 2019 |
88 |
Defense Against Adversarial Attacks Using Feature Scattering-based Adversarial Training
|
89.98% |
36.64% |
36.64% |
Unknown |
× |
WideResNet-28-10 |
NeurIPS 2019 |
89 |
Adversarial Interpolation Training: A Simple Approach for Improving Model Robustness
|
90.25% |
36.45% |
36.45% |
Unknown |
× |
WideResNet-28-10 |
OpenReview, Sep 2019 |
90 |
Adversarial Defense via Learning to Generate Diverse Attacks
|
78.91% |
34.95% |
34.95% |
Unknown |
× |
ResNet-20 |
ICCV 2019 |
91 |
Sensible adversarial learning
|
91.51% |
34.22% |
34.22% |
Unknown |
× |
WideResNet-34-10 |
OpenReview, Sep 2019 |
92 |
Towards Stable and Efficient Training of Verifiably Robust Neural Networks
|
44.73% |
32.64% |
32.64% |
Unknown |
× |
5-layer-CNN |
ICLR 2020 |
93 |
Bilateral Adversarial Training: Towards Fast Training of More Robust Models Against Adversarial Attacks
|
92.80% |
29.35% |
29.35% |
Unknown |
× |
WideResNet-28-10 |
ICCV 2019 |
94 |
Enhancing Adversarial Defense by k-Winners-Take-All
|
79.28% |
18.50% |
7.40% |
☑ |
× |
DenseNet-121 |
ICLR 2020 |
95 |
Manifold Regularization for Adversarial Robustness
|
90.84% |
1.35% |
1.35% |
Unknown |
× |
ResNet-18 |
arXiv, Mar 2020 |
96 |
Adversarial Defense by Restricting the Hidden Space of Deep Neural Networks
|
89.16% |
0.28% |
0.28% |
Unknown |
× |
ResNet-110 |
ICCV 2019 |
97 |
Jacobian Adversarially Regularized Networks for Robustness
|
93.79% |
0.26% |
0.26% |
Unknown |
× |
WideResNet-34-10 |
ICLR 2020 |
98 |
ClusTR: Clustering Training for Robustness
|
91.03% |
0.00% |
0.00% |
Unknown |
× |
WideResNet-28-10 |
arXiv, Jun 2020 |
99 |
Standardly trained model
|
94.78% |
0.0% |
0.0% |
Unknown |
× |
WideResNet-28-10 |
N/A |