Table 4: Performance of MIRP on CIFAR-10 dataset with
varying decaying factor.
Decaying factor PGD / Time
0.1 50.59 / 47 min
0.6 49.19 / 50.8 min
Here, Non-zero refers that the values are retained
between batches and Zero refers to that the values are
initialized to zero between batches.
4 CONCLUSIONS
Our findings show that MIRP adversarial training,
when used with random initialization, can in fact be
more effective as the more costly PGD adversarial
training. As a result, we are able to learn adversarially
robust classifiers for CIFAR10/100 in minutes. We
believe that leveraging these significant reductions in
time to train robust models will allow future work to
iterate even faster, and accelerate research in learning
models which are resistant to adversarial attacks.
REFERENCES
Athalye, A., Engstrom, L., Ilyas, A., and Kwok, K. (2018).
Synthesizing robust adversarial examples.
Coleman, C. A., Narayanan, D., Kang, D., Zhao, T., Zhang,
J., Nardi, L., Bailis, P., Olukotun, K., R
´
e, C., and Za-
haria, M. A. (2017). Dawnbench : An end-to-end deep
learning benchmark and competition.
Dong, Y., Liao, F., Pang, T., Su, H., Zhu, J., Hu, X., and Li,
J. (2018). Boosting adversarial attacks with momen-
tum. 2018 IEEE/CVF Conference on Computer Vision
and Pattern Recognition, pages 9185–9193.
Dong, Y., Su, H., Zhu, J., and Bao, F. (2017). Towards
interpretable deep neural networks by leveraging ad-
versarial examples. ArXiv, abs/1708.05493.
Eykholt, K., Evtimov, I., Fernandes, E., Li, B., Rahmati, A.,
Xiao, C., Prakash, A., Kohno, T., and Song, D. (2018).
Robust physical-world attacks on deep learning mod-
els.
Finlayson, S. G., Bowers, J., Ito, J., Zittrain, J., Beam, A.,
and Kohane, I. S. (2019). Adversarial attacks on med-
ical machine learning. Science, 363:1287 – 1289.
Goodfellow, I. J., Shlens, J., and Szegedy, C. (2015). Ex-
plaining and harnessing adversarial examples. CoRR,
abs/1412.6572.
He, K., Zhang, X., Ren, S., and Sun, J. (2016). Deep resid-
ual learning for image recognition. 2016 IEEE Con-
ference on Computer Vision and Pattern Recognition
(CVPR), pages 770–778.
Kurakin, A., Goodfellow, I. J., and Bengio, S. (2017).
Adversarial machine learning at scale. ArXiv,
abs/1611.01236.
Li, Y. and Gal, Y. (2017). Dropout inference in bayesian
neural networks with alpha-divergences. In ICML.
Metzen, J. H., Genewein, T., Fischer, V., and Bischoff, B.
(2017). On detecting adversarial perturbations. ArXiv,
abs/1702.04267.
Micikevicius, P., Narang, S., Alben, J., Diamos, G. F.,
Elsen, E., Garc
´
ıa, D., Ginsburg, B., Houston, M.,
Kuchaiev, O., Venkatesh, G., and Wu, H. (2018).
Mixed precision training. ArXiv, abs/1710.03740.
Pang, T., Du, C., and Zhu, J. (2017). Robust deep learn-
ing via reverse cross-entropy training and threshold-
ing test. ArXiv, abs/1706.00633.
Papernot, N., Mcdaniel, P., Wu, X., Jha, S., and Swami, A.
(2016). Distillation as a defense to adversarial per-
turbations against deep neural networks. 2016 IEEE
Symposium on Security and Privacy (SP), pages 582–
597.
Smith, L. N. and Topin, N. (2019). Super-convergence: very
fast training of neural networks using large learning
rates. In Defense + Commercial Sensing.
Szegedy, C., Zaremba, W., Sutskever, I., Bruna, J., Erhan,
D., Goodfellow, I. J., and Fergus, R. (2014). Intriguing
properties of neural networks. CoRR, abs/1312.6199.
Tram
`
er, F., Kurakin, A., Papernot, N., Boneh, D., and Mc-
daniel, P. (2018). Ensemble adversarial training: At-
tacks and defenses. ArXiv, abs/1705.07204.
Wang, J. (2019). Bilateral adversarial training: Towards
fast training of more robust models against adversarial
attacks. 2019 IEEE/CVF International Conference on
Computer Vision (ICCV), pages 6628–6637.
Wong, E., Rice, L., and Kolter, J. Z. (2020). Fast is better
than free: Revisiting adversarial training.
ICAART 2022 - 14th International Conference on Agents and Artificial Intelligence
916