model classification is unstable when classes were in-
creased. We should be able to improve the perfor-
mance of our method on CIFAR-100 by developing
a better method for creating the ground truth label or
optimizing the training process. In future work, we
will theoretically investigate our method and improve
the training process.
ACKNOWLEDGEMENTS
This work was supported by JST SPRING, Grant
Number JPMJSP2158.
REFERENCES
Brock, A., Donahue, J., and Simonyan, K. (2019). Large
scale gan training for high fidelity natural image syn-
thesis. In the International Conference on Learning
Representations.
Carlini, N. and Wagner, D. (2017). Towards evaluating the
robustness of neural networks. In the IEEE Sympo-
sium on Security and Privacy, volume 1, pages 39–57.
Goodfellow, I., Shlens, J., and Szegedy, C. (2015). Explain-
ing and harnessing adversarial examples. In the Inter-
national Conference on Learning Representations.
Guo, H., Mao, Y., and Zhang, R. (2019). Mixup as locally
linear out-of-manifold regularization. In the AAAI
Conference on Artificial Intelligence, volume 33,
pages 3714–3722.
He, K., Zhang, X., Ren, S., and Sun, J. (2016). Deep resid-
ual learning for image recognition. In the IEEE Con-
ference on Computer Vision and Pattern Recognition,
pages 770–778.
Lee, S., Lee, H., and Yoon, S. (2020). Adversarial vertex
mixup-toward better adversarially robust generaliza-
tion. In the IEEE/CVF Conference on Computer Vi-
sion and Pattern Recognition, pages 272–281.
Long, J., Shelhamer, E., and Darrell, T. (2015). Fully con-
volutional networks for semantic segmentation. In
the IEEE Conference on Computer Vision and Pattern
Recognition, pages 3431–3440.
Madry, A., Makelov, A., Schmidt, L., Tsipras, D., and
Vladu, A. (2018). Towards deep learning models re-
sistant to adversarial attacks. In the International Con-
ference on Learning Representations.
Meng, D. and Chen, H. (2017). Magnet: a two-pronged de-
fense against adversarial examples. In the ACM Con-
ference on Computer and Communications Security,
pages 135—147.
Moosavi-Dezfooli, S.-M., Fawzi, A., Fawzi, O., and
Frossard, P. (2017). Universal adversarial perturba-
tions. In the IEEE Conference on Computer Vision
and Pattern Recognition, pages 1765–1773.
Moosavi-Dezfooli, S.-M., Fawzi, A., and Frossard, P.
(2016). Deepfool: a simple and accurate method
to fool deep neural networks. In 2574-2582, editor,
the IEEE Conference on Computer Vision and Pattern
Recognition.
Papernot, N., McDaniel, P., and Goodfellow, I. (2016a).
Transferability in machine learning: from phenomena
to black-box attacks using adversarial samples. arXiv
preprint arXiv:1605.07277.
Papernot, N., McDaniel, P., Wu, X., Jha, S., and Swami, A.
(2016b). Distillation as a defense to adversarial per-
turbations against deep neural networks. In the IEEE
Symposium on Security and Privacy.
Qin, J., Fang, J., Zhang, Q., Liu, W., Wang, X., and Wang,
X. (2020). Resizemix: Mixing data with preserved
object information and true labels. arXiv preprint
arXiv:2012.11101.
Redmon, J., Divvala, S., Girshick, R., and Farhadi, A.
(2016). You only look once: Unified, real-time ob-
ject detection. In the IEEE Conference on Computer
Vision and Pattern Recognition, pages 779–788.
Samangouei, P., Kabkab, M., and Chellappa, R. (2018).
Defense-gan: Protecting classifiers against adversarial
attacks using generative models. In the International
Conference on Learning Representations.
Schmidt, L., Santurkar, S., Tsipras, D., Talwar, K., and
Madry, A. (2018). Adversarially robust generaliza-
tion requires more data. In the Advances in Neural
Information Processing Systems.
Shafahi, A., Najibi, M., Ghiasi, A., Xu, Z., Dickerson, J.,
Studer, C., Davis, L. S., Taylor, G., and Goldstein, T.
(2019). Adversarial training for free! In the Advances
in Neural Information Processing Systems.
Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., and Wo-
jna, Z. (2016). Rethinking the inception architecture
for computer vision. In the IEEE Conference on Com-
puter Vision and Pattern Recognition, pages 2818–
2826.
Szegedy, C., Zaremba, W., Sutskever, I., Bruna, J., Erhan,
D., Goodfellow, I., and Fergus, R. (2014). Intrigu-
ing properties of neural networks. In the International
Conference on Learning Representations.
Tram
`
er, F., Kurakin, A., Papernot, N., Goodfellow, I.,
Boneh, D., and McDaniel, P. (2018). Ensemble ad-
versarial training: Attacks and defenses. In the Inter-
national Conference on Learning Representations.
Tsipras, D., Santurkar, S., Engstrom, L., Turner, A., and
Madry, A. (2019). Robustness may be at odds with
accuracy. In the International Conference on Learning
Representations.
Wang, J. and Zhang, H. (2019). Bilateral adversarial train-
ing: Towards fast training of more robust models
against adversarial attacks. In the IEEE/CVF Interna-
tional Conference on Computer Vision, pages 6629–
6638.
Wong, E., Rice, L., and Kolter, J. Z. (2020). Fast is better
than free: Revisiting adversarial training. In the Inter-
national Conference on Learning Representations.
Xie, C., Wang, J., Zhang, Z., Zhou, Y., Xie, L., and Yuille,
A. (2017). Adversarial examples for semantic seg-
mentation and object detection. In the IEEE Interna-
Masking and Mixing Adversarial Training
81