
(a)
(b)
Figure 6: Validation accuracy and the training loss corre-
spond to three distinct approaches to generating attention
maps on CUB-200-2011 using ResNet18. Our OcCaMix
method of generating attention maps in red curves performs
the best, exhibiting a larger convergence speed without re-
quiring an additional network.
requiring no pre-trained models or additional training
modules. We propose a method utilizing regional op-
erations of arbitrary shapes in deep learning and ex-
pect that more work will be proposed to get rid of
the limitations of square-shaped region operations in
deep learning networks. Comprehensive experimen-
tal results have demonstrated top performance on var-
ious benchmarks and models. Moving forward, our
study will expand to include weakly supervised object
localization, unsupervised learning, self-supervised
learning and masked models.
REFERENCES
Achanta, R., Shaji, A., Smith, K., Lucchi, A., Fua, P., and
S
¨
usstrunk, S. (2012). Slic superpixels compared to
state-of-the-art superpixel methods. IEEE transac-
tions on pattern analysis and machine intelligence,
34(11):2274–2282.
Baek, K., Bang, D., and Shim, H. (2021). Gridmix: Strong
regularization through local context mapping. Pattern
Recognition, 109:107594.
DeVries, T. and Taylor, G. W. (2017). Improved regular-
ization of convolutional neural networks with cutout.
arXiv preprint arXiv:1708.04552.
Feng, Z., Fu, K., and Zhao, Q. (2019). Learning to fo-
cus and discriminate for fine-grained classification. In
2019 IEEE International Conference on Image Pro-
cessing (ICIP), pages 415–419. IEEE.
Hammoudi, K., Cabani, A., Slika, B., Benhabiles, H., Dor-
naika, F., and Melkemi, M. (2023). Superpixelgrid-
masks data augmentation: Application to precision
health and other real-world data. Journal of Health-
care Informatics Research.
He, K., Zhang, X., Ren, S., and Sun, J. (2016). Deep resid-
ual learning for image recognition. In Proceedings of
the IEEE conference on computer vision and pattern
recognition, pages 770–778.
Park, J., Yang, J. Y., Shin, J., Hwang, S. J., and Yang,
E. (2022). Saliency grafting: Innocuous attribution-
guided mixup with calibrated label mixing. In Pro-
ceedings of the AAAI Conference on Artificial Intelli-
gence, volume 36, pages 7957–7965.
Qin, J., Fang, J., Zhang, Q., Liu, W., Wang, X., and Wang,
X. (2020). Resizemix: Mixing data with preserved
object information and true labels. arXiv preprint
arXiv:2012.11101.
Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I.,
and Salakhutdinov, R. (2014). Dropout: a simple way
to prevent neural networks from overfitting. The jour-
nal of machine learning research, 15(1):1929–1958.
Uddin, A. S., Monira, M. S., Shin, W., Chung, T., and Bae,
S.-H. (2020). Saliencymix: A saliency guided data
augmentation strategy for better regularization. In In-
ternational Conference on Learning Representations.
Walawalkar, D., Shen, Z., Liu, Z., and Savvides, M. (2020).
Attentive cutmix: An enhanced data augmentation
approach for deep learning based image classifica-
tion. In ICASSP 2020-2020 IEEE International Con-
ference on Acoustics, Speech and Signal Processing
(ICASSP), pages 3642–3646. IEEE.
Yu, X., Zhao, Y., Gao, Y., and Xiong, S. (2021). Maskcov:
A random mask covariance network for ultra-fine-
grained visual categorization. Pattern Recognition,
119:108067.
Yun, S., Han, D., Oh, S. J., Chun, S., Choe, J., and Yoo,
Y. (2019). Cutmix: Regularization strategy to train
strong classifiers with localizable features. In Pro-
ceedings of the IEEE/CVF international conference
on computer vision, pages 6023–6032.
Zhang, H., Cisse, M., Dauphin, Y. N., and Lopez-Paz, D.
(2018). mixup: Beyond empirical risk minimization.
In International Conference on Learning Representa-
tions.
Zhou, B., Khosla, A., Lapedriza, A., Oliva, A., and Tor-
ralba, A. (2016). Learning deep features for discrim-
inative localization. In Proceedings of the IEEE con-
ference on computer vision and pattern recognition,
pages 2921–2929.
Image Augmentation Preserving Object Parts Using Superpixels of Variable Granularity
717