Figure 9: Classification rates among several value ranges of
tensors for each amount [%] of training data of CIFAR10.
Figure 10: Classification rates of several types of labels for
each amount [%] of training data of CIFAR10.
5 CONCLUSIONS
In this paper, we improved and evaluated the tensor
representation label proposed in (Chen, 2021) as a
different label representation in image recognition.
Specifically, improvements and evaluations were
conducted on image selection of a reference,
complexity increase, and tensor size setting. For the
reference image of the tensor representation, we
proposed sampling directly from the training data and
averaging procedures for each class. To increase
complexity, we proposed the addition of Gaussian
noise and the application of block encryption. We
also evaluated the expansion of the tensor size and the
number of channels. We also examined the varieties
of value range. In the recognition experiments
conducted for evaluating the proposed methods, our
proposed tensor representations with higher
complexity and larger sizes were as accurate as the
conventional one-hot vector for ordinary data and
more accurate than the conventional tensor
representation labels based on speech spectrograms.
In addition, in the evaluation of resistance to AEs and
experiments with reduced training data, we confirmed
that our proposed labels provide higher accuracy than
conventional labels, including one-hot vectors in
many cases. However, we unfortunately found that no
method was superior in all the cases, and that some
methods are not suitable for certain models and
datasets.
Future works include verification of the
compatibility between label types and the structure of
DNN models, and evaluation using other databases
and DNN models.
REFERENCES
Krizhevsky, A., Sutskever, I. Hinton, G. (2012). Imagenet
classification with deep convolutional neural networks.
In NIPS’12, 26th Conference on Neural Information
Processing Systems.
Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z.
(2016). Rethinking the inception architecture for
computer vision. In CVPR’16, The IEEE/CVF
Conference on Computer Vision and Pattern
Recognition.
Hinton, G., Vinyals, O. Dean, J. (2015). Distilling the
knowledge in a neural network. In NIPS’12, 29th
Conference on Neural Information Processing Systems
Workshop.
Chen, B., Li, Y., Raghupathi, S., Lipson, H. (2021). Beyond
categorical label representations for image
classification. In ICLR’21, 9th International
Conference on Learning Representation.
Szegedy C., Zaremba, W., Sutskever, I., Bruna, J., Erhan,
D., Goodfellow, I., Fergus, R. (2014). Intriguing
properties of neural networks. In ICLR’14, 2th
International Conference on Learning Representation.
Lukasik, M., Bhojanapalli, S., Menon, A., Kumar, S.
(2020). Does label smoothing mitigate label noise? In
ICML’20, 37th International Conference on Machine
Learning.
Krizhevsky, A. (2009). Learning multiple layers of features
from tiny images.
Ehrsam, W., Meyer, C., Smith, J., Tuchman, W. (1976).
Message verification and transmission error detection
by block chaining. US Patent 4074066.
Simonyan, K., Zisserman, A. (2015). Very deep
convolutional networks for large-scale image
recognition. In ICLR’15, 3th International Conference
on Learning Representation.
He, K., Zhang, X., Ren, S., Sun, J. (2016). Deep residual
learning for image recognition. In CVPR’16, The
IEEE/CVF Conference on Computer Vision and
Pattern Recognition.
Goodfellow, I., Shlens, J., Szededy, C. (2015). Explaining
and harnessing adversarial examples. In ICLR’15, 3th
International Conference on Learning Representation.
Kurakin, A., Goodfellow, I., Bengio, S. (2017). Adversarial
examples in the physical world. In ICLR’15,
International Conference on Learning Representation
Workshop.