A Combined Activation Function for Learning Performance
Improvement of CNN Image Classification
Guangliang Pan
1, a
, Jun Li
2, b, *
, Fei Lin
2, c
, Tingting Sun
3, d
, Yulin Sun
1, e
1
School of Electrical Engineering and Automation, Qilu University of Technology(Shandong Academy of Sciences), Jinan
250353, China
2
School of Electronic and Information Engineering (Department of Physics ), Qilu University of Technology (Shandong
Academy of Sciences),Jinan 250353,China
3
Weihai Ocean Vocational College, Weihai 264300, China
e
alanlylsun@163.com
Keywords: Convolutional neural network, LeNet-5, sigmoid, tanh-relu, tanh, relu,
Abstract: With the rise of artificial intelligence, it has unlimited possibilities for machines to replace human work.
Aiming at how to improve the learning performance of convolutional neural network (CNN) image
classification by changing the activation function, a combined Tanh-relu activation function is proposed
based on the single Sigmoid, Tanh and Relu activation functions. Based on CNN-LeNet-5, the size of the
convolution kernel and sampling window is changed and the number of layers of the convolutional neural
network is reduced. At the same time, the network structure of the LeNet-5 model is improved. On the
Mnist handwritten digital dataset, the combined Tanh-relu activation function was compared with a single
activation function. The experimental results show that the CNN model with combined Tanh-relu activation
function has faster accuracy fitting speed and higher accuracy, improves the convergence speed of loss and
enhances the convergence performance of CNN model.
1 INTRODUCTION
In the information age, the explosive growth of data
volume makes the display of deep learning (DL)
extraordinarily important (Meyer P, Noblet V ,
Mazzara C , et al, 2018). As a common model of
deep learning, convolutional neural networks have
achieved great success in image processing (Arena
P , Basile A , Bucolo M , et al, 2003; Al-Ajlan
Amani, El Allali Achraf, 2018). In order to make the
neural network learn more complex data, the
activation function introduces the nonlinear input
into the neural network by converting the input
signal of the node into the output signal, enhancing
the learning ability of the neural network model and
improving the classification performance.
At the beginning of the introduction of the
activation function, mainly the Sigmoid and Tanh
activation functions are mainly applied to various
neural network models. Both of these activation
functions are saturated S-type activation functions,
which are prone to gradient dispersion during neural
network training. For the study of activation
functions, people have never stopped. Kr-izhevsky
used Relu (corrected linear unit) as the activation
function for the first time in the ImageNet ILSVRC
competition (Krizhevsky, Alex, I. Sutskever, and G.
Hinton, 2012). All of the above studies have studied
a single activation function, without considering the
use of a single activation function. Improving the
convolutional neural network structure is also an
important way to optimize the learning performance
of the model (Horn Z C, Auret L, Mccoy J T, et al,
2017).
Inspired by the literature (Qian S , Liu H , Liu C ,
et al, 2017, Yao G , Lei T , Zhong J, 2018), this
paper combines the advantages and disadvantages of
the three activation functions, and reconstructs the
LeNet-5 model of the convolutional neural network
structure, reduces the layer of the fully connected
layer and changes the size of the convolution kernel.
A combined activation function is applied to
convolutional neural network image classification to
improve the image classification performance of
convolutional neural networks.
360
Pan, G., Li, J., Lin, F., Sun, T. and Sun, Y.
A Combined Activation Function for Learning Performance Improvement of CNN Image Classification.
DOI: 10.5220/0008851103600366
In Proceedings of 5th International Conference on Vehicle, Mechanical and Electrical Engineering (ICVMEE 2019), pages 360-366
ISBN: 978-989-758-412-1
Copyright
c
2020 by SCITEPRESS – Science and Technology Publications, Lda. All rights reserved