crease which give us 48.64 accuracy compared to the
average pooling which give us 46.5 accuracy. We see
that adding 0.25 % dropout and regularization weights
give us nearly 3.5% improvement. Similarly for the
case of CFAR100 we get increase of 0.75 % with-
out regularization to 1.1% increase with regulariza-
tion and dropout. We see similar trends for the case
of other datasets given in Table 1. For these settings
we get the best results for the case of Tiny ImageNet
which improves the result by almost 3.5 %. The con-
sistent improvement across various datasets show the
stability of the method.
5 CONCLUSIONS
In this paper, we present FlexPooling approach with
and without simple auxiliary classifier(SAC). Flex-
Pooling with SAC is a straightforward but efficient
adaptive pooling technique that learns weighted av-
erage pooling over activations together with the rest
of the network, thus generalizing the idea of aver-
age pooling with consistent improved performance.
In our approach, we make sure that each successive
layer repeats the most salient information from the
prior activations because pooling is a lossy opera-
tion but essential in separating high-level informa-
tion from low-level information. This improves the
network’s discriminability. Secondly for FlexPooling
with SAC our suggested network takes feature maps
from early stages and compress them into linear rep-
resentation via FlexPooling, where no convolutional
processing is involved, yet we downsample the fea-
ture maps into single pixel representation. The loss
values obtained from early stages of the CNN aid the
network in learning more abstract concepts by boost-
ing the gradient signal throughout the entire network.
Our approach learns jointly with the entire network
end to end, enhancing its ability to adapt and ex-
tract more meaningful feature, map representations
that help with the model’s discriminability and gener-
alizability We validate this claim by extensive exper-
iments in single-classifiers as well as multi-classifier
settings. We obtain the stable, increasing accuracy
trend in both settings from the average pool to flex
pool. We show further improvement in accuracy
when each of above mentioned settings are tested with
three different ablation studies, including FlexPool,
FlexPool with regularization and FlexPool with regu-
larization and dropout. Overall, FlexPool with(SAC)
settings attain higher accuracies on average compared
to a single classifier thanks to improved gradient sig-
nal throughout the CNN. Finally, we demonstrate that
our technique consistently outperforms baseline net-
works in image classification across a variety of pop-
ular datasets, resulting in accuracy gains of 1-3%.
REFERENCES
Allen, Z., Zeyuan, L., Yuanzhi, L., and Yingyu (2019).
Learning and generalization in overparameterized
neural networks, going beyond two layers. in ad-
vances in neural information processing systems.
Boureau, Y.-L., Ponce, J., Fr, J. P., and Lecun, Y. (2010a).
A Theoretical Analysis of Feature Pooling in Visual
Recognition.
Boureau, Y.-L., Ponce, J., and LeCun, Y. (2010b). A the-
oretical analysis of feature pooling in visual recogni-
tion. In Proceedings of the 27th International Confer-
ence on International Conference on Machine Learn-
ing, ICML’10, page 111–118, Madison, WI, USA.
Omnipress.
Gholamalinezhad, H. and Khosravi, H. (2020a). Pooling
methods in deep neural networks, a review. arXiv
preprint arXiv:2009.07485.
Gholamalinezhad, H. and Khosravi, H. (2020b). Pooling
Methods in Deep Neural Networks, a Review.
Goodfellow, I. J., Warde-Farley, D., Mirza, M., Courville,
A., and Bengio, Y. (2013). Maxout Networks. 30th In-
ternational Conference on Machine Learning, ICML
2013, (PART 3):2356–2364.
Gulcehre, C., Cho, K., Pascanu, R., and Bengio, Y. (2013).
Learned-Norm Pooling for Deep Feedforward and Re-
current Neural Networks.
He, K., Zhang, X., Ren, S., and Sun, J. Deep Residual
Learning for Image Recognition.
He, K., Zhang, X., Ren, S., and Sun, J. (2014). Spatial Pyra-
mid Pooling in Deep Convolutional Networks for Vi-
sual Recognition. Lecture Notes in Computer Science
(including subseries Lecture Notes in Artificial Intel-
ligence and Lecture Notes in Bioinformatics), 8691
LNCS(PART 3):346–361.
Ionescu, C., Vantzos, O., and Sminchisescu, C. (2015).
Training deep networks with structured lay-
ers by matrix backpropagation. arXiv preprint
arXiv:1509.07838.
Karen Simonyan & Andrew Zisserman (2018). very deep
convolutional networks for large-scale image recogni-
tion. American Journal of Health-System Pharmacy,
75(6):398–406.
Kingma, D. P. and Lei Ba, J. (2015). Adam: A method for
stochastic optimization. ICLR.
Krizhevsky, A. (2009). Learning multiple layers of features
from tiny images. arXiv.
Le, Y. and Yang, X. (2016). Tiny imagenet visual recogni-
tion challenge. arXiv.
Ng, A. (2020). Cs 230 - convolutional neural networks
cheatsheet. Stanford CS230.
Papineni, K., Roukos, S., Ward, T., and Zhu, W.-J. (2001).
BLEU. Proceedings of the 40th Annual Meeting on
Association for Computational Linguistics - ACL ’02,
page 311.
VISAPP 2023 - 18th International Conference on Computer Vision Theory and Applications
504