ral networks in fine-grained classification applications
as they contain more number of layers with more
number of trainable parameters which the model can
use for learning discriminative features allowing it to
distinguish between similar classes. Also, adding a
dense layer to the InceptionResnetv2 allows for more
features to be extracted while using L2 regularization
to prevent overfitting the training dataset.
Our future work will include training and evalu-
ating our models using (Buzzelli and Segantin, 2021)
dataset to provide us with a larger and a broader set of
annotations than the Stanford car dataset as it includes
Type-level annotations for all CompCars models.
REFERENCES
Alsahafi, Y., Lemmond, D., Ventura, J., and Boult, T.
(2019). Carvideos: A novel dataset for fine-grained
car classification in videos. In 16th International Con-
ference on Information Technology-New Generations
(ITNG 2019), pages 457–464. Springer.
Benjdira, B., Khursheed, T., Koubaa, A., Ammar, A., and
Ouni, K. (2019). Car detection using unmanned
aerial vehicles: Comparison between faster r-cnn and
yolov3. In 2019 1st International Conference on
Unmanned Vehicle Systems-Oman (UVS), pages 1–6.
IEEE.
Buzzelli, M. and Segantin, L. (2021). Revisiting the com-
pcars dataset for hierarchical car classification: New
annotations, experiments, and results. Sensors, 21(2).
Dehghan, A., Masood, S. Z., Shu, G., Ortiz, E., et al.
(2017). View independent vehicle make, model and
color recognition using convolutional neural network.
arXiv preprint arXiv:1702.01721.
Fomin, I., Nenahov, I., and Bakhshiev, A. (2020). Hierar-
chical system for car make and model recognition on
image using neural networks. In 2020 International
Conference on Industrial Engineering, Applications
and Manufacturing (ICIEAM), pages 1–6. IEEE.
Gao, Y. and Lee, H. J. (2015). Vehicle make recognition
based on convolutional neural network. In 2015 2nd
International Conference on Information Science and
Security (ICISS), pages 1–4. IEEE.
Gavali, P. and Banu, J. S. (2020). Bird species identification
using deep learning on gpu platform. In 2020 Interna-
tional Conference on Emerging Trends in Information
Technology and Engineering (ic-ETITE), pages 1–6.
Girshick, R., Donahue, J., Darrell, T., and Malik, J. (2015).
Region-based convolutional networks for accurate ob-
ject detection and segmentation. IEEE transactions on
pattern analysis and machine intelligence, 38(1):142–
158.
He, K., Zhang, X., Ren, S., and Sun, J. (2016). Deep resid-
ual learning for image recognition. In Proceedings of
the IEEE conference on computer vision and pattern
recognition, pages 770–778.
Hu, Q., Wang, H., Li, T., and Shen, C. (2017). Deep cnns
with spatially weighted pooling for fine-grained car
recognition. IEEE Transactions on Intelligent Trans-
portation Systems, 18(11):3147–3156.
Krause, J., Stark, M., Deng, J., and Fei-Fei, L. (2013). 3d
object representations for fine-grained categorization.
In Proceedings of the IEEE international conference
on computer vision workshops, pages 554–561.
Liu, D. and Wang, Y. (2017). Monza: image classification
of vehicle make and model using convolutional neural
networks and transfer learning.
Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S.,
Fu, C.-Y., and Berg, A. C. (2016). Ssd: Single shot
multibox detector. In European conference on com-
puter vision, pages 21–37. Springer.
Nguyen, N., Le, V., Le, T., Hai, V., Pantuwong, N., and
Yagi, Y. (2016). Flower species identification using
deep convolutional neural networks.
Redmon, J. and Farhadi, A. (2018). Yolov3: An incremental
improvement. arXiv preprint arXiv:1804.02767.
Ren, S., He, K., Girshick, R., and Sun, J. (2015). Faster
r-cnn: Towards real-time object detection with region
proposal networks. In Advances in neural information
processing systems, pages 91–99.
Rosebrock, A. (2018). Simple object tracking with opencv.
In Online.
Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., and
Chen, L.-C. (2018). Mobilenetv2: Inverted residu-
als and linear bottlenecks. In Proceedings of the IEEE
conference on computer vision and pattern recogni-
tion, pages 4510–4520.
Szegedy, C., Ioffe, S., Vanhoucke, V., and Alemi, A.
(2016a). Inception-v4, inception-resnet and the im-
pact of residual connections on learning. arXiv
preprint arXiv:1602.07261.
Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., and Wo-
jna, Z. (2016b). Rethinking the inception architecture
for computer vision. In Proceedings of the IEEE con-
ference on computer vision and pattern recognition,
pages 2818–2826.
Yang, L., Luo, P., Change Loy, C., and Tang, X. (2015).
A large-scale car dataset for fine-grained categoriza-
tion and verification. In Proceedings of the IEEE con-
ference on computer vision and pattern recognition,
pages 3973–3981.
Video-based Car Make, Model and Year Recognition
91