jects or their 3D models in the proposed method. It
can suggest the next viewpoint based on just the in-
formation of a single initial view, which along with
the property of considering both the 3D shape and ap-
pearance of objects offers an intrinsic advantage for
active object recognition tasks.
A dataset for testing active object recognition systems
was developed as a part of this work and was used to
evaluate the proposed next best view technique. In
the presence of heavy occlusions in the initial view,
we report 32.9% and 29.1% average accuracy and F
1
score improvements compared to the initial perfor-
mance values.
In continuation to this work, future efforts should be
directed toward probing alternative tiling schemes of
the initial view. Another area of work can be investi-
gating other ensemble methods in place of the current
voting scheme. A meta-learning approach would be a
potentially interesting way to combine the tile scores.
ACKNOWLEDGMENTS
This work has been supported in part by the Office
of Naval Research award N00014-16-1-2312 and US
Army Research Laboratory (ARO) award W911NF-
20-2-0084.
REFERENCES
Atanasov, N., Sankaran, B., Le Ny, J., Pappas, G. J., and
Daniilidis, K. (2014). Nonmyopic view planning for
active object classification and pose estimation. IEEE
Transactions on Robotics, 30(5):1078–1090.
Barzilay, O., Zelnik-Manor, L., Gutfreund, Y., Wagner, H.,
and Wolf, A. (2017). From biokinematics to a robotic
active vision system. Bioinspiration & Biomimetics,
12(5):056004.
Bircher, A., Kamel, M., Alexis, K., Oleynikova, H., and
Siegwart, R. (2016). Receding horizon” next-best-
view” planner for 3d exploration. In 2016 IEEE in-
ternational conference on robotics and automation
(ICRA), pages 1462–1468. IEEE.
Doumanoglou, A., Kouskouridas, R., Malassiotis, S., and
Kim, T.-K. (2016). Recovering 6d object pose and
predicting next-best-view in the crowd. In Proceed-
ings of the IEEE conference on computer vision and
pattern recognition, pages 3583–3592.
Gonzalez, R. C. (2018). Richard E. Woods Digital Image
Processing, Pearson. Prentice Hall.
Hoseini, P., Blankenburg, J., Nicolescu, M., Nicolescu, M.,
and Feil-Seifer, D. (2019a). Active eye-in-hand data
management to improve the robotic object detection
performance. Computers, 8(4):71.
Hoseini, P., Blankenburg, J., Nicolescu, M., Nicolescu, M.,
and Feil-Seifer, D. (2019b). An active robotic vi-
sion system with a pair of moving and stationary cam-
eras. In International Symposium on Visual Comput-
ing, pages 184–195. Springer.
Jia, Z., Chang, Y.-J., and Chen, T. (2010). A general
boosting-based framework for active object recogni-
tion. In British Machine Vision Conference (BMVC),
pages 1–11. Citeseer.
Krainin, M., Curless, B., and Fox, D. (2011). Autonomous
generation of complete 3d object models using next
best view manipulation planning. In 2011 IEEE In-
ternational Conference on Robotics and Automation,
pages 5031–5037. IEEE.
Otsu, N. (1979). A threshold selection method from gray-
level histograms. IEEE transactions on systems, man,
and cybernetics, 9(1):62–66.
Paul, S. K., Chowdhury, M. T., Nicolescu, M., Nicolescu,
M., and Feil-Seifer, D. (2020). Object detection and
pose estimation from rgb and depth data for real-time,
adaptive robotic grasping. In 24th International Con-
ference on Image Processing, Computer Vision, &
Pattern Recognition (IPCV). Springer.
Potthast, C. and Sukhatme, G. S. (2011). Next best view
estimation with eye in hand camera. In IEEE/RSJ Intl.
Conf. on Intelligent Robots and Systems (IROS). Cite-
seer.
Potthast, C. and Sukhatme, G. S. (2014). A probabilistic
framework for next best view estimation in a cluttered
environment. Journal of Visual Communication and
Image Representation, 25(1):148–164.
Rebull Mestres, J. (2017). Implementation of an automated
eye-in hand scanning system using best-path planning.
Master’s thesis, Universitat Polit
`
ecnica de Catalunya.
Wu, Z., Song, S., Khosla, A., Yu, F., Zhang, L., Tang, X.,
and Xiao, J. (2015). 3d shapenets: A deep representa-
tion for volumetric shapes. In Proceedings of the IEEE
conference on computer vision and pattern recogni-
tion, pages 1912–1920.