tional 15% of the total processing time of the full re-
construction technique and improving the results sig-
nificantly.
As future works, we believe that a study that seeks
to improve the accuracy of detecting the bounding
rectangle is needed for BGS-Grab. The detection of
the extreme points of BGS-Deep can improve the ac-
curacy and robustness of the background segmenta-
tion in both pipelines and, consequently, the quality
of the final 3D reconstruction. We also believe that a
study in which partial occlusion and sudden move-
ments are considered in the target object shooting
could also enable the reconstruction from test cases in
which the objects are moved manually. Such a study
would provide an alternative to use turntables during
capture, making the scenario of obtaining the datasets
closer to realistic scenarios. Complementarily, GPU
processing in parts of the 3D mapping pipeline still
processed in the CPU can significantly decrease the
technique processing time.
REFERENCES
Bozic, A., Zollhofer, M., Theobalt, C., and Nießner, M.
(2020). Deepdeform: Learning non-rigid rgb-d recon-
struction with semi-supervised data. In Proceedings
of the IEEE/CVF Conference on Computer Vision and
Pattern Recognition, pages 7002–7012.
Bradski, G. and Kaehler, A. (2000). Opencv. Dr. Dobb’s
journal of software tools, 3.
Goesele, M., Curless, B., and Seitz, S. M. (2006). Multi-
view stereo revisited. In 2006 IEEE Computer Society
Conference on Computer Vision and Pattern Recogni-
tion (CVPR’06), volume 2, pages 2402–2409. IEEE.
Kazhdan, M. and Hoppe, H. (2013). Screened poisson sur-
face reconstruction. ACM Transactions on Graphics
(ToG), 32(3):1–13.
Kuo, P.-C., Chen, C.-A., Chang, H.-C., Su, T.-F., and
Lai, S.-H. (2014). 3d reconstruction with automatic
foreground segmentation from multi-view images ac-
quired from a mobile device. In Asian Conference on
Computer Vision, pages 352–365. Springer.
Locher, A., Perdoch, M., Riemenschneider, H., and
Van Gool, L. (2016). Mobile phone and cloud—a
dream team for 3d reconstruction. In 2016 IEEE Win-
ter Conference on Applications of Computer Vision
(WACV), pages 1–8. IEEE.
Lyra, V. G. d. M., Pinto, A. H., Lima, G. C., Lima, J. P.,
Teichrieb, V., Quintino, J. P., da Silva, F. Q., San-
tos, A. L., and Pinho, H. (2020). Development of
an efficient 3d reconstruction solution from permis-
sive open-source code. In 2020 22nd Symposium on
Virtual and Augmented Reality (SVR), pages 232–241.
IEEE.
Maninis, K.-K., Caelles, S., Pont-Tuset, J., and Van Gool,
L. (2018). Deep extreme cut: From extreme points to
object segmentation. In Proceedings of the IEEE Con-
ference on Computer Vision and Pattern Recognition,
pages 616–625.
Newcombe, R. A., Fox, D., and Seitz, S. M. (2015). Dy-
namicfusion: Reconstruction and tracking of non-
rigid scenes in real-time. In Proceedings of the IEEE
conference on computer vision and pattern recogni-
tion, pages 343–352.
Pan, Q., Reitmayr, G., and Drummond, T. (2009).
Proforma: Probabilistic feature-based on-line rapid
model acquisition. In BMVC, volume 2, page 6. Cite-
seer.
Pokale, A., Aggarwal, A., Jatavallabhula, K. M., and Kr-
ishna, M. (2020). Reconstruct, rasterize and backprop:
Dense shape and pose estimation from a single im-
age. In Proceedings of the IEEE/CVF Conference on
Computer Vision and Pattern Recognition Workshops,
pages 40–41.
Reynolds, D. A. (2009). Gaussian mixture models. Ency-
clopedia of biometrics, 741:659–663.
Rother, C., Kolmogorov, V., and Blake, A. (2004). ” grab-
cut” interactive foreground extraction using iterated
graph cuts. ACM transactions on graphics (TOG),
23(3):309–314.
Schonberger, J. L. and Frahm, J.-M. (2016). Structure-
from-motion revisited. In Proceedings of the IEEE
conference on computer vision and pattern recogni-
tion, pages 4104–4113.
Shunli, W., Qingwu, H., Shaohua, W., Pengcheng, Z., and
Mingyao, A. (2018). A 3d reconstruction and vi-
sualization app using monocular vision service. In
2018 26th International Conference on Geoinformat-
ics, pages 1–5. IEEE.
Sobral, A. and Bouwmans, T. (2014). Bgs library:
A library framework for algorithm’s evaluation in
foreground/background segmentation. In Back-
ground Modeling and Foreground Detection for Video
Surveillance. CRC Press, Taylor and Francis Group.
Talbot, J. F. and Xu, X. (2006). Implementing grabcut.
Brigham Young University, 3.
Thompson, M. M., Eller, R. C., Radlinski, W. A., and
Speert, J. L. (1966). Manual of photogrammetry, vol-
ume 1. American Society of Photogrammetry Falls
Church, VA.
Ullman, S. (1979). The interpretation of structure from mo-
tion. Proceedings of the Royal Society of London. Se-
ries B. Biological Sciences, 203(1153):405–426.
Yu, R., Russell, C., Campbell, N. D., and Agapito, L.
(2015). Direct, dense, and deformable: Template-
based non-rigid 3d reconstruction from rgb video. In
Proceedings of the IEEE international conference on
computer vision, pages 918–926.
Zhang, Z., Feng, X., Liu, N., Geng, N., Hu, S., and Wang,
Z. (2019). From image sequence to 3d reconstructed
model. In 2019 Nicograph International (NicoInt),
pages 25–28. IEEE.
VISAPP 2022 - 17th International Conference on Computer Vision Theory and Applications
800