7 CONCLUSION
In this paper, we describe a novel two stage process
which allows for the generation of photo-realistic vir-
tual views inside colored point clouds. Our approach
has the advantage of being independent from the cap-
turing method and allows for free movements inside
the scene. High-resolution, realistic-looking panora-
mas are produced, which can be used for virtual real-
ity applications. Since this can be done at any posi-
tion and the process is locally and temporally consis-
tent, rendering stereoscopic views is possible as well.
However, the network still produces some unwanted
shadow artifacts. In future work, an even better suited
loss function may be found to reduce those errors. In
addition, the rendering step and network should be
modified such that it also handles smaller view ports
with the same resolution. Further, it would be of
interest to evaluate the performance in a consumer-
centered study.
ACKNOWLEDGEMENTS
This work is funded by Germany’s Federal Ministry
of Education and Research within the project KIMaps
(grant ID 01IS20031C).
REFERENCES
Aliev, K.-A., Ulyanov, D., and Lempitsky, V. (2019).
Neural Point-Based Graphics. arXiv preprint
arXiv:1906.08240.
Anguelov, D., Dulong, C., Filip, D., Frueh, C., Lafon,
S., Lyon, R., Ogale, A., Vincent, L., and Weaver, J.
(2010). Google street view: Capturing the world at
street level. Computer, 43(6):32–38.
Armeni, I., Sax, S., Zamir, A. R., and Savarese, S. (2017).
Joint 2d-3d-semantic data for indoor scene under-
standing. arXiv preprint arXiv:1702.01105.
Bui, G., Le, T., Morago, B., and Duan, Y. (2018). Point-
based rendering enhancement via deep learning. The
Visual Computer, 34(6-8):829–841.
Dai, P., Zhang, Y., Li, Z., Liu, S., and Zeng, B. (2020).
Neural point cloud rendering via multi-plane projec-
tion. In Proceedings of the IEEE/CVF Conference
on Computer Vision and Pattern Recognition, pages
7830–7839.
Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., and Fei-
Fei, L. (2009). Imagenet: A large-scale hierarchical
image database. In 2009 IEEE conference on com-
puter vision and pattern recognition, pages 248–255.
Ieee.
Flynn, J., Neulander, I., Philbin, J., and Snavely, N. (2016).
Deepstereo: Learning to predict new views from the
world’s imagery. In Proceedings of the IEEE Con-
ference on Computer Vision and Pattern Recognition,
pages 5515–5524.
Fu, Z., Hu, W., and Guo, Z. (2018). Point cloud inpainting
on graphs from non-local self-similarity. In 2018 25th
IEEE International Conference on Image Processing
(ICIP), pages 2137–2141. IEEE.
Hedman, P., Alsisan, S., Szeliski, R., and Kopf, J. (2017).
Casual 3d photography. ACM Transactions on Graph-
ics (TOG), 36(6):234.
Johnson, J., Alahi, A., and Fei-Fei, L. (2016). Percep-
tual Losses for Real-Time Style Transfer and Super-
Resolution.
Kerl, C., Sturm, J., and Cremers, D. (2013). Dense visual
SLAM for RGB-D cameras. In 2013 IEEE/RSJ In-
ternational Conference on Intelligent Robots and Sys-
tems, pages 2100–2106. IEEE.
Lai, P. K., Xie, S., Lang, J., and Laqaru
`
ere, R. (2019). Real-
time panoramic depth maps from omni-directional
stereo images for 6 dof videos in virtual reality. In
2019 IEEE Conference on Virtual Reality and 3D
User Interfaces (VR), pages 405–412. IEEE.
Moezzi, S., Tai, L.-C., and Gerard, P. (1997). Virtual view
generation for 3d digital video. IEEE multimedia,
4(1):18–26.
Mordvintsev, A., Olah, C., and Tyka, M. (2015). Inception-
ism: Going deeper into neural networks.
Mur-Artal, R. and Tard
´
os, J. D. (2017). Orb-slam2:
An open-source slam system for monocular, stereo,
and rgb-d cameras. IEEE Transactions on Robotics,
33(5):1255–1262.
NavVis GmbH (2019). https://www.navvis.com.
Pfister, H., Zwicker, M., Van Baar, J., and Gross, M. (2000).
Surfels: Surface elements as rendering primitives. In
Proceedings of the 27th annual conference on Com-
puter graphics and interactive techniques, pages 335–
342. ACM Press/Addison-Wesley Publishing Co.
Ronneberger, O., Fischer, P., and Brox, T. (2015). U-net:
Convolutional networks for biomedical image seg-
mentation. In International Conference on Medical
image computing and computer-assisted intervention,
pages 234–241. Springer.
Sahay, P. and Rajagopalan, A. N. (2015). Geometric in-
painting of 3d structures. In Proceedings of the IEEE
Conference on Computer Vision and Pattern Recogni-
tion Workshops, pages 1–7.
Schubert, S., Neubert, P., P
¨
oschmann, J., and Pretzel, P.
(2019). Circular Convolutional Neural Networks for
Panoramic Images and Laser Data. In 2019 IEEE
Intelligent Vehicles Symposium (IV), pages 653–660.
IEEE.
Simonyan, K. and Zisserman, A. (2014). Very Deep Con-
volutional Networks for Large-Scale Image Recogni-
tion.
Sung, J. W., Jeon, Y. J., Jeon, B. M., and Lim, J. H. (2014).
Virtual view image synthesis method and apparatus.
VISAPP 2021 - 16th International Conference on Computer Vision Theory and Applications
276