center on the panoramic image is derived. It was
shown analytically that the inclination may be
approximated by a sine wave.
In A, the sine wave is fitted by the least squares
method, whereas in B, robust and accurate fitting is
realized by eliminating outliers using RANSAC.
In A, the estimation accuracy is 3.3° for the first
correction and 1.5° for the second one, but in B, an
estimation accuracy of 0.1° is obtained without the
need for a second step (when the effect of noise can
be eliminated).
For A, it was reported that there were no failure
examples in an experiment of 40 examples, but only
one example was shown. For B, many experiments
were carried out in various environments, and
concrete examples are shown.
7 CONCLUSIONS
In this paper, we proposed a method for upright
adjustment of a panoramic image by detecting the
inclination of the camera from a pre-corrected
panoramic image with high accuracy and at high
speed using a vertical line existing in an indoor
environment or an outdoor environment near a
building.
Because of the nature of this method, it cannot be
used in an environment without vertical lines, but it is
useful in many environments in which autonomous
robots are expected to operate in the future, such as
normal indoor environments, construction sites, and
in and around warehouses. When the lengths of the
projection curves are extremely short, the method is
easily affected by noise, and the upright adjustment
tends to be unstable. However, even in this case, the
value of pthresh can be added as a certainty and the
correct handling can be performed in the post-
processing.
In future, this method will be applied to tasks such
as the self-localization of autonomous robots,
reconstruction of 3D environments, and object
recognition. In addition, by integrating self-
localization estimation and a 3D environment model,
speed-up and robustness will be achieved by learning
the parameters of the upright adjustment depending
on location.
REFERENCES
Bazin, J.-C., Demonceaux, C., et al. 2012. Rotation
estimation and vanishing point extraction by
omnidirectional vision in urban environment. The
International Journal of Robotics Research 31(1): 63-
81.
Bosse, M., Rikoski, R., et al. 2002. Vanishing points and 3d
lines from omnidirectional video. Proceedings.
International Conference on Image Processing, IEEE.
Canny, John 1986. A computational approach to edge
detection. IEEE Transactions on Pattern Analysis and
Machine Intelligence, 8(6): 679-698.
Cipolla, R., Robertson, D., et al. 1999. Photobuilder-3D
models of architectural scenes from uncalibrated
images. Proceedings IEEE International Conference
on Multimedia Computing and Systems, IEEE.
Coughlan, J. M. and Yuille, A. L. 1999. Manhattan world:
Compass direction from a single image by bayesian
inference. Proceedings of the seventh IEEE
international conference on computer vision, IEEE.
Demonceaux, C., Vasseur, P., et al. 2006. Omnidirectional
vision on UAV for attitude computation. Proceedings
2006 IEEE International Conference on Robotics and
Automation, 2006. ICRA 2006., IEEE.
Demonceaux, C., Vasseur, P., et al. 2007. UAV attitude
computation by omnidirectional vision in urban
environment. Proceedings 2007 IEEE International
Conference on Robotics and Automation, IEEE.
Fischler, A., Martin and Bolles, C., Robert 1981. Random
sample consensus: a paradigm for model fitting with
applications to image analysis and automated
cartography. Communications of the ACM, 24(6): 381-
395.
Gallagher, A. C. 2005. Using vanishing points to correct
camera rotation in images. The 2nd Canadian
Conference on Computer and Robot Vision (CRV'05),
IEEE.
Illingworth, John and Kittler, Josef 1988. A survey of the
Hough transform. Computer vision, graphics and
image processing, 44(1): 87-116.
Jayasuriya, M., Ranasinghe, R., et al. 2020. Active
Perception for Outdoor Localisation with an
Omnidirectional Camera. IEEE/RSJ International
Conference on Intelligent Robots and Systems.
Jeon, J., Jung, J., et al. 2018. Deep Upright Adjustment of
360 Panoramas Using Multiple Roll Estimations. Asian
Conference on Computer Vision, Springer.
Joo, K., Oh, T.-H., et al. 2018. Globally optimal inlier set
maximization for Atlanta frame estimation.
Proceedings of the IEEE Conference on Computer
Vision and Pattern Recognition.
Jung, J., Kim, B., et al. 2017. Robust upright adjustment of
360 spherical panoramas. The Visual Computer 33(6-
8): 737-747.
Jung, R., Lee, A. S. J., et al. 2019. Deep360Up: A Deep
Learning-Based Approach for Automatic VR Image
Upright Adjustment. 2019 IEEE Conference on Virtual
Reality and 3D User Interfaces (VR), IEEE.
Kawai, N. 2019. A method for rectifying inclination of
panoramic images. ACM SIGGRAPH 2019 Posters,
ACM.
Lee, H., Shechtman, E., et al. 2013. Automatic upright
adjustment of photographs with robust camera