0 5 10 15 2 0 2 5
m
− 5
0
5
10
15
m
0 5 10 15 20 25
m
− 5
0
5
10
15
m
0 5 10 15 20 2 5
m
− 5
0
5
10
15
m
(a) (b) (c)
Figure 10: Resulting trajectories with LS enabled using (a) noise level 1, (b) noise level 2 and (c) noise level 3. The triangles
represent the AUV orientation at some points and blue lines denote the detected loops.
During the experiments, the NN performed
195021 image comparisons spending, in average, less
than 0.5 ms per image pair. Only 5505 of the com-
pared image pairs were false positives. Even though
this is less than a 3%, they lead to more than 15 m of
mean ATE. Thanks to the LS, however, the number
of false positives was reduced to zero and the ATE to
values between 7.1 cm and 29.7 cm. By executing LS
only with the pairs classified as positives by the NN
the gain in speed is considerable.
Overall, our proposal is able to robust and fastly
search loops, being able to produce high quality Vi-
sual Graph SLAM algorithms even in front of large
odometric noise in underwater scenarios.
ACKNOWLEDGEMENTS
Grant PID2020-115332RB-C33 funded by MCIN /
AEI / 10.13039/501100011033 and, as appropriate,
by ”ERDF A way of making Europe”.
REFERENCES
Arshad, S. and Kim, G. W. (2021). Role of deep learning in
loop closure detection for visual and lidar SLAM: A
survey. Sensors (Switzerland), 21(4):1–17.
Bonin-Font, F., Burguera, A., and Oliver, G. (2013). New
solutions in underwater imaging and vision systems.
In Imaging Marine Life: Macrophotography and Mi-
croscopy Approaches for Marine Biology, pages 23–
47.
Burguera, A. and Bonin-Font, F. (2020). Towards visual
loop detection in underwater robotics using a deep
neural network. Proceedings of VISAPP, 5:667–673.
Ceriani, S., Fontana, G., Giusti, A., Marzorati, D., Mat-
teucci, M., Migliore, D., Rizzi, D., Sorrenti, D. G.,
and Taddei, P. (2009). Rawseeds ground truth collec-
tion systems for indoor self-localization and mapping.
Autonomous Robots, 27(4):353–371.
Irion, J. (2019). Python GraphSLAM. Available at: https:
//github.com/JeffLIrion/python-graphslam.
Latif, Y., Cadena, C., and Neira, J. (2014). Robust graph
SLAM back-ends: A comparative analysis. Proceed-
ings of IEEE/RSJ IROS, (3):2683–2690.
Liu, H., Zhao, C., Huang, W., and Shi, W. (2018). An
End-To-End Siamese Convolutional Neural Network
for Loop Closure Detection in Visual Slam System. In
Proceedings of the IEEE ICASSP, pages 3121–3125.
Lowry, S., Sunderhauf, N., Newman, P., Leonard, J. J.,
Cox, D., Corke, P., and Milford, M. J. (2016). Vi-
sual Place Recognition: A Survey. IEEE Transactions
on Robotics, 32(1):1–19.
Lu, F. and Milios, E. E. (1994). Robot pose estimation in
unknown environments by matching 2D range scans.
Proceedings of the IEEE Computer Society Confer-
ence on Computer Vision and Pattern Recognition,
pages 935–938.
Mahmut Kaya and Hasan Sakir Bilge (2019). Deep Metric
Learning : A Survey. Symmetry, 11.9:1066.
Merril, N. and Huang, G. (2018). Lightweight Unsuper-
vised Deep Loop Closure. In Robotics: Science and
Systems.
Smith, R., Self, M., and Cheeseman, P. (1988). A stochas-
tic map for uncertain spatial relationships. Proceed-
ings of the 4th international symposium on Robotics
Research, (0262022729):467–474.
Thrun, S. and Montemerlo, M. (2006). The graph SLAM
algorithm with applications to large-scale mapping of
urban structures. International Journal of Robotics
Research, 25(5-6):403–429.
VISAPP 2022 - 17th International Conference on Computer Vision Theory and Applications
598