values are higher in the case of WaPT.
All exposed data confirms that, in this test, the
WaPT algorithm improves the accuracy in the tem-
plate matching process, getting consequently more
accurate feature positions. In autonomous navigation
systems the tracking process has to be done in large
environments where the data from sensors help to im-
prove the tracking process. So, matching improve-
ments are not critical.
Nevertheless, in augmented reality applications,
the accuracy is very important to visualize the vir-
tual elements correctly, that is, drift and jitter must
be reduced as much as possible. In augmented reality
applications this contribution might provide a valu-
able improvement. It is more accurately than template
matching algorithms used in traditional methods.
5 CONCLUSIONS AND FUTURE
WORK
The work presented in this paper proposes a new in-
ternal representation of the environment for marker-
less tracking. Besides the point cloud, a normal vector
for each point is also stored.
In the 3D reconstruction process, not only the 3D
points of the environment are calculated. A minimiza-
tion process is also run in order to estimate the best
normal for each point.
In the tracking process these normals are used to
obtain an improved warped template in order to obtain
a more precise matching.
Perspective deformations are reduced when tem-
plate matching techniques and patches as features are
used.
The results obtained in Section 4 are a first vali-
dation. In this validation, the similarity of the patches
are calculated in order to know the matching process
precision. The results have been favourable to WaPT,
proving its higher accuracy. This approach provides
more precision than traditional methods, which as-
sume that the points are facing the camera.
When more experiments confirm the results pro-
vided by this paper, new research should be done in
order to find methods that accelerate the reconstruc-
tion process for surface normals. One possible ap-
proach could be to start assuming that all surfaces
face the camera as (Davison et al., 2007) and, on-line,
perform a progressive improvement. These normals
might be also useful for surface illumination com-
pensation when comparing images taken in different
lighting conditions.
As future work more validations are planed. Se-
quences with different perspectives, where points do
not face the camera, will be used. In this kind of se-
quences, the traditional methods should be even more
compromised.
In WaPT, the Reference patch is located in the
Reference keyFrame, which is the keyFrame where
the point was first seen. In order to improve more the
precision and the patch similarities, it is planed to use
as the Reference keyframe the one where the point is
visible and it is more similar to the current image.
ACKNOWLEDGEMENTS
Work partially funded by the Spanish Ministry of
Economy and Competitiveness. Project ELAS-
TRACK (TIN2012-33879).
REFERENCES
Alahi, A., Ortiz, R., and Vandergheynst, P. (2012). Freak:
Fast retina keypoint. In Computer Vision and Pat-
tern Recognition (CVPR), 2012 IEEE Conference on,
pages 510–517. Ieee.
Barron, J. L., Fleet, D. J., and Beauchemin, S. S. (1994).
Performance of optical flow techniques. 12:43–77.
Bay, H., Tuytelaars, T., and Van Gool, L. (2006). Surf:
Speeded up robust features. In Computer Vision–
ECCV 2006, pages 404–417. Springer.
Davison, A. J. (2003). Real-time simultaneous localisa-
tion and mapping with a single camera. In Computer
Vision, 2003. Proceedings. Ninth IEEE International
Conference on, pages 1403–1410. IEEE.
Davison, A. J., Reid, I. D., Molton, N. D., and Stasse, O.
(2007). Monoslam: Real-time single camera slam.
In Pattern Analysis and Machine Intelligence, IEEE
Transactions on, volume 29, pages 1052–1067. IEEE.
Dellaert, F., Seitz, S. M., Thorpe, C. E., and Thrun, S.
(2000). Structure from motion without correspon-
dence. In Computer Vision and Pattern Recognition,
2000. Proceedings. IEEE Conference on, volume 2,
pages 557–564. IEEE.
Furukawa, Y. and Ponce, J. (2010). Accurate, dense, and
robust multiview stereopsis. 32(8):1362–1376.
Hartley, R. and Zisserman, A. (2003). Multiple view geom-
etry in computer vision. Cambridge university press.
Klein, G. and Murray, D. (2007). Parallel tracking and map-
ping for small ar workspaces. In Mixed and Aug-
mented Reality, 2007. ISMAR 2007. 6th IEEE and
ACM International Symposium on, pages 225–234.
IEEE.
Lowe, D. G. (2004). Distinctive image features from scale-
invariant keypoints. 60:91–110.
Molton, N., Davison, A. J., and Reid, I. (2004). Locally pla-
nar patch features for real-time structure from motion.
In BMVC, pages 1–10.
VISAPP2015-InternationalConferenceonComputerVisionTheoryandApplications
502