curve for each point detector when averaged across
all datasets:
Dataset / Error Type E
rot
(θ) E
dist
E
proj
(%)
Harris 140 1.63 0.81
SIFT 149 1.70 0.85
Saliency 169 1.72 0.88
Whilst our method worked well on the Stanford
dragon, it was outperformed by the Harris corner de-
tector and SIFT; this may be due to the large amount
of small corners on it, allowing the other methods to
be better suited. In particular, Kadir-Brady saliency
may not work as well for small features since their en-
tropy may not be peaked in scale space (the entropy is
high for the smallest scale and then decreases). Many
of the results could potentially have been localised to
within 1
◦
if a better method had been used in the re-
finement stage. In particular, numerous edges were
extracted from images of the temple, producing a lot
of noise for the refinement. This could be solved by,
for example, using Mutual Information for refinement
as in (Mastin et al., 2009).
5 CONCLUSIONS
In this paper we have presented a generalisation of
Kadir-Brady saliency to an arbitrary number of di-
mensions and provided a novel curvature based ex-
tension to 3D. Further, we have used this as a fil-
ter for the 2D / 3D registration problem and shown
saliency to be a superior filter for point-based regis-
tration, demonstrating its consistency across different
modalities. This is due to its principled information-
theoretic approach, however it is only a proxy for true
saliency and improvements can be made here. Future
work may include line or curve-based saliency, esti-
mating the focal length of the camera as well as its ex-
trinsics and the extension of the principle of saliency-
based filtering to other multi-modal problems.
ACKNOWLEDGEMENTS
This research was executed with the financial sup-
port of the EPSRC and EU ICT FP7 project IMPART
(grant agreement No. 316564).
REFERENCES
Arvo, J. (1992). Fast random rotation matrices. pages 117–
120.
Chung, A. J., Deligianni, F., Hu, X.-P., and Yang, G.-Z.
(2004). Visual feature extraction via eye tracking for
saliency driven 2D/3D registration. In Proc. of the
2004 symposium on Eye tracking research & applica-
tions, ETRA ’04, pages 49–54.
David, P., DeMenthon, D., Duraiswami, R., and Samet, H.
(2002). Softposit: Simultaneous pose and correspon-
dence determination. In Proc. of the 7th European
Conference on Computer Vision-Part III, ECCV ’02,
pages 698–714.
Egnal, G. (2000). Mutual information as a stereo correspon-
dence measure. Technical report, Univ. of Pennsylva-
nia.
Enqvist, O., Josephson, K., and Kahl, F. (2009). Optimal
correspondences from pairwise constraints. In Inter-
national Conference on Computer Vision.
Fitzgibbon, A. W. (2001). Robust registration of 2D and
3D point sets. In British Machine Vision Conference,
pages 662–670.
Gendrin, C., Furtado, H., Weber, C., Bloch, C., Figl, M.,
Pawiro, S. A., Bergmann, H., Stock, M., Fichtinger,
G., Georg, D., and Birkfellner, W. (2011). Monitoring
tumor motion by real time 2d/3d registration during
radiotherapy. Radiother Oncol.
Granger, S. and Pennec, X. (2002). Multi-scale em-icp:
A fast and robust approach for surface registration.
In European Conference on Computer Vision (ECCV
2002), volume 2353 of LNCS, pages 418–432.
Guillemaut, J.-Y. and Hilton, A. (2011). Joint multi-layer
segmentation and reconstruction for free-viewpoint
video applications. Int. J. Comput. Vision, 93(1):73–
100.
Huynh, D. Q. (2009). Metrics for 3d rotations: Comparison
and analysis. J. Math. Imaging Vis., 35(2):155–164.
Itti, L. (2000). A saliency-based search mechanism for overt
and covert shifts of visual attention. Vision Research,
40(10-12):1489–1506.
Judd, T., Durand, F., and Torralba, A. (2012). A benchmark
of computational models of saliency to predict human
fixations. Technical report, MIT.
Kadir, T. and Brady, M. (2001). Saliency, scale and image
description. Int. J. Comput. Vision, 45(2):83–105.
Kotsas, P. and Dodd, T. (2011). A review of methods for
2d/3d registration. World Academy of Science, Engi-
neering and Technology.
Lee, C. H., Varshney, A., and Jacobs, D. W. (2005). Mesh
saliency. ACM Trans. Graph., 24(3):659–666.
Mastin, A., Kepner, J., and Fisher III, J. W. (2009). Auto-
matic registration of lidar and optical images of urban
scenes. In Computer Vision and Pattern Recognition,
2009. CVPR 2009. IEEE Conference on.
Moreno-Noguer, F., Lepetit, V., and Fua, P. (2008). Pose
priors for simultaneously solving alignment and cor-
respondence. In Proc. of the 10th European Confer-
ence on Computer Vision: Part II, ECCV ’08, pages
405–418.
Seitz, S. M., Curless, B., Diebel, J., Scharstein, D., and
Szeliski, R. (2006). A comparison and evaluation of
multi-view stereo reconstruction algorithms. In Proc.
of the 2006 IEEE Computer Society Conference on
VISAPP2014-InternationalConferenceonComputerVisionTheoryandApplications
272