the camera motion as it is more visible in Figure 2. in
case of the TIS approach (c). It was also shown that
the detected salient regions by the proposed approach
(MC-TIS) have a large overlap with the locations of
human eye movement fixations as compared to other
saliency algorithms.
Our proposedsystem may fail in some difficult sit-
uations, such as in case of more severe camera mo-
tions or camera motion is wrongly estimated. So the
success depends strongly on the quality and accuracy
of the used motion estimation method as well. Of
course the proposed approach can also prove effec-
tive in other computer vision problems, e.g. in object
categorization or object recognition, video encoding
where compression plays an important role.
Interesting avenues for future research are to in-
vestigate the combination of motion estimation and
saliency algorithms for the application of intelligent
video compression.
ACKNOWLEDGEMENTS
This work was supported by the ROMEO project
(grant number: 287896), funded by the EC FP7 ICT
collaborative research programme.
REFERENCES
Abdollahian, G. and Edward J, D. (2007). Finding regions
of interest in home videos based on camera motion. In
IEEE International Conference on Image Processing
(ICIP), volume 4.
Achanta, R., Hemami, S. S., Estrada, F. J., and S¨usstrunk,
S. (2009). Frequency-tuned salient region detection.
In CVPR, pages 1597–1604. IEEE.
Achanta, R. and S¨usstrunk, S. (2009). Saliency detection
for content-aware image resizing. In IEEE Intl. Conf.
on Image Processing.
Ali Borji, L. I. (2013). State-of-the-art in visual attention
modeling. In IEEE transactions on Pattern Analysis
and Machine Intelligence, volume 35, pages 185–207.
Borji. A, Tavakoli. H, S. D. and Itti, L. (2013). Analysis of
scores, datasets, and models in visual saliency predic-
tion. In Proceedings of the IEEE International Con-
ference on Computer Vision, pages 921–928.
C. Chamaret, J. C. C. and Meur, O. L. (2010). Spatio-
temporal combination of saliency maps and eye-
tracking assessment of different strategies. In Proc.
IEEE Int. Conf. Image Process, pages 1077–1080.
Chen, Y.-M. and Baji´c, I. V. (2010). Motion vector outlier
rejection cascade for global motion estimation. IEEE
Signal Process. Lett, 17(2):197–200.
Cheng, M.-M., Zhang, G.-X., Mitra, N. J., Huang, X., and
Hu, S.-M. (2011). Global contrast based salient region
detection. In CVPR, pages 409–416.
Cui, X., Liu, Q., and Metaxas, D. (2009). Temporal spec-
tral residual: fast motion saliency detection. In Pro-
ceedings of the 17th ACM international conference on
Multimedia, MM ’09, pages 617–620, New York, NY,
USA. ACM.
Deigmoeller, J. (2010). Intelligent image cropping and scal-
ing. In PhD thesis, Brunel University.
Guo, C., Ma, Q., and Zhang, L. (2008). Spatio-temporal
saliency detection using phase spectrum of quaternion
fourier transform. In CVPR’08.
H. Hadizadeh, M. J. Enriquez, I. V. B. (2012). Eye-tracking
database for a set of standard video sequences. IEEE
Trans. on Image Processing, 21(2):898–903.
Hadi Hadizadeh, I. V. B. (2014). Saliency-aware video
compression. IEEE Trans. on Image Processing,
23(1):19–33.
Han, J., Ngan, K. N., Li, M., and Zhang, H. (2006). Un-
supervised extraction of visual attention objects in
color images. IEEE Trans. Circuits Syst. Video Techn.,
16(1):141–145.
Hou, X., Harel, J., and Koch, C. (2012). Image signature:
Highlighting sparse salient regions. IEEE Trans. Pat-
tern Anal. Mach. Intell., 34(1):194–201.
Hou, X. and Zhang, L. (2007). Saliency detection: A
spectral residual approach. In IEEE Conference on
Computer Vision and Pattern Recognition (CVPR07).
IEEE Computer Society, pages 1–8.
Itti, L. and Koch, C. (2001). Computational modelling
of visual attention. Nature Review Neuroscience,
2(3):194–203.
Itti, L., Koch, C., and Niebur, E. (1998). A model
of saliency-based visual attention for rapid scene
analysis. IEEE Trans. Pattern Anal. Mach. Intell.,
20(11):1254–1259.
M. Cerf, E. P. F. and Koch, C. (2009). Faces and text attract
gaze independent of the task: Experimental data and
computer model. In Journal of vision, volume 9.
Qureshi, H. (2013). Dct based temporal image signature
approach. Proceedings of the 8th International Con-
ference on Computer Vision Theory and Applications
(VISAPP ’13), 1:208–212.
Qureshi, H. and Ludwig, M. (2013). Improving temporal
image signature approach by adding face conspicuity
map. Proceedings of the 2nd ROMEO Workshop.
Schauerte, B. and Stiefelhagen, R. (2012). Predicting
human gaze using quaternion dct image signature
saliency and face detection. In Proceedings of the
IEEE Workshop on the Applications of Computer Vi-
sion (WACV). IEEE.
Treisman, A. (1986). Features and objects in visual pro-
cessing. Sci. Am., 255(5):114–125.
Treisman, A. M. and Gelade, G. (1980). A feature-
integration theory of attention. Cognitive Psychology,
12:97–136.
Y.-F. Ma, X.-S. Hua, L. L. and Zhang, H.-J. (2005). A
generic framework of user attention model and its ap-
plication in video summarization. In Trans. Multi, vol-
ume 7, pages 907–919.
VISAPP2015-InternationalConferenceonComputerVisionTheoryandApplications
516