similarity to other templates maintained in the appear-
ance pool before adding them, together with their cor-
responding histograms, to it.
Instead of using local features to represent the ob-
ject (e.g. (Zhou et al., 2009) used SIFT features, (He
et al., 2009) used SURF features, (Kim, 2008) used
corner features), our approach utilise them to model
the target movement because local features are not
detected enough to cover the whole object. Besides
that, it is hard to decide the object boundary basing
on positions of (few) local features. Feature match-
ing, however, provides clues where the target might
go. In our framework, the MCMC-based search uses
the distribution of motion directions of local image
features from the feature pool to enhance target pre-
diction. These local motion directions are extracted
directly from two consecutive frames. The algorithm
can also handle variation in motion of a target with-
out using any prior knowledge of movement. More-
over, different from methods utilising multiple motion
models to predict the target, our method only uses one
motion model and it is directly derived from the cur-
rent state of the target. Overall, experiments showed
the FMM framework to have performance advantages
over other trackers.
FMM detects target appearance changes using the
templates maintained in the appearance pool. Should
the target change its appearance very often in a long
video sequences, many templates may be stored,
some of which will become irrelevant. To cope with
this problem, some learnt appearances should be re-
moved from the pool. Care must, however, be taken
not to remove appearances which would be useful
later. This will be the subject of future work. Note
also that there is no motion learning mechanism in
FMM. The target motion is derived by detecting and
matching sparse features. These matches could be
used to enhance learning of target motion.
REFERENCES
Adam, A., Rivlin, E., and Shimshoni, I. (2006). Ro-
bust fragments-based tracking using the integral his-
togram. In CVPR 2006, volume 1, pages 798–805.
Babenko, B., Yang, M.-H., and Belongie, S. (2011). Robust
object tracking with online multiple instance learning.
Birchfield, S. (1998). Elliptical head tracking using inten-
sity gradients and color histograms. In CVPR.
Bouguet, J.-Y. (2000). Pyramidal implementation of the lu-
cas kanade feature tracker.
Collins, R., Liu, Y., and Leordeanu, M. (2005). Online se-
lection of discriminative tracking features. PAMI.
Comaniciu, D., Ramesh, V., and Meer, P. (2003). Kernel-
based object tracking. PAMI, 25(5):564 – 577.
Everingham, M., Gool, L., Williams, C., Winn, J., and Zis-
serman, A. (2010). The pascal visual object classes
(voc) challenge. IJCV, 88(2):303–338.
Grabner, H. and Bischof, H. (2006). On-line boosting and
vision. In CVPR, volume 1, pages 260–267.
Grabner, H., Leistner, C., and Bischof, H. (2008). Semi-
supervised on-line boosting for robust tracking. In
ECCV, pages 234–247. Springer-Verlag.
He, W., Yamashita, T., Lu, H., and Lao, S. (2009). Surf
tracking. In ICCV, pages 1586–1592.
Isard, M. and Blake, A. (1996). Contour tracking by
stochastic propagation of conditional density. In
ECCV, pages 343–356, London, UK. Springer-Verlag.
Isard, M. and Blake, A. (1998). A mixed-state condensation
tracker with automatic model-switching. In ICCV.
Khan, Z., Balch, T., and Dellaert, F. (2005). Mcmc-based
particle filtering for tracking a variable number of in-
teracting targets. PAMI, 27(11):1805 –1819.
Kim, Z. (2008). Real time object tracking based on dy-
namic feature grouping with background subtraction.
In CVPR 2008, pages 1–8.
Klein, D., Schulz, D., Frintrop, S., and Cremers, A. (2010).
Adaptive real-time video-tracking for arbitrary ob-
jects. In IROS 2010, pages 772–777.
Kristan, M., Kovacic, S., Leonardis, A., and Pers, J.
(2010). A two-stage dynamic model for visual track-
ing. 40(6):1505–1520.
Kwon, J. and Lee, K. M. (2010). Visual tracking decompo-
sition. In CVPR.
Kwon, J. and Lee, K. M. (2013). Tracking by sampling and
integrating multiple trackers. PAMI, 99:1.
Li, X., Hu, W., Shen, C., Zhang, Z., Dick, A., and Hengel,
A. V. D. (2013). A survey of appearance models in vi-
sual object tracking. ACM Trans. Intell. Syst. Technol.
Matthews, I., Ishikawa, T., and Baker, S. (2004). The tem-
plate update problem. PAMI, 26(6):810–815.
Nummiaro, K., Koller-Meier, E., and Gool, L. V. (2002).
An adaptive color-based particle filter.
Okuma, K., Taleghani, A., Freitas, N. D., Freitas, O. D., Lit-
tle, J. J., and Lowe, D. G. (2004). A boosted particle
filter: Multitarget detection and tracking.
Prez, P., Hue, C., Vermaak, J., and Gangnet, M. (2002).
Color-based probabilistic tracking. In ECCV.
Pridmore, T. P., Naeem, A., and Mills, S. (2007). Managing
particle spread via hybrid particle filter/kernel mean
shift tracking. In Proc. BMVC, pages 70.1–70.10.
Ross, D. A., Lim, J., Lin, R.-S., and Yang, M.-H. (2008).
Incremental learning for robust visual tracking.
Serby, D., Meier, E., and Van Gool, L. (2004). Probabilistic
object tracking using multiple features. In ICPR 2004.
Shi, J. and Tomasi, C. (1994). Good features to track. In
CVPR, pages 593–600.
Wu, Y., Lim, J., and Yang, M.-H. (2013). Online object
tracking: A benchmark. In CVPR 2013.
Yang, F., Lu, H., and Yang, M.-H. (2014). Robust super-
pixel tracking. Image Processing, IEEE Transactions.
Yilmaz, A., Javed, O., and Shah, M. (2006). Object track-
ing: A survey.
Zhou, H., Yuan, Y., and Shi, C. (2009). Object tracking
using {SIFT} features and mean shift. CVIU.
AdaptiveTrackingviaMultipleAppearanceModelsandMultipleLinearSearches
495