Figure 10: Results for each character obtained from the user
study showing the mean and variance for each setting.
gested approach was to create new poses by estima-
ting the dense optical flow between the transitioning
frames in each direction and then move the pixels step
wise towards the flow and blend the frames from each
direction.
The suggested animation based on optical flow
was compared with animations using a direct al-
pha blending between the transitioning frames and
not performing any animation. This was evaluated
through a web based user study between three diffe-
rent characters. The results showed consistency over
all characters, where the suggested animation with
different similarity measure settings had a higher rate
of preference and perceived realism than using anima-
tions based on alpha blending and using no animation.
Creating an adaptive number of intermediate fra-
mes, which should vary according to the the pose si-
milarity and speed of motion before and after transi-
tioning, could be a possible area for future investiga-
tion.
REFERENCES
Budd, C., Huang, P., Klaudiny, M., and Hilton, A. (2013).
Global non-rigid alignment of surface sequences.
International Journal of Computer Vision, 102(1-
3):256–270.
Casas, D., Tejera, M., Guillemaut, J.-Y., and Hilton, A.
(2013). Interactive animation of 4d performance cap-
ture. IEEE transactions on visualization and computer
graphics, 19(5):762–773.
Casas, D., Volino, M., Collomosse, J., and Hilton, A.
(2014). 4d video textures for interactive character ap-
pearance. In Computer Graphics Forum, volume 33,
pages 371–380. Wiley Online Library.
De Aguiar, E., Stoll, C., Theobalt, C., Ahmed, N., Sei-
del, H.-P., and Thrun, S. (2008). Performance capture
from sparse multi-view video. In ACM Transactions
on Graphics (TOG), volume 27, page 98. ACM.
Farneb
¨
ack, G. (2003). Two-frame motion estimation based
on polynomial expansion. In Scandinavian conference
on Image analysis, pages 363–370. Springer.
Fechteler, P., Paier, W., and Eisert, P. (2014). Articulated
3d model tracking with on-the-fly texturing. In 2014
IEEE International Conference on Image Processing
(ICIP), pages 3998–4002. IEEE.
Huang, P., Hilton, A., and Starck, J. (2010). Shape simila-
rity for 3d video sequences of people. International
Journal of Computer Vision, 89(2-3):362–381.
Kovar, L., Gleicher, M., and Pighin, F. (2002). Motion
graphs. In ACM transactions on graphics (TOG), vo-
lume 21, pages 473–482. ACM.
Møgelmose, A., Bahnsen, C., and Moeslund, T. B. (2015).
Comparison of multi-shot models for short-term re-
identification of people using rgb-d sensors. In Inter-
national Joint Conference on Computer Vision, Ima-
ging and Computer Graphics Theory and Applicati-
ons 2015.
Starck, J. and Hilton, A. (2007). Surface capture
for performance-based animation. IEEE Computer
Graphics and Applications, 27(3):21–31.
Xu, F., Liu, Y., Stoll, C., Tompkin, J., Bharaj, G., Dai,
Q., Seidel, H.-P., Kautz, J., and Theobalt, C. (2011).
Video-based characters: creating new human perfor-
mances from a multi-view video database. ACM Tran-
sactions on Graphics (TOG), 30(4):32.
Low Cost Video Animation of People using a RGBD Sensor
249