Figure 12: Exemplary results of the view synthesis: (a) original camera view, (b) synthesis with source data, (c) voxel
resampling, (d) mesh resampling, (d) and (e) temporal filtering depth of k = 2 for voxel resampling and mesh resampling,
respectively.
resampling (d) without temporal filtering, the respec-
tive synthesis results with a temporal filtering depth
k = 2 in (e) and (f). While effects in the temporal
domain are hard to notice from single frames, synthe-
sis artifacts related to depth map artifacts on object
boundaries and revealed occlusions can be noticed
in (b) which do not occur in (c) to (f). Overall, the
quality of synthesis results with enhanced depth maps
does not vary significantly with a negligible positive
margin for the voxel based approach and temporal fil-
tering.
4 SUMMARY AND
CONCLUSIONS
This work presents algorithms to resample and filter
point cloud data reconstructed from multiple cameras
and multiple time instants. In an initial resampling
stage a voxel and a surface mesh based approach are
presented to resample the point cloud data into a com-
mon sampling grid. Subsequently, the resampled data
undergoes a filtering stage based on clustering to re-
move artifacts of depth estimation and achieve spa-
tiotemporal consistency. The presented algorithms
are evaluated in a view synthesis scenario. Results
show that view synthesis with enhanced depth maps
as produced by the algorithms leads to less artifacts
than synthesis with the original source data. The dif-
ference in achieved quality between the voxel and the
surface mesh based approach is negligible and with
regard to the computational complexity of the surface
mesh reconstruction, the voxel based approach is the
desirable solution for resampling.
Filtering in the temporal domain shows slight syn-
thesis quality improvements when moving objects are
confined to a limited region of the scene as in the bal-
let data set. For data sets in which moving objects
cover larger areas such as in the breakdancer data set,
temporal filtering does not improve synthesis results
compared to filtering across cameras. The presented
motion masking excludes samples within a relatively
wide image area with respect to the actual moving ob-
ject. Therefore, depth map artifacts in the correspond-
ing areas are not interpolated in the filtering stage and
thus affect synthesis.
REFERENCES
Belaifa, O., Skupin, R., Kurutepe, E., and Sikora, T. (2012).
Resampling of Multiple Camera Point Cloud Data. In
Consumer Communications and Networking Confer-
ence (CCNC), 2012 IEEE, pages 15–19. IEEE.
Dodgson, N. A. (2005). Autostereoscopic 3D Displays.
Computer, 38(8):31–36.
Farneb¨ack, G. (2003). Two-Frame Motion Estimation
Based On Polynomial Expansion. In Image Analysis,
pages 363–370. Springer.
Hastie, T., Tibshirani, R., and Friedman, J. J. H. (2001). The
Elements of Statistical Learning, volume 1. Springer
New York.
Merkle, P., Morvan, Y., Smolic, A., et al. (2009). The
Effects of Multiview Depth Video Compression On
Multiview Rendering. Signal Processing: Image
Communication, 24(1-2):73–88.
Merrell, P., Akbarzadeh, A., Wang, L., Mordohai, P.,
Frahm, J.-M., Yang, R., Nister, D., and Pollefeys, M.
(2007). Real-Time Visibility-Based Fusion of Depth
Maps. Computer Vision, IEEE International Confer-
ence on, 0:1–8.
Scharstein, D. and Szeliski, R. (2002). A Taxonomy and
Evaluation of Dense Two-Frame Stereo Correspon-
dence Algorithms. International journal of computer
vision, 47(1):7–42.
Seitz, S., Curless, B., Diebel, J., Scharstein, D., and
Szeliski, R. (2006). A Comparison and Evaluation of
Multi-View Stereo Reconstruction Algorithms.
Smolic, A., Mueller, K., Merkle, P., Fehn, C., Kauff,
P., Eisert, P., and Wiegand, T. (2006). 3d Video
and Free Viewpoint Video-Technologies, Applica-
tions and Mpeg Standards. In Multimedia and Expo,
2006 IEEE International Conference on, pages 2161–
2164. IEEE.
Starck, J. and Hilton, A. (2005). Virtual View Synthesis of
People From Multiple View Video Sequences. Graph-
ical Models, 67(6):600–620.
MultiviewPointCloudFilteringforSpatiotemporalConsistency
537