
ACKNOWLEDGEMENTS
This work was co-funded by the European Union
under Horizon Europe, grant number 101092889,
project SHARESPACE. We thank Ren
´
e Schuster
for constructive discussions and feedback on earlier
drafts of this paper.
REFERENCES
Barron, J. T., Mildenhall, B., Tancik, M., Hedman, P.,
Martin-Brualla, R., and Srinivasan, P. P. (2021). Mip-
nerf: A multiscale representation for anti-aliasing
neural radiance fields. ICCV.
Barron, J. T., Mildenhall, B., Verbin, D., Srinivasan, P. P.,
and Hedman, P. (2022). Mip-nerf 360: Unbounded
anti-aliased neural radiance fields. CVPR.
Gao, J., Gu, C., Lin, Y., Zhu, H., Cao, X., Zhang, L., and
Yao, Y. (2023). Relightable 3d gaussian: Real-time
point cloud relighting with brdf decomposition and
ray tracing. arXiv:2311.16043.
Gu
´
edon, A. and Lepetit, V. (2024). Sugar: Surface-aligned
gaussian splatting for efficient 3d mesh reconstruction
and high-quality mesh rendering. CVPR.
Hedman, P., Srinivasan, P. P., Mildenhall, B., Barron, J. T.,
and Debevec, P. (2021). Baking neural radiance fields
for real-time view synthesis. ICCV.
Huang, B., Yu, Z., Chen, A., Geiger, A., and Gao, S. (2024).
2d gaussian splatting for geometrically accurate radi-
ance fields. In SIGGRAPH 2024 Conference Papers.
Association for Computing Machinery.
Jensen, R., Dahl, A., Vogiatzis, G., Tola, E., and Aanæs,
H. (2014). Large scale multi-view stereopsis evalua-
tion. In 2014 IEEE Conference on Computer Vision
and Pattern Recognition, pages 406–413. IEEE.
Kerbl, B., Kopanas, G., Leimk
¨
uhler, T., and Drettakis,
G. (2023). 3d gaussian splatting for real-time radi-
ance field rendering. ACM Transactions on Graphics,
42(4).
Mildenhall, B., Srinivasan, P. P., Tancik, M., Barron, J. T.,
Ramamoorthi, R., and Ng, R. (2020). Nerf: Repre-
senting scenes as neural radiance fields for view syn-
thesis. In ECCV.
Ravi, N., Gabeur, V., Hu, Y.-T., Hu, R., Ryali, C., Ma,
T., Khedr, H., R
¨
adle, R., Rolland, C., Gustafson, L.,
Mintun, E., Pan, J., Alwala, K. V., Carion, N., Wu,
C.-Y., Girshick, R., Doll
´
ar, P., and Feichtenhofer, C.
(2024). Sam 2: Segment anything in images and
videos. arXiv preprint arXiv:2408.00714.
Reiser, C., Szeliski, R., Verbin, D., Srinivasan, P. P.,
Mildenhall, B., Geiger, A., Barron, J. T., and Hedman,
P. (2023). Merf: Memory-efficient radiance fields for
real-time view synthesis in unbounded scenes. SIG-
GRAPH.
Sch
¨
onberger, J. L. and Frahm, J.-M. (2016). Structure-
from-motion revisited. In Conference on Computer
Vision and Pattern Recognition (CVPR).
Sch
¨
onberger, J. L., Zheng, E., Pollefeys, M., and Frahm, J.-
M. (2016). Pixelwise view selection for unstructured
multi-view stereo. In European Conference on Com-
puter Vision (ECCV).
Wang, P., Liu, L., Liu, Y., Theobalt, C., Komura, T., and
Wang, W. (2021). Neus: Learning neural implicit sur-
faces by volume rendering for multi-view reconstruc-
tion. NeurIPS.
Xie, T., Zong, Z., Qiu, Y., Li, X., Feng, Y., Yang, Y., and
Jiang, C. (2023). Physgaussian: Physics-integrated
3d gaussians for generative dynamics. arXiv preprint
arXiv:2311.12198.
Yang, C., Li, S., Fang, J., Liang, R., Xie, L., Zhang, X.,
Shen, W., and Tian, Q. (2024). Gaussianobject: High-
quality 3d object reconstruction from four views with
gaussian splatting. ACM Transactions on Graphics,
43(6).
Yariv, L., Gu, J., Kasten, Y., and Lipman, Y. (2021). Vol-
ume rendering of neural implicit surfaces. In Thirty-
Fifth Conference on Neural Information Processing
Systems.
Yariv, L., Hedman, P., Reiser, C., Verbin, D., Srinivasan,
P. P., Szeliski, R., Barron, J. T., and Mildenhall, B.
(2023). Bakedsdf: Meshing neural sdfs for real-time
view synthesis. arXiv.
Zhou, Q.-Y., Park, J., and Koltun, V. (2018). Open3D:
A modern library for 3D data processing.
arXiv:1801.09847.
APPENDIX
Downstream Applications
Our method produces an isolated representation of a
target object from the scene. Whether using Gaus-
sians or mesh, the representation can be directly used
without any need for additional processing. This en-
ables quick and easy use for downstream applications,
such as appearance editing and physics simulations.
An example of appearance editing is shown in Fig-
ure 9.
Pruning in 3D Gaussian Splatting
We demonstrate the versatility of our pruning strat-
egy by implementing it in 3DGS. We perform a sin-
gle experiment on the Bicycle scene from the Mip-
NeRF360 dataset as proof of principle. The results
are shown in Table 8. Similar to the results for 2DGS,
our pruning strategy effectively reduces the number
of occluded Gaussians while preserving quality. This
reduces memory requirements and positively impacts
the training time. However, we notice that the num-
ber of occluded Gaussians is lower compared to the
2DGS scenes. This is likely due to 2DGS encourag-
ing surfaces that are fully opaque. The result is a clear
Object-Centric 2D Gaussian Splatting: Background Removal and Occlusion-Aware Pruning for Compact Object Models
529