ZPressor: Bottleneck-Aware Compression for Scalable Feed-Forward 3DGS
Abstract
ZPressor, a lightweight module, compresses multi-view inputs for feed-forward 3D Gaussian Splatting models, enhancing their scalability and performance under dense view settings.
Feed-forward 3D Gaussian Splatting (3DGS) models have recently emerged as a promising solution for novel view synthesis, enabling one-pass inference without the need for per-scene 3DGS optimization. However, their scalability is fundamentally constrained by the limited capacity of their encoders, leading to degraded performance or excessive memory consumption as the number of input views increases. In this work, we analyze feed-forward 3DGS frameworks through the lens of the Information Bottleneck principle and introduce ZPressor, a lightweight architecture-agnostic module that enables efficient compression of multi-view inputs into a compact latent state Z that retains essential scene information while discarding redundancy. Concretely, ZPressor enables existing feed-forward 3DGS models to scale to over 100 input views at 480P resolution on an 80GB GPU, by partitioning the views into anchor and support sets and using cross attention to compress the information from the support views into anchor views, forming the compressed latent state Z. We show that integrating ZPressor into several state-of-the-art feed-forward 3DGS models consistently improves performance under moderate input views and enhances robustness under dense view settings on two large-scale benchmarks DL3DV-10K and RealEstate10K. The video results, code and trained models are available on our project page: https://lhmd.top/zpressor.
Community
ZPressor is a plug-and-play module that compresses multi-view inputs for scalable feed-forward 3DGS.
Project Page: https://lhmd.top/zpressor
Code: https://github.com/ziplab/ZPressor
Existing feed-forward 3DGS models struggle with dense views, facing performance drops & massive redundancy. ZPressor leverages Information Bottleneck Theory to compress multi-view features, significantly boosting scalability and reconstruction quality for robust dense-view synthesis.
Plug & play, lightweight, and powerful.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- AnySplat: Feed-forward 3D Gaussian Splatting from Unconstrained Views (2025)
- MV-CoLight: Efficient Object Compositing with Consistent Lighting and Shadow Generation (2025)
- SpatialSplat: Efficient Semantic 3D from Sparse Unposed Images (2025)
- SHaDe: Compact and Consistent Dynamic 3D Reconstruction via Tri-Plane Deformation and Latent Diffusion (2025)
- MonoSplat: Generalizable 3D Gaussian Splatting from Monocular Depth Foundation Models (2025)
- 3DGS Compression with Sparsity-guided Hierarchical Transform Coding (2025)
- VideoScene: Distilling Video Diffusion Model to Generate 3D Scenes in One Step (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
is this compatible with deformable beta splatting ? https://github.com/RongLiu-Leo/beta-splatting
Our current model is primarily designed for feed-forward 3D reconstruction models. Therefore, it is not directly applicable to optimization-based, single-scene training 3D reconstruction methods.
Theoretically, our model exhibits compatibility with any feed-forward 3D reconstruction approach, irrespective of its underlying 3D representation. This is because our method operates at the information compression level, rather than performing explicit compression operations on the 3D representation itself (e.g., Gaussian Pruning/Merging techniques).
Should researchers extend Beta Splatting into a feed-forward model in the future, we would be very interested to explore the application of our method within such a framework.
Models citing this paper 1
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper