AnySplat: Feed-forward 3D Gaussian Splatting from Unconstrained Views
Abstract
AnySplat is a feed forward network that performs novel view synthesis without camera poses, using 3D Gaussian primitives and unified design for efficiency and quality across sparse and dense datasets.
We introduce AnySplat, a feed forward network for novel view synthesis from uncalibrated image collections. In contrast to traditional neural rendering pipelines that demand known camera poses and per scene optimization, or recent feed forward methods that buckle under the computational weight of dense views, our model predicts everything in one shot. A single forward pass yields a set of 3D Gaussian primitives encoding both scene geometry and appearance, and the corresponding camera intrinsics and extrinsics for each input image. This unified design scales effortlessly to casually captured, multi view datasets without any pose annotations. In extensive zero shot evaluations, AnySplat matches the quality of pose aware baselines in both sparse and dense view scenarios while surpassing existing pose free approaches. Moreover, it greatly reduce rendering latency compared to optimization based neural fields, bringing real time novel view synthesis within reach for unconstrained capture settings.Project page: https://city-super.github.io/anysplat/
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Diffusion-Guided Gaussian Splatting for Large-Scale Unconstrained 3D Reconstruction and Novel View Synthesis (2025)
- SparSplat: Fast Multi-View Reconstruction with Generalizable 2D Gaussian Splatting (2025)
- Styl3R: Instant 3D Stylized Reconstruction for Arbitrary Scenes and Styles (2025)
- Sparfels: Fast Reconstruction from Sparse Unposed Imagery (2025)
- MonoSplat: Generalizable 3D Gaussian Splatting from Monocular Depth Foundation Models (2025)
- Recollection from Pensieve: Novel View Synthesis via Learning from Uncalibrated Videos (2025)
- MV-CoLight: Efficient Object Compositing with Consistent Lighting and Shadow Generation (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper