Recognition: 2 theorem links
· Lean TheoremAnchorSplat: Feed-Forward 3D Gaussian Splatting with 3D Geometric Priors
Pith reviewed 2026-05-10 18:41 UTC · model grok-4.3
The pith
AnchorSplat uses 3D geometric priors to anchor Gaussians directly in 3D space, decoupling them from 2D pixels for efficient scene reconstruction.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
The paper establishes that representing 3D scenes with anchor-aligned Gaussians guided by 3D geometric priors allows direct 3D-space modeling in a feed-forward manner, reducing the number of Gaussians needed while enhancing reconstruction fidelity and view consistency compared to pixel-aligned approaches.
What carries the argument
The anchor-aligned Gaussian representation, which uses 3D priors to determine Gaussian positions and attributes independently of 2D image features.
If this is right
- Substantially fewer Gaussian primitives are required for high-quality scene representation.
- Reconstruction becomes independent of the resolution and number of input views.
- View consistency in novel view synthesis improves due to the 3D-centric design.
- Computational efficiency increases from the reduced primitive count.
Where Pith is reading between the lines
- This approach could be combined with existing 3D scanning hardware that outputs point clouds or voxels as priors.
- Future work might explore using learned priors when explicit 3D data is unavailable.
- Similar anchoring ideas could apply to other primitive-based rendering methods like surfels.
Load-bearing premise
The input must include reliable 3D geometric priors that capture enough scene structure to guide accurate Gaussian placement.
What would settle it
Running the model on input where the 3D priors are removed or replaced with random points and verifying whether it loses its performance edge over pixel-aligned baselines on the same benchmark.
Figures
read the original abstract
Recent feed-forward Gaussian reconstruction models adopt a pixel-aligned formulation that maps each 2D pixel to a 3D Gaussian, entangling Gaussian representations tightly with the input images. In this paper, we propose AnchorSplat, a novel feed-forward 3DGS framework for scene-level reconstruction that represents the scene directly in 3D space. AnchorSplat introduces an anchor-aligned Gaussian representation guided by 3D geometric priors (e.g., sparse point clouds, voxels, or RGB-D point clouds), enabling a more geometry-aware renderable 3D Gaussians that is independent of image resolution and number of views. This design substantially reduces the number of required Gaussians, improving computational efficiency while enhancing reconstruction fidelity. Beyond the anchor-aligned design, we utilize a Gaussian Refiner to adjust the intermediate Gaussiansy via merely a few forward passes. Experiments on the ScanNet++ v2 NVS benchmark demonstrate the SOTA performance, outperforming previous methods with more view-consistent and substantially fewer Gaussian primitives.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The manuscript proposes AnchorSplat, a feed-forward 3D Gaussian Splatting framework for scene-level reconstruction. Unlike prior pixel-aligned methods that map 2D pixels directly to 3D Gaussians, it introduces an anchor-aligned Gaussian representation conditioned on explicit 3D geometric priors (sparse point clouds, voxels, or RGB-D clouds). This is claimed to produce resolution- and view-independent renderable Gaussians, reduce the total number of primitives, and improve view consistency. A Gaussian Refiner module is added to adjust intermediate Gaussians with a small number of forward passes. The central empirical claim is state-of-the-art performance on the ScanNet++ v2 novel-view-synthesis benchmark, outperforming previous methods in consistency and efficiency.
Significance. If the empirical claims hold after proper controls, the work could be significant for efficient feed-forward 3D reconstruction pipelines. Decoupling Gaussian placement from image resolution via 3D anchors and reducing primitive count address practical bottlenecks in 3DGS. The refiner idea is a lightweight post-processing step that might generalize. However, the significance is conditional on the availability of 3D priors and on whether gains are attributable to the anchor design rather than the richer input modality.
major comments (3)
- [Experiments] Experiments section: The SOTA claim on ScanNet++ v2 NVS (better view consistency, substantially fewer Gaussians) is central but unsupported by any reported quantitative metrics, baseline tables, or ablation studies in the manuscript. Without these, it is impossible to verify the claim or to isolate the contribution of the anchor-aligned design from the use of 3D geometric priors as input.
- [Method] Method section: The anchor-aligned Gaussian representation is presented as guided by 3D priors and independent of image resolution, yet no equations, pseudocode, or algorithmic details are supplied for anchor placement, Gaussian alignment to anchors, or the exact conditioning mechanism. This omission is load-bearing for assessing novelty and correctness relative to prior 3DGS formulations.
- [§3.2] §3.2 (Gaussian Refiner): The statement that the refiner adjusts intermediate Gaussians “via merely a few forward passes” is used to support efficiency and consistency claims, but the architecture, training objective, input format, and number of passes are not specified. Without these details the efficiency advantage cannot be evaluated.
minor comments (3)
- [Abstract] Abstract: Typo “intermediate Gaussiansy” should read “intermediate Gaussians.”
- [Introduction] Introduction and related work: The distinction from prior pixel-aligned feed-forward models would be strengthened by citing specific representative works and clearly stating their input assumptions.
- [Figures] Figures: Qualitative results should include side-by-side comparisons against baselines to illustrate the claimed improvements in view consistency and primitive count.
Simulated Author's Rebuttal
We thank the referee for the constructive feedback. We address each major comment below and will make substantial revisions to the manuscript to provide the requested details, metrics, and clarifications.
read point-by-point responses
-
Referee: [Experiments] Experiments section: The SOTA claim on ScanNet++ v2 NVS (better view consistency, substantially fewer Gaussians) is central but unsupported by any reported quantitative metrics, baseline tables, or ablation studies in the manuscript. Without these, it is impossible to verify the claim or to isolate the contribution of the anchor-aligned design from the use of 3D geometric priors as input.
Authors: We acknowledge that the current version of the manuscript presents the SOTA claim at a high level without including the supporting quantitative tables, specific metrics, or ablations. In the revised manuscript we will add comprehensive results on ScanNet++ v2, including PSNR/SSIM/LPIPS scores against relevant baselines, Gaussian primitive counts, and dedicated ablations that separate the anchor-aligned representation from the 3D prior input. These additions will allow direct verification of the claims and isolation of the design contributions. revision: yes
-
Referee: [Method] Method section: The anchor-aligned Gaussian representation is presented as guided by 3D priors and independent of image resolution, yet no equations, pseudocode, or algorithmic details are supplied for anchor placement, Gaussian alignment to anchors, or the exact conditioning mechanism. This omission is load-bearing for assessing novelty and correctness relative to prior 3DGS formulations.
Authors: We agree that the method description would be strengthened by explicit formalization. The revised manuscript will include the mathematical formulation for generating anchors from the 3D geometric priors, the alignment procedure that maps Gaussians to anchors, and the conditioning mechanism that makes the representation resolution-independent. We will also add pseudocode for the overall feed-forward pipeline to clarify the differences from pixel-aligned approaches. revision: yes
-
Referee: [§3.2] §3.2 (Gaussian Refiner): The statement that the refiner adjusts intermediate Gaussians “via merely a few forward passes” is used to support efficiency and consistency claims, but the architecture, training objective, input format, and number of passes are not specified. Without these details the efficiency advantage cannot be evaluated.
Authors: We will expand §3.2 with the missing specifications: the network architecture of the refiner, the training objective (including loss terms), the precise input format for intermediate Gaussians, and the number of forward passes (typically 2–3) used at inference time. These details will substantiate the efficiency and consistency benefits claimed for the refiner module. revision: yes
Circularity Check
No circularity: method and claims are empirically grounded without self-referential reductions
full rationale
The paper defines AnchorSplat via an anchor-aligned Gaussian representation that takes 3D geometric priors as explicit input and reports SOTA empirical results on ScanNet++ v2 NVS. No equations, fitted parameters, or predictions are described that reduce by construction to the inputs (e.g., no self-definitional scaling, no 'prediction' that is a refit of the same data, no uniqueness theorem imported from self-citation). The central design choice and performance claims remain independent of the listed circularity patterns and are presented as falsifiable experimental outcomes.
Axiom & Free-Parameter Ledger
axioms (1)
- domain assumption 3D geometric priors such as sparse point clouds, voxels, or RGB-D point clouds are available and sufficiently accurate to guide anchor placement and Gaussian representation.
invented entities (2)
-
Anchor-aligned Gaussian representation
no independent evidence
-
Gaussian Refiner
no independent evidence
Lean theorems connected to this paper
-
IndisputableMonolith/Foundation/AbsoluteFloorClosure.leanreality_from_one_distinction unclear?
unclearRelation between the paper passage and the cited Recognition theorem.
AnchorSplat introduces an anchor-aligned Gaussian representation guided by 3D geometric priors (e.g., sparse point clouds, voxels, or RGB-D point clouds), enabling a more geometry-aware renderable 3D Gaussians that is independent of image resolution and number of views.
-
IndisputableMonolith/Cost/FunctionalEquation.leanwashburn_uniqueness_aczel unclear?
unclearRelation between the paper passage and the cited Recognition theorem.
We leverage the pretrained MapAnything [18] ... to predict depths and camera poses ... downsampled into a sparser set of anchors by using farthest point sampling (FPS) algorithm
What do these tags mean?
- matches
- The paper's claim is directly supported by a theorem in the formal canon.
- supports
- The theorem supports part of the paper's argument, but the paper may add assumptions or extra steps.
- extends
- The paper goes beyond the formal theorem; the theorem is a base layer rather than the whole result.
- uses
- The paper appears to rely on the theorem as machinery.
- contradicts
- The paper's claim conflicts with a theorem or certificate in the canon.
- unclear
- Pith found a possible connection, but the passage is too broad, indirect, or ambiguous to say the theorem truly supports the claim.
Reference graph
Works this paper leans on
-
[1]
Accurate 3-d reconstruction under iot environ- ments and its applications to augmented reality.IEEE Trans- actions on Industrial Informatics, 17(3):2090–2100, 2020
Mingwei Cao, Liping Zheng, Wei Jia, Huimin Lu, and Xi- aoping Liu. Accurate 3-d reconstruction under iot environ- ments and its applications to augmented reality.IEEE Trans- actions on Industrial Informatics, 17(3):2090–2100, 2020. 1
2090
-
[2]
pixelsplat: 3d gaussian splats from image pairs for scalable generalizable 3d reconstruction
David Charatan, Sizhe Lester Li, Andrea Tagliasacchi, and Vincent Sitzmann. pixelsplat: 3d gaussian splats from image pairs for scalable generalizable 3d reconstruction. InPro- ceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 19457–19467, 2024. 1, 3
2024
-
[3]
A survey on 3d gaussian splatting, 2025
Guikun Chen and Wenguan Wang. A survey on 3d gaussian splatting, 2025. 1
2025
-
[4]
Quantifying and alleviating co-adaptation in sparse-view 3d gaussian splatting,
Kangjie Chen, Yingji Zhong, Zhihao Li, Jiaqi Lin, Youyu Chen, Minghan Qin, and Haoqian Wang. Quantifying and alleviating co-adaptation in sparse-view 3d gaussian splat- ting.arXiv preprint arXiv:2508.12720, 2025. 2
-
[5]
Splat-nav: Safe real-time robot navigation in gaussian splatting maps.IEEE Transactions on Robotics,
Timothy Chen, Ola Shorinwa, Joseph Bruno, Aiden Swann, Javier Yu, Weijia Zeng, Keiko Nagami, Philip Dames, and Mac Schwager. Splat-nav: Safe real-time robot navigation in gaussian splatting maps.IEEE Transactions on Robotics,
-
[6]
G3r: Gradient guided gen- eralizable reconstruction
Yun Chen, Jingkang Wang, Ze Yang, Sivabalan Mani- vasagam, and Raquel Urtasun. G3r: Gradient guided gen- eralizable reconstruction. InEuropean Conference on Com- puter Vision, pages 305–323. Springer, 2024. 4
2024
-
[7]
Mvsplat: Efficient 3d gaussian splatting from sparse multi-view images
Yuedong Chen, Haofei Xu, Chuanxia Zheng, Bohan Zhuang, Marc Pollefeys, Andreas Geiger, Tat-Jen Cham, and Jianfei Cai. Mvsplat: Efficient 3d gaussian splatting from sparse multi-view images. InEuropean Conference on Computer Vision, pages 370–386. Springer, 2024. 1, 3
2024
-
[8]
Dashgaussian: Optimizing 3d gaussian splatting in 200 seconds
Youyu Chen, Junjun Jiang, Kui Jiang, Xiao Tang, Zhihao Li, Xianming Liu, and Yinyu Nie. Dashgaussian: Optimizing 3d gaussian splatting in 200 seconds. InCVPR, 2025. 2
2025
-
[9]
Youyu Chen, Junjun Jiang, Yueru Luo, Kui Jiang, Xianming Liu, Xu Yan, and Dave Zhenyu Chen. Reliev3r: Relieving feed-forward reconstruction from multi-view geometric an- notations.arXiv preprint arXiv:2604.00548, 2026. 3
-
[10]
Nerf: Neural radiance field in 3d vision, a comprehensive review,
Kyle Gao, Yina Gao, Hongjie He, Dening Lu, Linlin Xu, and Jonathan Li. Nerf: Neural radiance field in 3d vision, a comprehensive review.arXiv preprint arXiv:2210.00379,
-
[11]
Scene reconstruction with functional objects for robot autonomy.International Journal of Computer Vision, 130(12):2940–2961, 2022
Muzhi Han, Zeyu Zhang, Ziyuan Jiao, Xu Xie, Yixin Zhu, Song-Chun Zhu, and Hangxin Liu. Scene reconstruction with functional objects for robot autonomy.International Journal of Computer Vision, 130(12):2940–2961, 2022. 1
2022
-
[12]
Deep residual learning for image recognition
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. InProceed- ings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016. 4
2016
-
[13]
Kohei Honda, Takeshi Ishita, Yasuhiro Yoshimura, and Ryo Yonetani. Gsplatvnm: Point-of-view synthesis for visual navigation models using gaussian splatting.arXiv preprint arXiv:2503.05152, 2025. 2
-
[14]
LRM: Large Reconstruction Model for Single Image to 3D
Yicong Hong, Kai Zhang, Jiuxiang Gu, Sai Bi, Yang Zhou, Difan Liu, Feng Liu, Kalyan Sunkavalli, Trung Bui, and Hao Tan. Lrm: Large reconstruction model for single image to 3d.arXiv preprint arXiv:2311.04400, 2023. 2
work page internal anchor Pith review arXiv 2023
-
[15]
Mvsanywhere: Zero-shot multi-view stereo
Sergio Izquierdo, Mohamed Sayed, Michael Firman, Guillermo Garcia-Hernando, Daniyar Turmukhambetov, Javier Civera, Oisin Mac Aodha, Gabriel Brostow, and Jamie Watson. Mvsanywhere: Zero-shot multi-view stereo. In Proceedings of the Computer Vision and Pattern Recognition Conference, pages 11493–11504, 2025. 3
2025
-
[16]
Lihan Jiang, Yucheng Mao, Linning Xu, Tao Lu, Kerui Ren, Yichen Jin, Xudong Xu, Mulin Yu, Jiangmiao Pang, Feng Zhao, et al. Anysplat: Feed-forward 3d gaussian splatting from unconstrained views.arXiv preprint arXiv:2505.23716,
-
[17]
ilrm: An iterative large 3d reconstruction model.arXiv preprint arXiv:2507.23277, 2025
Gyeongjin Kang, Seungtae Nam, Seungkwon Yang, Xi- angyu Sun, Sameh Khamis, Abdelrahman Mohamed, and Eunbyung Park. ilrm: An iterative large 3d reconstruction model.arXiv preprint arXiv:2507.23277, 2025. 2
-
[18]
MapAnything: Universal Feed-Forward Metric 3D Reconstruction
Nikhil Keetha, Norman M ¨uller, Johannes Sch ¨onberger, Lorenzo Porzi, Yuchen Zhang, Tobias Fischer, Arno Knapitsch, Duncan Zauss, Ethan Weber, Nelson Antunes, et al. Mapanything: Universal feed-forward metric 3d re- construction.arXiv preprint arXiv:2509.13414, 2025. 2, 3
work page internal anchor Pith review arXiv 2025
-
[19]
3d gaussian splatting for real-time radiance field rendering.ACM Trans
Bernhard Kerbl, Georgios Kopanas, Thomas Leimk ¨uhler, and George Drettakis. 3d gaussian splatting for real-time radiance field rendering.ACM Trans. Graph., 42(4):139–1,
-
[20]
Gaussnav: Gaussian splatting for visual navigation.IEEE Transactions on Pattern Analysis and Machine Intelligence,
Xiaohan Lei, Min Wang, Wengang Zhou, and Houqiang Li. Gaussnav: Gaussian splatting for visual navigation.IEEE Transactions on Pattern Analysis and Machine Intelligence,
-
[21]
Ascend: a scalable and uni- fied architecture for ubiquitous deep neural network com- puting: Industry track paper
Heng Liao, Jiajin Tu, Jing Xia, Hu Liu, Xiping Zhou, Honghui Yuan, and Yuxing Hu. Ascend: a scalable and uni- fied architecture for ubiquitous deep neural network com- puting: Industry track paper. In2021 IEEE International Symposium on High-Performance Computer Architecture (HPCA), pages 789–801. IEEE, 2021. 5
2021
-
[22]
Vastgaussian: Vast 3d gaussians for large scene reconstruction
Jiaqi Lin, Zhihao Li, Xiao Tang, Jianzhuang Liu, Shiyong Liu, Jiayue Liu, Yangdi Lu, Xiaofei Wu, Songcen Xu, You- liang Yan, et al. Vastgaussian: Vast 3d gaussians for large scene reconstruction. InCVPR, 2024. 2
2024
-
[23]
Mvsgaussian: Fast generalizable gaussian splatting recon- struction from multi-view stereo
Tianqi Liu, Guangcong Wang, Shoukang Hu, Liao Shen, Xinyi Ye, Yuhang Zang, Zhiguo Cao, Wei Li, and Ziwei Liu. Mvsgaussian: Fast generalizable gaussian splatting recon- struction from multi-view stereo. InEuropean Conference on Computer Vision, pages 37–53. Springer, 2024. 3
2024
-
[24]
Yifan Liu, Zhiyuan Min, Zhenwei Wang, Junta Wu, Tengfei Wang, Yixuan Yuan, Yawei Luo, and Chunchao Guo. World- mirror: Universal 3d world reconstruction with any-prior prompting.arXiv preprint arXiv:2510.10726, 2025. 2
-
[25]
Rea- songrounder: Lvlm-guided hierarchical feature splatting for open-vocabulary 3d visual grounding and reasoning
Zhenyang Liu, Yikai Wang, Sixiao Zheng, Tongying Pan, Longfei Liang, Yanwei Fu, and Xiangyang Xue. Rea- songrounder: Lvlm-guided hierarchical feature splatting for open-vocabulary 3d visual grounding and reasoning. InPro- ceedings of the Computer Vision and Pattern Recognition Conference, pages 3718–3727, 2025. 2
2025
-
[26]
Decoupled Weight Decay Regularization
Ilya Loshchilov, Frank Hutter, et al. Fixing weight decay regularization in adam.arXiv preprint arXiv:1711.05101, 5 (5):5, 2017. 5
work page internal anchor Pith review Pith/arXiv arXiv 2017
-
[27]
Saswat Subhajyoti Mallick, Rahul Goel, Bernhard Kerbl, Francisco Vicente Carrasco, Markus Steinberger, and Fer- nando De La Torre. Taming 3dgs: High-quality ra- diance fields with limited resources.arXiv preprint arXiv:2406.15643, 2024. 2
-
[28]
Srinivasan, Matthew Tancik, Jonathan T
Ben Mildenhall, Pratul P. Srinivasan, Matthew Tancik, Jonathan T. Barron, Ravi Ramamoorthi, and Ren Ng. Nerf: Representing scenes as neural radiance fields for view syn- thesis. InProceedings of the European Conference on Com- puter Vision (ECCV), 2020. 1, 2
2020
-
[29]
Reg- nerf: Regularizing neural radiance fields for view synthesis from sparse inputs
Michael Niemeyer, Jonathan T Barron, Ben Mildenhall, Mehdi SM Sajjadi, Andreas Geiger, and Noha Radwan. Reg- nerf: Regularizing neural radiance fields for view synthesis from sparse inputs. InProceedings of the IEEE/CVF con- ference on computer vision and pattern recognition, pages 5480–5490, 2022. 2
2022
-
[30]
Slarm: Streaming and language-aligned reconstruction model for dynamic scenes
Zhicheng Qiu, Jiarui Meng, Tong-an Luo, Yican Huang, Xuan Feng, Xuanfu Li, and ZHan Xu. Slarm: Streaming and language-aligned reconstruction model for dynamic scenes. arXiv preprint arXiv:2603.22893, 2026. 2
-
[31]
Schonberger and Jan-Michael Frahm
Johannes L. Schonberger and Jan-Michael Frahm. Structure- from-motion revisited. InProceedings of the IEEE Confer- ence on Computer Vision and Pattern Recognition (CVPR),
-
[32]
Language embedded 3d gaussians for open- vocabulary scene understanding
Jin-Chuan Shi, Miao Wang, Hao-Bin Duan, and Shao- Hua Guan. Language embedded 3d gaussians for open- vocabulary scene understanding. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5333–5343, 2024. 2
2024
-
[33]
Rupprecht, and Andrea Vedaldi
Stanislaw Szymanowicz, C. Rupprecht, and Andrea Vedaldi. Splatter image: Ultra-fast single-view 3d reconstruction. 2024 IEEE/CVF Conference on Computer Vision and Pat- tern Recognition (CVPR), pages 10208–10217, 2023. 3
2024
-
[34]
Vggt: Vi- sual geometry grounded transformer
Jianyuan Wang, Minghao Chen, Nikita Karaev, Andrea Vedaldi, Christian Rupprecht, and David Novotny. Vggt: Vi- sual geometry grounded transformer. InProceedings of the Computer Vision and Pattern Recognition Conference, pages 5294–5306, 2025. 2
2025
-
[35]
Continuous 3d per- ception model with persistent state
Qianqian Wang, Yifei Zhang, Aleksander Holynski, Alexei A Efros, and Angjoo Kanazawa. Continuous 3d per- ception model with persistent state. InProceedings of the Computer Vision and Pattern Recognition Conference, pages 10510–10522, 2025. 2
2025
-
[36]
Dust3r: Geometric 3d vi- sion made easy
Shuzhe Wang, Vincent Leroy, Yohann Cabon, Boris Chidlovskii, and Jerome Revaud. Dust3r: Geometric 3d vi- sion made easy. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 20697– 20709, 2024. 2
2024
-
[37]
Weijie Wang, Yeqing Chen, Zeyu Zhang, Hengyu Liu, Haoxiao Wang, Zhiyuan Feng, Wenkang Qin, Zheng Zhu, Donny Y Chen, and Bohan Zhuang. V olsplat: Rethinking feed-forward 3d gaussian splatting with voxel-aligned pre- diction.arXiv preprint arXiv:2509.19297, 2025. 2
-
[38]
Freesplat: Generalizable 3d gaussian splatting towards free view synthesis of indoor scenes.Advances in Neural Information Processing Systems, 37:107326–107349, 2024
Yunsong Wang, Tianxin Huang, Hanlin Chen, and Gim Hee Lee. Freesplat: Generalizable 3d gaussian splatting towards free view synthesis of indoor scenes.Advances in Neural Information Processing Systems, 37:107326–107349, 2024. 3
2024
-
[39]
Heightfields for efficient scene reconstruction for ar
Jamie Watson, Sara Vicente, Oisin Mac Aodha, Cl´ement Go- dard, Gabriel Brostow, and Michael Firman. Heightfields for efficient scene reconstruction for ar. InProceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pages 5850–5860, 2023. 1
2023
-
[40]
latentsplat: Autoencoding variational gaussians for fast generalizable 3d reconstruction
Christopher Wewer, Kevin Raj, Eddy Ilg, Bernt Schiele, and Jan Eric Lenssen. latentsplat: Autoencoding variational gaussians for fast generalizable 3d reconstruction. InECCV 2024 Workshop on Wild 3D: 3D Modeling, Reconstruction, and Generation in the Wild, 2024. 3
2024
-
[41]
Point transformer v3: Simpler faster stronger
Xiaoyang Wu, Li Jiang, Peng-Shuai Wang, Zhijian Liu, Xi- hui Liu, Yu Qiao, Wanli Ouyang, Tong He, and Hengshuang Zhao. Point transformer v3: Simpler faster stronger. InPro- ceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 4840–4851, 2024. 4
2024
-
[42]
Resplat: Learning recurrent gaussian splats.arXiv preprint arXiv:2510.08575, 2025
Haofei Xu, Daniel Barath, Andreas Geiger, and Marc Polle- feys. Resplat: Learning recurrent gaussian splats.arXiv preprint arXiv:2510.08575, 2025. 3
-
[43]
Depthsplat: Connecting gaussian splatting and depth
Haofei Xu, Songyou Peng, Fangjinhua Wang, Hermann Blum, Daniel Barath, Andreas Geiger, and Marc Pollefeys. Depthsplat: Connecting gaussian splatting and depth. In Proceedings of the Computer Vision and Pattern Recognition Conference, pages 16453–16463, 2025. 1, 3
2025
-
[44]
Freenerf: Im- proving few-shot neural rendering with free frequency reg- ularization
Jiawei Yang, Marco Pavone, and Yue Wang. Freenerf: Im- proving few-shot neural rendering with free frequency reg- ularization. InProceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 8254–8263,
-
[45]
Fast3r: Towards 3d reconstruction of 1000+ images in one forward pass
Jianing Yang, Alexander Sax, Kevin J Liang, Mikael Henaff, Hao Tang, Ang Cao, Joyce Chai, Franziska Meier, and Matt Feiszli. Fast3r: Towards 3d reconstruction of 1000+ images in one forward pass. InProceedings of the Computer Vision and Pattern Recognition Conference, pages 21924–21935,
-
[46]
Botao Ye, Boqi Chen, Haofei Xu, Daniel Barath, and Marc Pollefeys. Yonosplat: You only need one model for feedforward 3d gaussian splatting.arXiv preprint arXiv:2511.07321, 2025. 2
-
[47]
Scannet++: A high-fidelity dataset of 3d in- door scenes
Chandan Yeshwanth, Yueh-Cheng Liu, Matthias Nießner, and Angela Dai. Scannet++: A high-fidelity dataset of 3d in- door scenes. InProceedings of the IEEE/CVF International Conference on Computer Vision, pages 12–22, 2023. 2, 5
2023
-
[48]
Mip-splatting: Alias-free 3d gaussian splat- ting
Zehao Yu, Anpei Chen, Binbin Huang, Torsten Sattler, and Andreas Geiger. Mip-splatting: Alias-free 3d gaussian splat- ting. InProceedings of the IEEE/CVF conference on com- puter vision and pattern recognition, pages 19447–19456,
-
[49]
Cor-gs: sparse-view 3d gaussian splatting via co-regularization
Jiawei Zhang, Jiahe Li, Xiaohan Yu, Lei Huang, Lin Gu, Jin Zheng, and Xiao Bai. Cor-gs: sparse-view 3d gaussian splatting via co-regularization. InECCV, 2024. 2
2024
-
[50]
Gs-lrm: Large recon- struction model for 3d gaussian splatting
Kai Zhang, Sai Bi, Hao Tan, Yuanbo Xiangli, Nanxuan Zhao, Kalyan Sunkavalli, and Zexiang Xu. Gs-lrm: Large recon- struction model for 3d gaussian splatting. InEuropean Con- ference on Computer Vision, pages 1–19. Springer, 2024. 2
2024
-
[51]
Part-level scene reconstruction affords robot interac- tion
Zeyu Zhang, Lexing Zhang, Zaijin Wang, Ziyuan Jiao, Muzhi Han, Yixin Zhu, Song-Chun Zhu, and Hangxin Liu. Part-level scene reconstruction affords robot interac- tion. In2023 IEEE/RSJ International Conference on Intelli- gent Robots and Systems (IROS), pages 11178–11185. IEEE,
-
[52]
Taming video diffusion prior with scene- grounding guidance for 3d gaussian splatting from sparse in- puts
Yingji Zhong, Zhihao Li, Dave Zhenyu Chen, Lanqing Hong, and Dan Xu. Taming video diffusion prior with scene- grounding guidance for 3d gaussian splatting from sparse in- puts. InCVPR, 2025. 2
2025
-
[53]
Hugs: Holistic urban 3d scene understanding via gaus- sian splatting
Hongyu Zhou, Jiahao Shao, Lu Xu, Dongfeng Bai, Weichao Qiu, Bingbing Liu, Yue Wang, Andreas Geiger, and Yiyi Liao. Hugs: Holistic urban 3d scene understanding via gaus- sian splatting. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 21336– 21345, 2024. 2
2024
-
[54]
Fsgs: Real-time few-shot view synthesis using gaussian splatting
Zehao Zhu, Zhiwen Fan, Yifan Jiang, and Zhangyang Wang. Fsgs: Real-time few-shot view synthesis using gaussian splatting. InECCV, 2024. 2 AnchorSplat: Feed-Forward 3D Gaussian Splatting with 3D Geometric Priors Supplementary Material A. Appendix This supplementary material provides the following addi- tional information: (B) additional ablation experiment...
2024
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.