Recognition: unknown
Real-Time Human Reconstruction and Animation using Feed-Forward Gaussian Splatting
Pith reviewed 2026-05-10 16:17 UTC · model grok-4.3
The pith
A single forward pass on multi-view images creates an animatable 3D human by predicting Gaussians attached to SMPL-X vertices.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
Our method predicts, in a canonical pose, a set of 3D Gaussian primitives associated with each SMPL-X vertex. One Gaussian is regularized to remain close to the SMPL-X surface, providing a strong geometric prior and stable correspondence to the parametric body model, while an additional small set of unconstrained Gaussians per vertex allows the representation to capture geometric structures that deviate from the parametric surface, such as clothing and hair. In contrast to recent approaches that require repeated network inference to synthesize novel poses, our method produces an animatable human representation from a single forward pass that can be efficiently animated via linear blend skinn
What carries the argument
Per-vertex Gaussian prediction in canonical space, where one Gaussian is constrained to the SMPL-X surface for geometric prior and the rest are unconstrained to capture deviations like clothing.
If this is right
- Reconstruction quality matches state-of-the-art on THuman 2.1, AvatarReX, and THuman 4.0 datasets
- Supports real-time animation and interactive applications
- Requires only a single forward pass instead of repeated inferences for each new pose or view
- The model generalizes to novel subjects and poses using the SMPL-X association
Where Pith is reading between the lines
- Such an approach could lower the barrier for creating personalized avatars in consumer applications by removing the need for heavy per-pose computation.
- Connecting the representation to an existing parametric model like SMPL-X may allow easy integration with motion capture systems and physics simulations for more realistic animations.
- If the association holds under deformation, it suggests a path toward hybrid representations that blend explicit body models with implicit details.
Load-bearing premise
The prediction of Gaussians in canonical space remains accurate and free of artifacts when the model is deformed to arbitrary new poses and subjects using linear blend skinning.
What would settle it
A direct test would be to animate the reconstructed model to a pose far from the training distribution on the THuman 4.0 dataset and measure if the rendered quality drops significantly compared to per-pose optimized methods, or if the clothing and hair details collapse or distort.
Figures
read the original abstract
We present a generalizable feed-forward Gaussian splatting framework for human 3D reconstruction and real-time animation that operates directly on multi-view RGB images and their associated SMPL-X poses. Unlike prior methods that rely on depth supervision, fixed input views, UV map, or repeated feed-forward inference for each target view or pose, our approach predicts, in a canonical pose, a set of 3D Gaussian primitives associated with each SMPL-X vertex. One Gaussian is regularized to remain close to the SMPL-X surface, providing a strong geometric prior and stable correspondence to the parametric body model, while an additional small set of unconstrained Gaussians per vertex allows the representation to capture geometric structures that deviate from the parametric surface, such as clothing and hair. In contrast to recent approaches such as HumanRAM, which require repeated network inference to synthesize novel poses, our method produces an animatable human representation from a single forward pass; by explicitly associating Gaussian primitives with SMPL-X vertices, the reconstructed model can be efficiently animated via linear blend skinning without further network evaluation. We evaluate our method on the THuman 2.1, AvatarReX and THuman 4.0 datasets, where it achieves reconstruction quality comparable to state-of-the-art methods while uniquely supporting real-time animation and interactive applications. Code and pre-trained models are available at https://github.com/Devdoot57/HumanGS .
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The manuscript proposes a feed-forward Gaussian splatting framework for human 3D reconstruction and real-time animation from multi-view RGB images and associated SMPL-X poses. It predicts, in a single forward pass and in canonical space, a set of 3D Gaussian primitives per SMPL-X vertex: one regularized to lie near the parametric surface and a small number of additional unconstrained Gaussians to capture deviations such as clothing and hair. The resulting model is then animated via standard linear blend skinning without any further network inference. The authors evaluate on THuman 2.1, AvatarReX and THuman 4.0, claiming reconstruction quality comparable to prior art while uniquely enabling real-time animation and interactive use; code and pre-trained models are released.
Significance. If the single-pass animatability claim holds, the work would provide a practical advance for real-time human avatar applications in VR/AR and graphics by combining Gaussian splatting expressiveness with SMPL-X parametric control in an efficient feed-forward pipeline. The explicit per-vertex Gaussian-to-SMPL-X association and the public release of code and models are clear strengths that aid reproducibility.
major comments (2)
- [Experiments] Experiments section: the manuscript states that the method 'achieves reconstruction quality comparable to state-of-the-art methods while uniquely supporting real-time animation' on THuman 2.1, AvatarReX and THuman 4.0, yet supplies no quantitative metrics (PSNR, SSIM, LPIPS), error bars, ablation tables, or animation-specific results (e.g., per-pose quality drop on held-out extreme poses or cross-subject generalization). This evidence gap is load-bearing for the central single-pass animation claim.
- [Method] Method (Gaussian association and LBS animation): the animatability claim rests on the assumption that Gaussians predicted in canonical space—one constrained to the SMPL-X surface and others free—remain stable and accurate when deformed by linear blend skinning for novel poses and subjects. No explicit regularization, stability analysis, or quantitative tests for this assumption under LBS (a first-order approximation ignoring secondary dynamics) are provided, directly undermining the 'without further network evaluation' guarantee.
Simulated Author's Rebuttal
We thank the referee for the constructive feedback. The comments highlight important areas for strengthening the experimental evidence and methodological justification. We address each major comment below and have prepared revisions to the manuscript.
read point-by-point responses
-
Referee: [Experiments] Experiments section: the manuscript states that the method 'achieves reconstruction quality comparable to state-of-the-art methods while uniquely supporting real-time animation' on THuman 2.1, AvatarReX and THuman 4.0, yet supplies no quantitative metrics (PSNR, SSIM, LPIPS), error bars, ablation tables, or animation-specific results (e.g., per-pose quality drop on held-out extreme poses or cross-subject generalization). This evidence gap is load-bearing for the central single-pass animation claim.
Authors: We agree that the original submission would be strengthened by explicit quantitative reporting. In the revised manuscript we have added tables reporting PSNR, SSIM and LPIPS on all three datasets, with standard-error bars computed across subjects. We have also inserted an ablation study and animation-specific evaluations that measure reconstruction quality on held-out extreme poses and cross-subject generalization, showing only modest degradation relative to the canonical-pose results. These additions directly substantiate the single-pass animation claim. revision: yes
-
Referee: [Method] Method (Gaussian association and LBS animation): the animatability claim rests on the assumption that Gaussians predicted in canonical space—one constrained to the SMPL-X surface and others free—remain stable and accurate when deformed by linear blend skinning for novel poses and subjects. No explicit regularization, stability analysis, or quantitative tests for this assumption under LBS (a first-order approximation ignoring secondary dynamics) are provided, directly undermining the 'without further network evaluation' guarantee.
Authors: The design already includes an explicit surface-regularization term that anchors one Gaussian per SMPL-X vertex, providing a stable correspondence under LBS. The remaining per-vertex Gaussians are still rigidly attached to the same vertex and therefore deform identically. While the initial submission did not contain a dedicated stability analysis, the revised version adds a short section discussing the first-order nature of LBS together with quantitative fidelity measurements on novel poses. We acknowledge that LBS omits secondary dynamics; the feed-forward canonical representation nevertheless enables real-time animation without re-inference, which is the central practical contribution. revision: partial
Circularity Check
No circularity: feed-forward prediction and LBS animation are independent of fitted outputs
full rationale
The derivation chain consists of a neural network predicting per-vertex Gaussian parameters (positions, covariances, etc.) in canonical space from multi-view RGB + SMPL-X input, followed by explicit attachment of one Gaussian to each SMPL-X vertex plus a small unconstrained set, then standard linear blend skinning for animation. None of these steps reduces to a self-definition, a fitted parameter renamed as prediction, or a self-citation chain. The network output is learned from data and evaluated on held-out datasets (THuman 2.1/4.0, AvatarReX); LBS is an external, non-learned deformation model. The single-pass animatability claim rests on the learned generalization, not on any equation that is tautological by construction. No load-bearing uniqueness theorem or ansatz is imported from prior author work.
Axiom & Free-Parameter Ledger
axioms (1)
- domain assumption SMPL-X parametric model supplies accurate vertex locations and skinning weights for stable correspondence and animation
Reference graph
Works this paper leans on
-
[1]
In: Proceedings of the IEEE/CVF International Conference on Computer Vision
Chen, A., Xu, Z., Zhao, F., Zhang, X., Xiang, F., Yu, J., Su, H.: Mvsnerf: Fast generalizable radiance field reconstruction from multi-view stereo. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 14124–14133 (2021)
2021
-
[2]
In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
Chen, J., Yi, W., Ma, L., Jia, X., Lu, H.: Gm-nerf: Learning generalizable model-based neural radiance fields from multi-view images. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 20648– 20658 (2023) Real-Time Human Animation 17
2023
-
[3]
In: Euro- pean Conference on Computer Vision
Chen, M., Zhang, J., Xu, X., Liu, L., Cai, Y., Feng, J., Yan, S.: Geometry-guided progressive nerf for generalizable and efficient neural human rendering. In: Euro- pean Conference on Computer Vision. pp. 222–239. Springer (2022)
2022
-
[4]
arXiv preprint arXiv:2204.11798 (2022)
Cheng, W., Xu, S., Piao, J., Qian, C., Wu, W., Lin, K.Y., Li, H.: Generalizable neural performer: Learning robust radiance fields for human novel view synthesis. arXiv preprint arXiv:2204.11798 (2022)
-
[5]
ACM Transactions on Graphics (TOG)34(4), 69:1–69:13 (2015)
Collet, A., Chuang, M., Sweeney, P., Gillett, D., Evseev, D., Calabrese, D., Hoppe, H., Kirk, A., Sullivan, S.: High-quality streamable free-viewpoint video. ACM Transactions on Graphics (TOG)34(4), 69:1–69:13 (2015)
2015
-
[6]
Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., Uszkoreit, J., Houlsby, N.: An image is worth 16x16 words: Transformers for image recognition at scale (10 2020).https://doi.org/10.48550/arXiv.2010.11929
work page internal anchor Pith review Pith/arXiv arXiv doi:10.48550/arxiv.2010.11929 2020
-
[7]
IEEE Transactions on Visualization and Computer Graphics (2023)
Gao, Q., Wang, Y., Liu, L., Liu, L., Theobalt, C., Chen, B.: Neural novel actor: Learning a generalized animatable neural representation for human actors. IEEE Transactions on Visualization and Computer Graphics (2023)
2023
-
[8]
IEEE Transactions on Pattern Analysis and Machine Intelligence (2022)
Gao, X., Yang, J., Kim, J., Peng, S., Liu, Z., Tong, X.: Mps-nerf: Generalizable 3d human rendering from multiview images. IEEE Transactions on Pattern Analysis and Machine Intelligence (2022)
2022
-
[9]
In: Proc
Geng, C., Peng, S., Xu, Z., Bao, H., Zhou, X.: Learning neural volumetric repre- sentations of dynamic humans in minutes. In: Proc. of CVPR (2023)
2023
-
[10]
In: Proc
Guo, C., Jiang, T., Chen, X., Song, J., Hilliges, O.: Vid2avatar: 3d avatar recon- struction from videos in the wild via self-supervised scene decomposition. In: Proc. of CVPR (2023)
2023
-
[11]
ACM Transactions on Graphics (ToG)38(6), 1–19 (2019)
Guo, K., Lincoln, P., Davidson, P., Busch, J., Yu, X., Whalen, M., Harvey, G., Orts-Escolano, S., Pandey, R., Dourgarian, J., et al.: The relightables: Volumetric performance capture of humans with realistic relighting. ACM Transactions on Graphics (ToG)38(6), 1–19 (2019)
2019
-
[12]
LRM: Large Reconstruction Model for Single Image to 3D
Hong, Y., Zhang, K., Gu, J., Bi, S., Zhou, Y., Liu, D., Liu, F., Sunkavalli, K., Bui, T., Tan, H.: Lrm: Large reconstruction model for single image to 3d. arXiv preprint arXiv:2311.04400 (2023)
work page internal anchor Pith review arXiv 2023
-
[13]
In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)
Hu, S., Liu, Z.: Gauhuman: Articulated gaussian splatting from monocular human videos. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). pp. 20418–20431. IEEE (2024)
2024
-
[14]
In: Proc
Jiang, T., Chen, X., Song, J., Hilliges, O.: Instantavatar: Learning avatars from monocular video in 60 seconds. In: Proc. of CVPR (2023)
2023
-
[15]
In: Proc
Jiang, W., Yi, K.M., Samei, G., Tuzel, O., Ranjan, A.: Neuman: Neural human radiance field from a single video. In: Proc. of ECCV (2022)
2022
-
[16]
Jin, H., Jiang, H., Tan, H., Zhang, K., Bi, S., Zhang, T., Luan, F., Snavely, N., Xu, Z.: Lvsm: A large view synthesis model with minimal 3d inductive bias. arXiv preprint arXiv:2410.17242 (2024)
-
[17]
ACM Trans
Kerbl, B., Kopanas, G., Leimkühler, T., Drettakis, G.: 3d gaussian splatting for real-time radiance field rendering. ACM Trans. Graph.42(4), 139–1 (2023)
2023
-
[18]
In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)
Kocabas, M., Chang, J.H.R., Gabriel, J., Tuzel, O., Ranjan, A.: Hugs: Human gaussian splats. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). pp. 505–515. IEEE (2024)
2024
-
[19]
Kwon, Y., Fang, B., Lu, Y., Dong, H., Zhang, C., Carrasco, F.V., Mosella-Montoro, A., Xu, J., Takagi, S., Kim, D., Prakash, A., la Torre, F.D.: Generalizable human gaussians for sparse view synthesis (2024),https://arxiv.org/abs/2407.12777 18 Devdoot Chatterjee, Zakaria Laskar, and C.V. Jawahar
-
[20]
In: European Conference on Computer Vision
Kwon, Y., Fang, B., Lu, Y., Dong, H., Zhang, C., Carrasco, F.V., Mosella- Montoro, A., Xu, J., Takagi, S., Kim, D., et al.: Generalizable human gaussians for sparse view synthesis. In: European Conference on Computer Vision. pp. 451–468. Springer (2024)
2024
-
[21]
Advances in Neural Information Processing Systems34, 24741–24752 (2021)
Kwon, Y., Kim, D., Ceylan, D., Fuchs, H.: Neural human performer: Learning generalizable radiance fields for human performance rendering. Advances in Neural Information Processing Systems34, 24741–24752 (2021)
2021
-
[22]
In: International Conference on Learning Representations (2023)
Kwon, Y., Kim, D., Ceylan, D., Fuchs, H.: Neural image-based avatars: General- izable radiance fields for human avatar modeling. In: International Conference on Learning Representations (2023)
2023
-
[23]
Association for Comput- ing Machinery, New York, NY, USA, 1 edn
Lewis, J.P., Cordner, M., Fong, N.: Pose Space Deformation: A Unified Approach to Shape Interpolation and Skeleton-Driven Deformation. Association for Comput- ing Machinery, New York, NY, USA, 1 edn. (2023),https://doi.org/10.1145/ 3596711.3596796
-
[24]
Li, M., Yao, S., Xie, Z., Chen, K.: Gaussianbody: Clothed human reconstruction via 3d gaussian splatting (2024)
2024
-
[25]
In: Proc
Li, R., Tanke, J., Vo, M., Zollhoefer, M., Gall, J., Kanazawa, A., Lassner, C.: Tava: Template-free animatable volumetric actors. In: Proc. of ECCV (2022)
2022
-
[26]
In: Proceed- ings of the IEEE/CVF conference on computer vision and pattern recognition
Li, Z., Zheng, Z., Wang, L., Liu, Y.: Animatable gaussians: Learning pose- dependent gaussian maps for high-fidelity human avatar modeling. In: Proceed- ings of the IEEE/CVF conference on computer vision and pattern recognition. pp. 19711–19722 (2024)
2024
-
[27]
In: Proc
Li, Z., Zheng, Z., Wang, L., Liu, Y.: Animatable gaussians: Learning pose- dependent gaussian maps for high-fidelity human avatar modeling. In: Proc. of CVPR (2024)
2024
-
[28]
Liu, L., Habermann, M., Rudnev, V., Sarkar, K., Gu, J., Theobalt, C.: Neural actor: Neural free-view synthesis of human actors with pose control (2021)
2021
-
[29]
Commu- nications of the ACM65(1), 99–106 (2021)
Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: Nerf: Representing scenes as neural radiance fields for view synthesis. Commu- nications of the ACM65(1), 99–106 (2021)
2021
-
[30]
In: CVPR (2024)
Moreau, A., Song, J., Dhamo, H., Shaw, R., Zhou, Y., Pérez-Pellitero, E.: Human gaussian splatting: Real-time rendering of animatable avatars. In: CVPR (2024)
2024
-
[31]
ACM Transasctions Graphics41(4) (2022)
Müller, T., Evans, A., Schied, C., Keller, A.: Instant neural graphics primitives with a multiresolution hash encoding. ACM Transasctions Graphics41(4) (2022)
2022
-
[32]
In: Proc
Noguchi, A., Sun, X., Lin, S., Harada, T.: Neural articulated radiance field. In: Proc. of ICCV (2021)
2021
-
[33]
In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)
Pang,H.,Zhu,H.,Kortylewski,A.,Theobalt,C.,Habermann,M.:Ash:Animatable gaussian splats for efficient and photoreal human rendering. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). pp. 1165–1175 (2024)
2024
-
[34]
In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition
Pavlakos, G., Choutas, V., Ghorbani, N., Bolkart, T., Osman, A.A., Tzionas, D., Black, M.J.: Expressive body capture: 3d hands, face, and body from a single im- age. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. pp. 10975–10985 (2019)
2019
-
[35]
In: Proceedings of the IEEE/CVF International Conference on Computer Vision
Peng, S., Dong, J., Wang, Q., Zhang, S., Shuai, Q., Zhou, X., Bao, H.: Animatable neural radiance fields for modeling dynamic human bodies. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 14314–14323 (2021)
2021
-
[36]
ArXivabs/2203.08133 (2022) Real-Time Human Animation 19
Peng, S., Zhang, S., Xu, Z., Geng, C., Jiang, B., Bao, H., Zhou, X.: Animatable neural implicit surfaces for creating avatars from videos. ArXivabs/2203.08133 (2022) Real-Time Human Animation 19
-
[37]
In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition
Peng, S., Zhang, Y., Xu, Y., Wang, Q., Shuai, Q., Bao, H., Zhou, X.: Neural body: Implicit neural representations with structured latent codes for novel view synthesis of dynamic humans. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. pp. 9054–9063 (2021)
2021
-
[38]
Plucker, J.: Xvii. on a new geometry of space. Philosophical Transactions of the Royal Society of London (155), 725–791 (12 1865).https://doi.org/10.1098/ rstl.1865.0017,https://doi.org/10.1098/rstl.1865.0017
-
[39]
arXiv preprint arXiv:2409.04196 (2024)
Prospero, L., Hamdi, A., Henriques, J.F., Rupprecht, C.: Gst: Precise 3d human body from a single image with gaussian splatting transformers. arXiv preprint arXiv:2409.04196 (2024)
-
[40]
Prospero, L., Hamdi, A., Henriques, J.F., Rupprecht, C.: Gst: Precise 3d human bodyfromasingleimagewithgaussiansplattingtransformers.In:2025IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). pp. 5997–6007 (2025).https://doi.org/10.1109/CVPRW67362.2025.00598
-
[41]
In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2024)
Qian, Z., Wang, S., Mihajlovic, M., Geiger, A., Tang, S.: 3dgs-avatar: Animatable avatars via deformable 3d gaussian splatting. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2024)
2024
-
[42]
In: Proceedings of the IEEE/CVF international conference on computer vision
Saito, S., Huang, Z., Natsume, R., Morishima, S., Kanazawa, A., Li, H.: Pifu: Pixel-aligned implicit function for high-resolution clothed human digitization. In: Proceedings of the IEEE/CVF international conference on computer vision. pp. 2304–2314 (2019)
2019
-
[43]
In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
Shao, Z., Wang, Z., Li, Z., Wang, D., Lin, X., Zhang, Y., Fan, M., Wang, Z.: Splattingavatar: Realistic real-time human avatars with mesh-embedded gaussian splatting. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 1606–1616 (2024)
2024
-
[44]
In: Proc
Su, S.Y., Yu, F., Zollhoefer, M., Rhodin, H.: A-neRF: Articulated neural radiance fields for learning human shape, appearance, and pose. In: Proc. of NeurIPS (2021)
2021
-
[45]
ACM Transactions on Graphics (TOG)38(4), 1–12 (2019)
Thies, J., Zollhöfer, M., Nießner, M.: Deferred neural rendering: Image synthesis using neural textures. ACM Transactions on Graphics (TOG)38(4), 1–12 (2019)
2019
- [46]
-
[47]
In: Proc
Wang, S., Schwarz, K., Geiger, A., Tang, S.: Arah: Animatable volume rendering of articulated human sdfs. In: Proc. of ECCV (2022)
2022
-
[48]
IEEE Transactions on Visualization and Computer Graphics (2024)
Wang, S., Wang, Z., Schmelzle, R., Zheng, L., Kwon, Y., Sengupta, R., Fuchs, H.: Learning view synthesis for desktop telepresence with few rgbd cameras. IEEE Transactions on Visualization and Computer Graphics (2024)
2024
-
[49]
Weng, C.Y., Curless, B., Kemelmacher-Shlizerman, I.: Vid2actor: Free-viewpoint animatable person synthesis from video in the wild (2020)
2020
-
[50]
Weng, C.Y., Curless, B., Srinivasan, P.P., Barron, J.T., Kemelmacher-Shlizerman, I.: Humannerf: Free-viewpoint rendering of moving people from monocular video (2022)
2022
-
[51]
In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)
Xiu, Y., Yang, J., Tzionas, D., Black, M.J.: ICON: Implicit Clothed humans Ob- tained from Normals. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). pp. 13296–13306 (June 2022)
2022
-
[52]
In: Proc
Xu,H.,Alldieck,T.,Sminchisescu,C.:H-neRF:Neuralradiancefieldsforrendering and temporal reconstruction of humans in motion. In: Proc. of NeurIPS (2021)
2021
-
[53]
In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition
Yu,A.,Ye,V.,Tancik,M.,Kanazawa, A.:pixelnerf:Neuralradiance fieldsfromone or few images. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. pp. 4578–4587 (2021)
2021
-
[54]
In: Proc
Yu, Z., Cheng, W., Liu, x., Wu, W., Lin, K.Y.: MonoHuman: Animatable human neural field from monocular video. In: Proc. of CVPR (2023) 20 Devdoot Chatterjee, Zakaria Laskar, and C.V. Jawahar
2023
-
[55]
In: Proceedings of the Spe- cial Interest Group on Computer Graphics and Interactive Techniques Conference Conference Papers
Yu, Z., Li, Z., Bao, H., Yang, C., Zhou, X.: Humanram: Feed-forward human re- construction and animation model using transformers. In: Proceedings of the Spe- cial Interest Group on Computer Graphics and Interactive Techniques Conference Conference Papers. pp. 1–13 (2025)
2025
-
[56]
ACM Trans- actions on Graphics (TOG)40(4) (2021)
Zhang, J., Liu, X., Ye, X., Zhao, F., Zhang, Y., Wu, M., Zhang, Y., Xu, L., Yu, J.: Editable free-viewpoint video using a layered neural representation. ACM Trans- actions on Graphics (TOG)40(4) (2021)
2021
-
[57]
In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
Zhao, F., Yang, W., Zhang, J., Lin, P., Zhang, Y., Yu, J., Xu, L.: Humannerf: Efficiently generated human radiance field from sparse inputs. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 7743–7753 (2022)
2022
-
[58]
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2024)
Zheng, S., Zhou, B., Shao, R., Liu, B., Zhang, S., Nie, L., Liu, Y.: Gps-gaussian: Generalizable pixel-wise 3d gaussian splatting for real-time human novel view syn- thesis. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (2024)
2024
-
[59]
In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (June 2022)
Zheng,Z.,Huang,H.,Yu,T.,Zhang,H.,Guo,Y.,Liu,Y.:Structuredlocalradiance fields for human avatar modeling. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (June 2022)
2022
-
[60]
Zheng, Z., Zhao, X., Zhang, H., Liu, B., Liu, Y.: Avatarrex: Real-time expressive full-body avatars. ACM Transactions on Graphics (TOG)42(4) (2023) Real-Time Human Animation 1 Supplementary Material Real-Time Human Reconstruction and Animation using Feed-Forward Gaussian Splatting S1 Experiment Details Evaluation Setup.To rigorously evaluate our method, w...
2023
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.