pith. machine review for the scientific record. sign in

arxiv: 2604.17807 · v1 · submitted 2026-04-20 · 💻 cs.CV · cs.RO

Recognition: unknown

Re²MoGen: Open-Vocabulary Motion Generation via LLM Reasoning and Physics-Aware Refinement

Authors on Pith no claims yet

Pith reviewed 2026-05-10 05:39 UTC · model grok-4.3

classification 💻 cs.CV cs.RO
keywords open-vocabulary motion generationtext-to-motion synthesisLLM reasoningMonte Carlo tree searchphysics-aware refinementkeyframe planningreinforcement learning post-training
0
0 comments X

The pith

Re²MoGen generates open-vocabulary motions by planning keyframes with LLM reasoning and then refining them for physical plausibility through pose optimization and reinforcement learning.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper sets out to solve the limitation of existing text-to-motion models, which work well only on descriptions similar to their training data but fail on novel prompts. It proposes a three-stage process that first uses Monte Carlo tree search to guide an LLM toward reasonable keyframe plans specifying root and key joint positions, then optimizes full-body poses with a human prior and fine-tunes a motion generator using dynamic temporal matching for completion, and finally applies physics-aware reinforcement learning to remove implausibilities. If the approach holds, it would let users produce character animations from arbitrary text instructions without requiring new paired training examples for every scenario. The central mechanism is the staged separation of semantic planning from physical correction, which aims to keep motions both faithful to the prompt and free of artifacts like foot sliding or joint violations.

Core claim

Re²MoGen shows that open-vocabulary motion generation becomes feasible when an LLM first produces sparse keyframe plans via enhanced reasoning, a pose model then completes full-body trajectories under a dynamic matching objective, and reinforcement learning finally enforces physical constraints on the resulting motion sequence, yielding outputs that remain semantically aligned with the original text while satisfying biomechanical rules.

What carries the argument

Three-stage pipeline that uses Monte Carlo tree search on an LLM to generate sparse keyframes, followed by pose optimization and RL-based physics refinement to complete and correct the full motion.

Load-bearing premise

Monte Carlo tree search on the LLM will produce keyframes that remain reasonable and semantically faithful even for motion descriptions far outside the training distribution, and the later optimization and refinement steps will not introduce new semantic drift or physical errors.

What would settle it

Test the system on a set of highly novel prompts such as 'a person juggling while riding a unicycle on ice' and measure whether human evaluators or a physics simulator rate the outputs as both matching the described action and free of violations like interpenetration or unstable balance; failure on either criterion for a majority of cases would falsify the central claim.

Figures

Figures reproduced from arXiv: 2604.17807 by Chenjia Bai, Jiakun Zheng, Shiqin Cao, Ting Xiao, Xinran Li, Zhe Wang.

Figure 1
Figure 1. Figure 1: The framework of Re2MoGen, which consists of three key parts: (i) MCTS-enhanced LLM Reasoning; (ii) Motion Completion and Finetuning, and (iii) Physics-aware refinement. In this paper, we adopt Motion Latent Diffusion (MLD) [4] as the basic model, denoted as pˆ 1:L = πmld(ci). For the basic motion generation process, MLD aims to model the conditional distribution of motions given a set of text de￾scription… view at source ↗
Figure 2
Figure 2. Figure 2: Qualitative comparison results of motions generated by different methods. [PITH_FULL_IMAGE:figures/full_fig_p007_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Visualization of our generated motions on the MuJoCo [PITH_FULL_IMAGE:figures/full_fig_p008_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Deploy generated motions on the real world robot. [PITH_FULL_IMAGE:figures/full_fig_p008_4.png] view at source ↗
Figure 5
Figure 5. Figure 5: The initial pose for LLM planning. 8. B. Experiments Details 8.1. LLM-planned Keyframes As shown in Figs.(10-16), we present the JSON data of keyframes planned by the LLM of different motion lengths alongside the rendered pose images after full-body opti￾mization. These results demonstrate that LLM can reason￾ably plan actions based on text descriptions. 8.2. Evaluation Metrics As mentioned in the main tex… view at source ↗
Figure 6
Figure 6. Figure 6: Our prompt template for LLM reasoning. 9.1. Experiments on SnapMotion [PITH_FULL_IMAGE:figures/full_fig_p014_6.png] view at source ↗
Figure 7
Figure 7. Figure 7: Examples for LLM reasoning. 0.05, 0.5, 1}. The results are shown in [PITH_FULL_IMAGE:figures/full_fig_p015_7.png] view at source ↗
Figure 8
Figure 8. Figure 8: Related reasons for the given examples. Please evaluate the alignment between a given generative motion clip and the corresponding text description ('{DESCRIPTION}'). The motion clip is represented as a sequence of frames {INPUT_LENGTH}. The motion only represents the character's actions, and is not concerned with the presence or absence of objects. Rating from 0 to 5. Scoring Criteria: 0-1: The motion doe… view at source ↗
Figure 9
Figure 9. Figure 9: VLM evaluation prompt. 5 [PITH_FULL_IMAGE:figures/full_fig_p016_9.png] view at source ↗
Figure 10
Figure 10. Figure 10: LLM-planned key joint positions and rendered images. [PITH_FULL_IMAGE:figures/full_fig_p017_10.png] view at source ↗
Figure 11
Figure 11. Figure 11: LLM-planned key joint positions and rendered images. [PITH_FULL_IMAGE:figures/full_fig_p017_11.png] view at source ↗
Figure 12
Figure 12. Figure 12: LLM-planned key joint positions and rendered images. [PITH_FULL_IMAGE:figures/full_fig_p018_12.png] view at source ↗
Figure 13
Figure 13. Figure 13: LLM-planned key joint positions and rendered images. [PITH_FULL_IMAGE:figures/full_fig_p018_13.png] view at source ↗
Figure 14
Figure 14. Figure 14: LLM-planned key joint positions and rendered images. [PITH_FULL_IMAGE:figures/full_fig_p019_14.png] view at source ↗
Figure 15
Figure 15. Figure 15: LLM-planned key joint positions and rendered images. [PITH_FULL_IMAGE:figures/full_fig_p019_15.png] view at source ↗
Figure 16
Figure 16. Figure 16: LLM-planned key joint positions and rendered images. [PITH_FULL_IMAGE:figures/full_fig_p020_16.png] view at source ↗
Figure 17
Figure 17. Figure 17: Additional qualitative comparison results of motions generated by different methods. [PITH_FULL_IMAGE:figures/full_fig_p020_17.png] view at source ↗
Figure 18
Figure 18. Figure 18: Additional qualitative comparison results of motions generated by different methods. [PITH_FULL_IMAGE:figures/full_fig_p021_18.png] view at source ↗
Figure 19
Figure 19. Figure 19: Additional results on the MuJoCo platform. [PITH_FULL_IMAGE:figures/full_fig_p021_19.png] view at source ↗
read the original abstract

Text-to-motion (T2M) generation aims to control the behavior of a target character via textual descriptions. Leveraging text-motion paired datasets, existing T2M models have achieved impressive performance in generating high-quality motions within the distribution of their training data. However, their performance deteriorates notably when the motion descriptions differ significantly from the training texts. To address this issue, we propose Re$^2$MoGen, a Reasoning and Refinement open-vocabulary Motion Generation framework that leverages enhanced Large Language Model (LLM) reasoning to generate an initial motion planning and then refine its physical plausibility via reinforcement learning (RL) post-training. Specifically, Re$^2$MoGen consists of three stages: We first employ Monte Carlo tree search to enhance the LLM's reasoning ability in generating reasonable keyframes of the motion based on text prompts, specifying only the root and several key joints' positions to ease the reasoning process. Then, we apply a human pose model as a prior to optimize the full-body poses based on the planned keyframes and use the resulting incomplete motion to supervise fine-tuning a pre-trained motion generator via a dynamic temporal matching objective, enabling spatiotemporal completion. Finally, we use post-training with physics-aware reward to refine motion quality to eliminate physical implausibility in LLM-planned motions. Extensive experiments demonstrate that our framework can generate semantically consistent and physically plausible motions and achieve state-of-the-art performance in open-vocabulary motion generation.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 3 minor

Summary. The paper introduces Re²MoGen, a three-stage framework for open-vocabulary text-to-motion generation. Stage 1 uses Monte Carlo tree search to enhance LLM reasoning for planning sparse keyframes (root and key joint positions) from text prompts. Stage 2 applies a human pose prior to optimize full-body poses and uses dynamic temporal matching to fine-tune a pre-trained motion generator for completion. Stage 3 performs physics-aware RL post-training to improve physical plausibility. The central claim is that this produces semantically consistent, physically plausible motions and achieves SOTA results on prompts outside the training distribution of standard T2M models.

Significance. If the quantitative results and component validations hold, the work would be significant for extending motion generation to open-vocabulary settings without requiring new large-scale paired datasets, by leveraging LLM reasoning plus physics refinement. It could influence hybrid LLM-physics approaches in CV and robotics. However, the significance is limited by the absence of isolated validation for the LLM+MCTS stage on out-of-distribution cases, which underpins the semantic consistency claim.

major comments (3)
  1. [§3.1] §3.1 (LLM Reasoning with MCTS): The central claim of semantic consistency for novel prompts rests on MCTS-enhanced LLM keyframe planning producing reasonable root/key-joint positions. No isolated quantitative evaluation (e.g., keyframe position error, semantic alignment score, or failure rate) is reported on deliberately out-of-distribution prompts. Later stages (pose optimization, temporal matching, RL) have no recovery mechanism if keyframes are semantically drifted, making this dependency load-bearing and unverified.
  2. [§4] §4 (Experiments): The SOTA and physical-plausibility claims lack ablation studies that isolate each stage's contribution (MCTS planning, dynamic temporal matching, physics RL) and direct comparisons against recent open-vocabulary baselines using standard metrics (FID, R-Precision, foot-skating, penetration). Without these, it is unclear whether performance gains are attributable to the proposed components or to the underlying pre-trained models.
  3. [§4.3] §4.3 (Quantitative Results): Reported metrics for open-vocabulary performance are not accompanied by error bars, number of evaluation runs, or statistical significance tests, and no failure-case analysis on prompts far outside training distribution is provided. This weakens the assertion that the framework reliably handles open-vocabulary inputs.
minor comments (3)
  1. [§3.2] Notation for the dynamic temporal matching objective is introduced without an explicit equation or pseudocode, making it difficult to reproduce the fine-tuning loss.
  2. [Figure 3] Figure 3 (qualitative results) would benefit from side-by-side comparison with a strong baseline on the same novel prompts to illustrate the claimed improvements.
  3. [§2] The manuscript cites prior T2M works but omits recent LLM-based motion planning papers from 2023-2024; adding these would strengthen the related-work section.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for the constructive and detailed feedback on our manuscript. We have carefully reviewed each major comment and provide point-by-point responses below. In the revised manuscript, we have incorporated additional experiments, ablations, and statistical analyses to address the concerns raised regarding validation of individual components and reporting rigor.

read point-by-point responses
  1. Referee: [§3.1] §3.1 (LLM Reasoning with MCTS): The central claim of semantic consistency for novel prompts rests on MCTS-enhanced LLM keyframe planning producing reasonable root/key-joint positions. No isolated quantitative evaluation (e.g., keyframe position error, semantic alignment score, or failure rate) is reported on deliberately out-of-distribution prompts. Later stages (pose optimization, temporal matching, RL) have no recovery mechanism if keyframes are semantically drifted, making this dependency load-bearing and unverified.

    Authors: We agree that an isolated quantitative evaluation of the MCTS-enhanced LLM keyframe planning stage on out-of-distribution prompts is important to substantiate the semantic consistency claim, given the dependency of downstream stages. In the revised manuscript, we have added a new subsection (§4.4) that reports keyframe position errors for root and key joints, semantic alignment scores via CLIP-based text-to-keyframe similarity, and observed failure rates on a curated set of deliberately out-of-distribution prompts. We also discuss the robustness of this stage and its influence on subsequent optimization and refinement. revision: yes

  2. Referee: [§4] §4 (Experiments): The SOTA and physical-plausibility claims lack ablation studies that isolate each stage's contribution (MCTS planning, dynamic temporal matching, physics RL) and direct comparisons against recent open-vocabulary baselines using standard metrics (FID, R-Precision, foot-skating, penetration). Without these, it is unclear whether performance gains are attributable to the proposed components or to the underlying pre-trained models.

    Authors: We thank the referee for highlighting the need for clearer attribution of performance gains. The revised manuscript now includes comprehensive ablation studies that isolate the contribution of each stage by ablating MCTS planning, dynamic temporal matching, and physics-aware RL individually, with results reported on FID, R-Precision, foot-skating ratio, and penetration depth. We have also added direct comparisons against recent open-vocabulary baselines using these standard metrics to demonstrate that the improvements stem from the proposed framework components. revision: yes

  3. Referee: [§4.3] §4.3 (Quantitative Results): Reported metrics for open-vocabulary performance are not accompanied by error bars, number of evaluation runs, or statistical significance tests, and no failure-case analysis on prompts far outside training distribution is provided. This weakens the assertion that the framework reliably handles open-vocabulary inputs.

    Authors: We acknowledge the importance of statistical rigor and failure analysis for claims of reliability on open-vocabulary inputs. In the revision, all quantitative tables now report results over 5 independent runs with different random seeds, including error bars as standard deviations. We have added statistical significance tests (paired t-tests with p-values) against baselines. Additionally, a new failure-case analysis subsection has been included, presenting qualitative and quantitative examples of challenging out-of-distribution prompts along with discussions of limitations. revision: yes

Circularity Check

0 steps flagged

No circularity: compositional pipeline using external pre-trained components

full rationale

The paper describes a three-stage engineering framework (LLM+MCTS keyframe planning, human-pose-prior optimization plus dynamic temporal matching fine-tuning of a pre-trained generator, and physics-aware RL post-training) that composes standard external models and techniques. No equations, derivations, or load-bearing claims reduce the final performance metric to a fitted parameter, self-definition, or self-citation chain; all stages are conditioned on independently trained priors whose correctness is not asserted by construction within this work.

Axiom & Free-Parameter Ledger

0 free parameters · 2 axioms · 0 invented entities

Based solely on the abstract, the framework assumes standard components from prior literature (pre-trained motion generators, human pose models, RL) without introducing new free parameters or invented entities in the high-level description.

axioms (2)
  • domain assumption A human pose model can serve as a reliable prior to optimize full-body poses from sparse keyframes.
    Invoked in the second stage to complete poses from planned keyframes.
  • domain assumption Physics-aware rewards in RL can eliminate implausibility without degrading semantic consistency from the LLM plan.
    Central to the final refinement stage.

pith-pipeline@v0.9.0 · 5576 in / 1410 out tokens · 44469 ms · 2026-05-10T05:39:32.555909+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

56 extracted references · 17 canonical work pages · 1 internal anchor

  1. [1]

    Cosmos-reason1: From physical common sense to embodied reasoning.arXiv preprint arXiv:2503.15558, 2025

    Alisson Azzolini, Junjie Bai, Hannah Brandon, Jiaxin Cao, Prithvijit Chattopadhyay, Huayu Chen, Jinju Chu, Yin Cui, Jenna Diamond, Yifan Ding, et al. Cosmos-reason1: From physical common sense to embodied reasoning.arXiv preprint arXiv:2503.15558, 2025. 2

  2. [2]

    Do as i can, not as i say: Grounding language in robotic affordances

    Anthony Brohan, Yevgen Chebotar, Chelsea Finn, Karol Hausman, Alexander Herzog, Daniel Ho, Julian Ibarz, Alex Irpan, Eric Jang, Ryan Julian, et al. Do as i can, not as i say: Grounding language in robotic affordances. InConference on robot learning, pages 287–318. PMLR, 2023. 2

  3. [3]

    Lan- guage models are few-shot learners.Advances in neural in- formation processing systems, 33:1877–1901, 2020

    Tom Brown, Benjamin Mann, Nick Ryder, Melanie Sub- biah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakan- tan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. Lan- guage models are few-shot learners.Advances in neural in- formation processing systems, 33:1877–1901, 2020. 2

  4. [4]

    Executing your commands via motion diffusion in latent space

    Xin Chen, Biao Jiang, Wen Liu, Zilong Huang, Bin Fu, Tao Chen, and Gang Yu. Executing your commands via motion diffusion in latent space. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 18000–18010, 2023. 1, 2, 3, 6

  5. [5]

    Reproducible scal- ing laws for contrastive language-image learning

    Mehdi Cherti, Romain Beaumont, Ross Wightman, Mitchell Wortsman, Gabriel Ilharco, Cade Gordon, Christoph Schuh- mann, Ludwig Schmidt, and Jenia Jitsev. Reproducible scal- ing laws for contrastive language-image learning. InPro- ceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2818–2829, 2023. 1

  6. [6]

    Anyskill: Learning open- vocabulary physical skill for interactive agents

    Jieming Cui, Tengyu Liu, Nian Liu, Yaodong Yang, Yixin Zhu, and Siyuan Huang. Anyskill: Learning open- vocabulary physical skill for interactive agents. InProceed- ings of the IEEE/CVF conference on computer vision and pattern recognition, pages 852–862, 2024. 1, 2, 6

  7. [7]

    Grove: A gen- eralized reward for learning open-vocabulary physical skill

    Jieming Cui, Tengyu Liu, Meng Ziyu, Yu Jiale, Ran Song, Wei Zhang, Yixin Zhu, and Siyuan Huang. Grove: A gen- eralized reward for learning open-vocabulary physical skill. InConference on Computer Vision and Pattern Recognition (CVPR), 2025. 1, 2

  8. [8]

    Soft-dtw: a differen- tiable loss function for time-series

    Marco Cuturi and Mathieu Blondel. Soft-dtw: a differen- tiable loss function for time-series. InInternational confer- ence on machine learning, pages 894–903. PMLR, 2017. 5

  9. [9]

    Dynamic planning with a llm

    Gautier Dagan, Frank Keller, and Alex Lascarides. Dynamic planning with a llm.arXiv preprint arXiv:2308.06391, 2023. 2

  10. [10]

    Textual decomposition then sub-motion-space scatter- ing for open-vocabulary motion generation.arXiv preprint arXiv:2411.04079, 2024

    Ke Fan, Jiangning Zhang, Ran Yi, Jingyu Gong, Yabiao Wang, Yating Wang, Xin Tan, Chengjie Wang, and Lizhuang Ma. Textual decomposition then sub-motion-space scatter- ing for open-vocabulary motion generation.arXiv preprint arXiv:2411.04079, 2024. 2

  11. [11]

    Dpok: Reinforcement learning for fine-tuning text-to-image diffu- sion models.Advances in Neural Information Processing Systems, 36:79858–79885, 2023

    Ying Fan, Olivia Watkins, Yuqing Du, Hao Liu, Moonkyung Ryu, Craig Boutilier, Pieter Abbeel, Moham- mad Ghavamzadeh, Kangwook Lee, and Kimin Lee. Dpok: Reinforcement learning for fine-tuning text-to-image diffu- sion models.Advances in Neural Information Processing Systems, 36:79858–79885, 2023. 5

  12. [12]

    Layoutgpt: Compositional visual plan- ning and generation with large language models.Advances in Neural Information Processing Systems, 36, 2024

    Weixi Feng, Wanrong Zhu, Tsu-jui Fu, Varun Jampani, Ar- jun Akula, Xuehai He, Sugato Basu, Xin Eric Wang, and William Yang Wang. Layoutgpt: Compositional visual plan- ning and generation with large language models.Advances in Neural Information Processing Systems, 36, 2024. 2

  13. [13]

    Generating diverse and natural 3d human motions from text

    Chuan Guo, Shihao Zou, Xinxin Zuo, Sen Wang, Wei Ji, Xingyu Li, and Li Cheng. Generating diverse and natural 3d human motions from text. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 5152–5161, 2022. 2, 6

  14. [14]

    Momask: Generative masked model- ing of 3d human motions

    Chuan Guo, Yuxuan Mu, Muhammad Gohar Javed, Sen Wang, and Li Cheng. Momask: Generative masked model- ing of 3d human motions. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1900–1910, 2024. 1, 2, 3

  15. [15]

    Snapmogen: Human motion generation from expressive texts, 2025

    Chuan Guo, Inwoo Hwang, Jian Wang, and Bing Zhou. Snapmogen: Human motion generation from expressive texts, 2025. 3

  16. [16]

    Reindiffuse: Craft- ing physically plausible motions with reinforced diffusion model

    Gaoge Han, Mingjiang Liang, Jinglei Tang, Yongkang Cheng, Wei Liu, and Shaoli Huang. Reindiffuse: Craft- ing physically plausible motions with reinforced diffusion model. In2025 IEEE/CVF Winter Conference on Applica- tions of Computer Vision (WACV), pages 2218–2227. IEEE,

  17. [17]

    Avatarclip: Zero-shot text- driven generation and animation of 3d avatars.ACM Trans- actions on Graphics (TOG), 41(4):1–19, 2022

    Fangzhou Hong, Mingyuan Zhang, Liang Pan, Zhongang Cai, Lei Yang, and Ziwei Liu. Avatarclip: Zero-shot text- driven generation and animation of 3d avatars.ACM Trans- actions on Graphics (TOG), 41(4):1–19, 2022. 1, 2

  18. [18]

    RoboBrain: A Unified Brain Model for Robotic Manipula- tion from Abstract to Concrete.arXiv preprint arXiv:2502.21257, 2025

    Yuheng Ji, Huajie Tan, Jiayu Shi, Xiaoshuai Hao, Yuan Zhang, Hengyuan Zhang, Pengwei Wang, Mengdi Zhao, Yao Mu, Pengju An, et al. Robobrain: A unified brain model for robotic manipulation from abstract to concrete.arXiv preprint arXiv:2502.21257, 2025. 2, 3

  19. [19]

    Motiongpt: Human motion as a foreign language.Ad- vances in Neural Information Processing Systems, 36, 2024

    Biao Jiang, Xin Chen, Wen Liu, Jingyi Yu, Gang Yu, and Tao Chen. Motiongpt: Human motion as a foreign language.Ad- vances in Neural Information Processing Systems, 36, 2024. 1, 2, 6

  20. [20]

    Harmon: Whole-bodymotiongenerationofhumanoidrobotsfromlan- guage descriptions

    Zhenyu Jiang, Yuqi Xie, Jinhan Li, Ye Yuan, Yifeng Zhu, and Yuke Zhu. Harmon: Whole-body motion generation of humanoid robots from language descriptions.arXiv preprint arXiv:2410.12773, 2024. 3

  21. [21]

    HumanoidGen: Data generation for bimanual dexterous manipulation via LLM reasoning

    Zhi Jing, Siyuan Yang, Jicong Ao, Ting Xiao, Yugang Jiang, and Chenjia Bai. Humanoidgen: Data generation for biman- ual dexterous manipulation via llm reasoning.arXiv preprint arXiv:2507.00833, 2025. 2

  22. [22]

    Pchc: Enabling prefer- ence conditioned humanoid control via multi-objective rein- 9 forcement learning.arXiv preprint arXiv:2603.24047, 2026

    Huanyu Li, Dewei Wang, Xinmiao Wang, Xinzhe Liu, Peng Liu, Chenjia Bai, and Xuelong Li. Pchc: Enabling prefer- ence conditioned humanoid control via multi-objective rein- 9 forcement learning.arXiv preprint arXiv:2603.24047, 2026. 1

  23. [23]

    Interreal: A unified physics-based imitation framework for learning human-object interaction skills.arXiv preprint arXiv:2603.07516, 2026

    Dayang Liang, Yuhang Lin, Xinzhe Liu, Jiyuan Shi, Yunlong Liu, and Chenjia Bai. Interreal: A unified physics-based imitation framework for learning human-object interaction skills.arXiv preprint arXiv:2603.07516, 2026. 1

  24. [24]

    Motion-x: A large- scale 3d expressive whole-body human motion dataset.Ad- vances in Neural Information Processing Systems, 2023

    Jing Lin, Ailing Zeng, Shunlin Lu, Yuanhao Cai, Ruimao Zhang, Haoqian Wang, and Lei Zhang. Motion-x: A large- scale 3d expressive whole-body human motion dataset.Ad- vances in Neural Information Processing Systems, 2023. 4, 6

  25. [25]

    Pro-hoi: Perceptive root-guided humanoid-object interaction.arXiv preprint arXiv:2603.01126, 2026

    Yuhang Lin, Jiyuan Shi, Dewei Wang, Jipeng Kong, Yong Liu, Chenjia Bai, and Xuelong Li. Pro-hoi: Perceptive root-guided humanoid-object interaction.arXiv preprint arXiv:2603.01126, 2026. 1

  26. [26]

    Plan, posture and go: To- wards open-world text-to-motion generation.arXiv preprint arXiv:2312.14828, 2023

    Jinpeng Liu, Wenxun Dai, Chunyu Wang, Yiji Cheng, Yan- song Tang, and Xin Tong. Plan, posture and go: To- wards open-world text-to-motion generation.arXiv preprint arXiv:2312.14828, 2023. 1, 2

  27. [27]

    Smpl: A skinned multi- person linear model

    Matthew Loper, Naureen Mahmood, Javier Romero, Gerard Pons-Moll, and Michael J Black. Smpl: A skinned multi- person linear model. InSeminal Graphics Papers: Pushing the Boundaries, Volume 2, pages 851–866. 2023. 1

  28. [28]

    Winkler, Kris Ki- tani, and Weipeng Xu

    Zhengyi Luo, Jinkun Cao, Alexander W. Winkler, Kris Ki- tani, and Weipeng Xu. Perpetual humanoid control for real- time simulated avatars. InInternational Conference on Com- puter Vision (ICCV), 2023. 1

  29. [29]

    Troje, Ger- ard Pons-Moll, and Michael J

    Naureen Mahmood, Nima Ghorbani, Nikolaus F. Troje, Ger- ard Pons-Moll, and Michael J. Black. Amass: Archive of motion capture as surface shapes. InThe IEEE International Conference on Computer Vision (ICCV), 2019. 4, 6

  30. [31]

    Generation of com- plex 3d human motion by temporal and spatial composition of diffusion models

    Lorenzo Mandelli and Stefano Berretti. Generation of com- plex 3d human motion by temporal and spatial composition of diffusion models. In2025 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), pages 1279–

  31. [32]

    Learning from massive human videos for universal humanoid pose control,

    Jiageng Mao, Siheng Zhao, Siqi Song, Tianheng Shi, Jun- jie Ye, Mingtong Zhang, Haoran Geng, Jitendra Malik, Vi- tor Guizilini, and Yue Wang. Learning from massive human videos for universal humanoid pose control.arXiv preprint arXiv:2412.14172, 2024. 1

  32. [33]

    Georgios Pavlakos, Vasileios Choutas, Nima Ghorbani, Timo Bolkart, Ahmed A. A. Osman, Dimitrios Tzionas, and Michael J. Black. Expressive body capture: 3d hands, face, and body from a single image. InProceedings IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2019. 4

  33. [34]

    Learning transferable visual models from natural language supervi- sion

    Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervi- sion. InInternational conference on machine learning, pages 8748–8763. PmLR, 2021. 1, 2

  34. [35]

    Proximal Policy Optimization Algorithms

    John Schulman, Filip Wolski, Prafulla Dhariwal, Alec Rad- ford, and Oleg Klimov. Proximal policy optimization algo- rithms.arXiv preprint arXiv:1707.06347, 2017. 5

  35. [36]

    Robotic navigation with large pre-trained models of language.Vision, and Action, 2022

    Dhruv Shah, Blazej Osinski, Brian Ichter, and Sergey Levine. Robotic navigation with large pre-trained models of language.Vision, and Action, 2022. 2

  36. [37]

    Progprompt: Generating situated robot task plans using large language models

    Ishika Singh, Valts Blukis, Arsalan Mousavian, Ankit Goyal, Danfei Xu, Jonathan Tremblay, Dieter Fox, Jesse Thoma- son, and Animesh Garg. Progprompt: Generating situated robot task plans using large language models.arXiv preprint arXiv:2209.11302, 2022. 2

  37. [38]

    Llm-planner: Few-shot grounded planning for embodied agents with large language models

    Chan Hee Song, Jiaman Wu, Clayton Washington, Brian M Sadler, Wei-Lun Chao, and Yu Su. Llm-planner: Few-shot grounded planning for embodied agents with large language models. InProceedings of the IEEE/CVF international con- ference on computer vision, pages 2998–3009, 2023. 2

  38. [39]

    Motionclip: Exposing human mo- tion generation to clip space

    Guy Tevet, Brian Gordon, Amir Hertz, Amit H Bermano, and Daniel Cohen-Or. Motionclip: Exposing human mo- tion generation to clip space. InComputer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, Octo- ber 23–27, 2022, Proceedings, Part XXII, pages 358–374. Springer, 2022. 1, 2, 6

  39. [40]

    Human motion diffu- sion model

    Guy Tevet, Sigal Raab, Brian Gordon, Yoni Shafir, Daniel Cohen-or, and Amit Haim Bermano. Human motion diffu- sion model. InThe Eleventh International Conference on Learning Representations, 2023. 1, 2, 6

  40. [41]

    CLoSD: Closing the loop between simulation and diffusion for multi-task character control

    Guy Tevet, Sigal Raab, Setareh Cohan, Daniele Reda, Zhengyi Luo, Xue Bin Peng, Amit Haim Bermano, and Michiel van de Panne. CLoSD: Closing the loop between simulation and diffusion for multi-task character control. In The Thirteenth International Conference on Learning Rep- resentations, 2025. 1

  41. [42]

    X-loco: To- wards generalist humanoid locomotion control via synergetic policy distillation.arXiv preprint arXiv:2603.03733, 2026

    Dewei Wang, Xinmiao Wang, Chenyun Zhang, Jiyuan Shi, Yingnan Zhao, Chenjia Bai, and Xuelong Li. X-loco: To- wards generalist humanoid locomotion control via synergetic policy distillation.arXiv preprint arXiv:2603.03733, 2026. 1

  42. [43]

    Halo: Closing sim-to- real gap for heavy-loaded humanoid agile motion skills via differentiable simulation.arXiv preprint arXiv:2603.15084,

    Xingyi Wang, Chenyun Zhang, Weiji Xie, Chao Yu, Wei Song, Chenjia Bai, and Shiqiang Zhu. Halo: Closing sim-to- real gap for heavy-loaded humanoid agile motion skills via differentiable simulation.arXiv preprint arXiv:2603.15084,

  43. [44]

    Scenegenagent: Precise industrial scene generation with coding agent

    Xiao Xia, Dan Zhang, Zibo Liao, Zhenyu Hou, Tianrui Sun, Jing Li, Ling Fu, and Yuxiao Dong. Scenegenagent: Precise industrial scene generation with coding agent. 2024. 2

  44. [45]

    Kungfubot: Physics-based humanoid whole-body control for learning highly-dynamic skills,

    Weiji Xie, Jinrui Han, Jiakun Zheng, Huanyu Li, Xinzhe Liu, Jiyuan Shi, Weinan Zhang, Chenjia Bai, and Xue- long Li. Kungfubot: Physics-based humanoid whole-body control for learning highly-dynamic skills.arXiv preprint arXiv:2506.12851, 2025. 8

  45. [46]

    arXiv: 2602.07439 [cs.RO]

    Weiji Xie, Jiakun Zheng, Jinrui Han, Jiyuan Shi, Weinan Zhang, Chenjia Bai, and Xuelong Li. Textop: Real-time in- teractive text-driven humanoid robot motion generation and control.arXiv preprint arXiv:2602.07439, 2026. 1

  46. [47]

    T2m-gpt: Generating human motion from textual de- scriptions with discrete representations

    Jianrong Zhang, Yangsong Zhang, Xiaodong Cun, Shaoli Huang, Yong Zhang, Hongwei Zhao, Hongtao Lu, and Xi Shen. T2m-gpt: Generating human motion from textual de- scriptions with discrete representations. InProceedings of 10 the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2023. 2

  47. [48]

    DartControl: A diffusion-based autoregressive motion model for real-time text-driven motion control

    Kaifeng Zhao, Gen Li, and Siyu Tang. DartControl: A diffusion-based autoregressive motion model for real-time text-driven motion control. InThe Thirteenth International Conference on Learning Representations (ICLR), 2025. 1

  48. [49]

    Navgpt: Explicit reasoning in vision-and-language navigation with large lan- guage models

    Gengze Zhou, Yicong Hong, and Qi Wu. Navgpt: Explicit reasoning in vision-and-language navigation with large lan- guage models. InProceedings of the AAAI Conference on Artificial Intelligence, pages 7641–7649, 2024. 2 11 Re2MoGen: Open-Vocabulary Motion Generation via LLM Reasoning and Physics-Aware Refinement Supplementary Material

  49. [50]

    Implementation Details 7.1

    A. Implementation Details 7.1. Hyperparameter In the LLM planning, we employDeepSeek-R1as the rea- soning model, which exhibits superior spatial understand- ing and reasoning capabilities. The CLIP model mentioned in the paper uses theCLIP-ViT-L/14model from Open- CLIP [5]. The VLM used in our experiments isQwen-VL- Max. The detailed hyperparameters invol...

  50. [51]

    The reasons are:

    To simplify the reasoning task, we do not ask the LLM to plan motions for full-body joints. The reasons are:

  51. [52]

    Therefore, we only make the LLM infer motions for5key joints and plan only the displacement of each keyframe relative to the previous one

    excessive data processing would significantly degrade the reasoning quality of the model; 2) the strong inter- dependencies among full-body joints make the reason- ing less tolerant to errors. Therefore, we only make the LLM infer motions for5key joints and plan only the displacement of each keyframe relative to the previous one

  52. [53]

    Forward-Backward

    We define a set of foundational information as follows: • Human skeleton information, derived from the SMPL [27] Neutral skeleton. • The displacement direction information is not repre- sented in XYZ coordinates, but in directional terms such as “Forward-Backward”, “Up-Down”, and “Left- Right”, which helps the LLM better grasp movement orientation. • The ...

  53. [54]

    We provide a formatted output template to facilitate information ex- traction and subsequent usage

    Additionally, we require the LLM to output the rea- soning behind its planned keyframes, encouraging more thorough deliberation during inference. We provide a formatted output template to facilitate information ex- traction and subsequent usage

  54. [55]

    We include two examples to achieve few-shot fine- tuning, enabling the LLM to quickly adapt to our task during inference. Fig. 7 shows these two examples, and Fig. 8 explains the reasons. With the above prompt template, the LLM’s capability in planning such a task can be significantly enhanced. 7.3. Motion Representations In our implementation, we utilize...

  55. [56]

    Experiments Details 8.1

    B. Experiments Details 8.1. LLM-planned Keyframes As shown in Figs.(10-16), we present the JSON data of keyframes planned by the LLM of different motion lengths alongside the rendered pose images after full-body opti- mization. These results demonstrate that LLM can reason- ably plan actions based on text descriptions. 8.2. Evaluation Metrics As mentioned...

  56. [57]

    F0":{"Pelvis

    C. Additional Experiments Table 5. Results on SnapMotion dataset. Methods CLIP S↑VLM S↑ MLD23.39±0.401.64±0.24 MotionGPT22.36±0.511.42±0.14 MoMask23.92±0.361.75±0.17 Ours24.68±0.642.17±0.40 2 You are an expert on Kinematics and Human Motion. You are tasked to plan the movement of the four end controllers (L_Ankle, R_Ankle, L_Wrist, R_Wrist) and the root j...