pith. machine review for the scientific record. sign in

arxiv: 2604.07335 · v1 · submitted 2026-04-08 · 💻 cs.RO

Recognition: unknown

TAMEn: Tactile-Aware Manipulation Engine for Closed-Loop Data Collection in Contact-Rich Tasks

Authors on Pith no claims yet

Pith reviewed 2026-05-10 17:35 UTC · model grok-4.3

classification 💻 cs.RO
keywords bimanual manipulationtactile sensingdata collectionvisuo-tactile learningrobot manipulationcontact-rich taskswearable interfacedemonstration learning
0
0 comments X

The pith

A cross-morphology wearable interface with dual-modal tracking collects tactile-rich data that raises bimanual manipulation success from 34 percent to 75 percent.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper introduces TAMEn to solve persistent problems in collecting usable demonstrations for contact-rich bimanual robot tasks. Existing handheld methods lack gripper flexibility, cannot check feasibility in real time, and miss authentic tactile signals during recovery from failures. TAMEn supplies a single wearable interface that adapts to different grippers, switches between motion-capture precision and VR portability, and feeds a pyramid of pretraining, demonstrations, and human-guided recovery data into policy training. If the collected tactile information is faithful, policies learn to manage physical contacts that vision alone cannot resolve. Experiments confirm better replayability of demonstrations and the reported jump in task completion rates.

Core claim

TAMEn is a tactile-aware manipulation engine that uses a cross-morphology wearable interface and a dual-modal acquisition pipeline of precision motion-capture mode plus portable VR-based mode. This hardware supports a pyramid-structured data regime that unifies large-scale tactile pretraining, task-specific bimanual demonstrations, and human-in-the-loop recovery data with visualized tactile feedback, enabling closed-loop policy refinement that improves demonstration replayability and lifts success rates from 34 percent to 75 percent across diverse bimanual manipulation tasks.

What carries the argument

The cross-morphology wearable interface together with the dual-modal precision and portable acquisition pipeline that supplies a pyramid-structured regime of tactile pretraining, demonstrations, and recovery data.

If this is right

  • The feasibility-aware pipeline produces demonstrations that replay more reliably on the robot.
  • Visuo-tactile policies trained under the pyramid regime reach 75 percent success on the tested contact-rich bimanual tasks.
  • Human-in-the-loop recovery sessions supply interactive data that refines policies beyond static demonstrations alone.
  • The open-sourced hardware and dataset allow direct reproduction and extension of the visuo-tactile collection method.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The same wearable hardware could be adapted to collect data for single-arm or multi-robot contact tasks where gripper morphology varies.
  • Portable VR mode might enable gathering of tactile demonstrations outside controlled lab spaces in more varied settings.
  • Adding recovery data with real tactile feedback may lower the volume of simulation pretraining required for contact-rich skills.
  • Similar closed-loop collection pipelines could be tested with other sensory streams such as force-torque or audio to check transfer of the pyramid structure.

Load-bearing premise

The wearable interface and dual-modal pipeline supply sufficiently authentic tactile signals during both high-precision demonstrations and human-in-the-loop recovery without adding artifacts that harm policy learning.

What would settle it

Training the same policy architecture on the collected demonstrations but without the tactile channel and observing success rates that stay at or below the 34 percent baseline on the reported bimanual tasks.

Figures

Figures reproduced from arXiv: 2604.07335 by Chenghang Jiang, Guoying Gu, Hongyang Li, Jieji Ren, Junxi Zhou, Li Chen, Longyan Wu, Ran Huang, Shijia Peng.

Figure 1
Figure 1. Figure 1: Introducing TAMEn, a Tactile-Aware Manipulation Engine for closed-loop data collection in contact-rich bimanual tasks, which builds upon the UMI paradigm with key enhancements in multimodality, precision-portability synergy, replayability, and data flywheel. (a) Wearable visuo-tactile interface captures rich multimodal data while breaking the precision-portability trade-off through a dual-mode pipeline tha… view at source ↗
Figure 2
Figure 2. Figure 2: Hardware system. Left: Structure of TAMEn. Right: two data collection modes, supporting high-precision demonstration collection and portable in-the-wild or recovery data acquisition. dataset to facilitate reproducibility and support research in visuo￾tactile manipulation. Index Terms—Visuo-Tactile, Handheld Interface, Closed-Loop Data Collection, Bimanual Manipulation. I. INTRODUCTION P HYSICAL interaction… view at source ↗
Figure 3
Figure 3. Figure 3: Mechanisms of the proposed handheld gripper interface. (a) Flexion–extension gripper. (b) Parallel-jaw gripper. Left: Overall view of the interface. Right: Kinematic schematic with key geometric parameters. in unstructured environments. In the motion-capture mode, the interface is tracked by the NOKOV system, enabling sub-millimeter pose tracking. The markers are arranged in a structured layout, among whic… view at source ↗
Figure 4
Figure 4. Figure 4: Pyramid-structured visuo-tactile learning framework. Large-scale single-arm visuo-tactile data provide broad priors for pretraining, task-specific bimanual data support coordination-aware fine-tuning, and recovery data further refine the policy around realistic failure states. embedding at the current timestep and an additional temporal positive from the next timestep. The loss is written as: Lcon = − 1 B … view at source ↗
Figure 5
Figure 5. Figure 5: Robot setup. A dual-arm platform equipped with wrist￾mounted cameras and fingertip visuo-tactile sensors. in [PITH_FULL_IMAGE:figures/full_fig_p006_5.png] view at source ↗
Figure 6
Figure 6. Figure 6: Trajectory visualization. We test TAMEn on a variety of contact-rich tasks. to a folder, detaches it, and moves it toward the drawer, while the right arm opens the drawer and closes it after the clip is placed inside. Successful execution requires firm yet controlled interaction with the spring-loaded clip, as tactile feedback helps stabilize the grasp and judge detachment, followed by coordinated sequenti… view at source ↗
Figure 7
Figure 7. Figure 7: Tracking accuracy. Trajectory errors of VR-based tracking and GoPro-based SLAM tracking relative to the motion-capture reference. TABLE II: Data validity across collection settings. The table reports replay success rates (%). Method Herbal Transfer Cable Mounting Avg. No Feasibility Screening 39 12 26 Online Validation (Ours) 100 100 100 with improved stability, while maintaining trajectory errors within 1… view at source ↗
Figure 8
Figure 8. Figure 8: Generalization. TAMEn generalizes to unseen objects with representative robust and fragile cases shown. replayable on robots and organizes heterogeneous data for staged learning. Furthermore, TAMEn establishes a closed￾loop data flywheel through AR-based teleoperation with tactile feedback (tAmeR), enabling recovery-oriented data collection and continuous policy refinement under realistic failure states. E… view at source ↗
Figure 9
Figure 9. Figure 9: Robustness. Under visual disturbances, TAMEn exhibits improved robustness during contact-rich execution, with representative robust and fragile cases shown. REFERENCES [1] C. Li, C. Liu, D. Wang, S. Zhang, L. Li, Z. Zeng, F. Liu, J. Xu, and R. Chen, “Vitamin-b: A reliable and efficient visuo-tactile bimanual manipulation interface,” arXiv preprint arXiv:2511.05858, 2025. [2] J. Hu, D. Jones, M. R. Dogar, a… view at source ↗
Figure 11
Figure 11. Figure 11: Compatibility with multiple tactile sensors. TAMEn supports seamless integration of different tactile sensors, includ￾ing GelSight, Xense, DW-Tac, PaXini, and ours, demonstrating its adaptability across heterogeneous sensing modalities. ① Precision Mode ② Portable Mode Right Left Right Left R5R4R3 R2 R1 L4 L3 L2 L1 Flange Flange Flange Flange [PITH_FULL_IMAGE:figures/full_fig_p013_11.png] view at source ↗
Figure 12
Figure 12. Figure 12: Local frame construction and unified flange representation. In the precision mode, a local frame is constructed from the marker configuration on each collector, while in the portable mode, the VR handle provides the tracked pose. Both are mapped to a shared flange-based reference frame for consistent pose representation [PITH_FULL_IMAGE:figures/full_fig_p013_12.png] view at source ↗
Figure 13
Figure 13. Figure 13: In-the-wild visuo-tactile data collection. The portable configuration of TAMEn enables data acquisition across diverse real-world scenes. configuration, we record synchronized visual, tactile, and motion data for high-fidelity demonstration collection. Visual observations are captured using a fisheye camera equipped with a 180◦ field-of-view lens at 30 FPS and a resolution of 640 × 480. The visuo-tactile … view at source ↗
Figure 14
Figure 14. Figure 14: Visualization of tAmeR. Wrist-mounted RGB and tactile streams are visualized above the scene. The interface supports multiple visuo-tactile sensors, where (a) GelSight, (b) DW-Tac, (c) Xense, and (d) our sensor show the tactile stream from the left gripper. adjustments. In our implementation, the maximum joint velocity is set to 180◦/s, and the TCP velocity limit is set to 250, mm/s. The joint soft limits… view at source ↗
Figure 15
Figure 15. Figure 15: Dish washing from data collection to robot execution. Demonstrations collected with proxy materials are successfully transferred to real cleaning scenarios, enabling stable contact and effective wiping on real stains. Policy architecture. Our downstream policy follows an ACT￾style transformer architecture. Visual and tactile observations are encoded separately using ResNet-18 backbones. Their features are… view at source ↗
read the original abstract

Handheld paradigms offer an efficient and intuitive way for collecting large-scale demonstration of robot manipulation. However, achieving contact-rich bimanual manipulation through these methods remains a pivotal challenge, which is substantially hindered by hardware adaptability and data efficacy. Prior hardware designs remain gripper-specific and often face a trade-off between tracking precision and portability. Furthermore, the lack of online feasibility checking during demonstration leads to poor replayability. More importantly, existing handheld setups struggle to collect interactive recovery data during robot execution, lacking the authentic tactile information necessary for robust policy refinement. To bridge these gaps, we present TAMEn, a tactile-aware manipulation engine for closed-loop data collection in contact-rich tasks. Our system features a cross-morphology wearable interface that enables rapid adaptation across heterogeneous grippers. To balance data quality and environmental diversity, we implement a dual-modal acquisition pipeline: a precision mode leveraging motion capture for high-fidelity demonstrations, and a portable mode utilizing VR-based tracking for in-the-wild acquisition and tactile-visualized recovery teleoperation. Building on this hardware, we unify large-scale tactile pretraining, task-specific bimanual demonstrations, and human-in-the-loop recovery data into a pyramid-structured data regime, enabling closed-loop policy refinement. Experiments show that our feasibility-aware pipeline significantly improves demonstration replayability, and that the proposed visuo-tactile learning framework increases task success rates from 34% to 75% across diverse bimanual manipulation tasks. We further open-source the hardware and dataset to facilitate reproducibility and support research in visuo-tactile manipulation.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The manuscript introduces TAMEn, a tactile-aware manipulation engine for closed-loop data collection in contact-rich bimanual tasks. It proposes a cross-morphology wearable interface for rapid gripper adaptation, a dual-modal pipeline (precision mode with motion capture for high-fidelity data and portable mode with VR for in-the-wild collection and tactile-visualized human-in-the-loop recovery), and a pyramid-structured data regime that combines large-scale tactile pretraining, task demonstrations, and recovery data for policy refinement. The authors claim this yields significantly better demonstration replayability and raises task success rates from 34% to 75% across diverse bimanual tasks, while open-sourcing the hardware and dataset.

Significance. If the empirical outcomes are substantiated, the work could provide a practical advance in scalable data collection for visuo-tactile robot learning by solving hardware adaptability and enabling authentic interactive recovery data. The open-sourcing of hardware and dataset is a clear strength that supports reproducibility and further research in contact-rich manipulation.

major comments (2)
  1. [Abstract] Abstract: the central claim that the visuo-tactile framework increases success rates from 34% to 75% is presented without any experimental details on trial counts, baselines, statistical tests, variance, or error analysis, preventing evaluation of whether the reported gain is robust or reproducible.
  2. [Hardware and data pipeline sections] Hardware and data pipeline sections: the load-bearing assumption that the cross-morphology wearable and dual-modal (mocap/VR) streams deliver sufficiently authentic contact forces and slip events for policy training is not supported by any quantitative validation such as force RMSE against calibrated robot sensors, latency measurements, or cross-gripper calibration residuals.
minor comments (2)
  1. [Data regime description] The description of the pyramid-structured data regime would benefit from a diagram or explicit breakdown of data volumes and weighting at each level to clarify how pretraining, demonstrations, and recovery interact.
  2. [Figures] Ensure all figures showing the wearable interface include scale bars and clear labels for sensor placement to aid hardware replication.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the detailed and constructive review. We address the two major comments point by point below, indicating the revisions we will incorporate to improve clarity and substantiation of our claims.

read point-by-point responses
  1. Referee: [Abstract] Abstract: the central claim that the visuo-tactile framework increases success rates from 34% to 75% is presented without any experimental details on trial counts, baselines, statistical tests, variance, or error analysis, preventing evaluation of whether the reported gain is robust or reproducible.

    Authors: We agree that the abstract, being a high-level summary, omits the granular experimental statistics that appear in the Experiments section. To address this, we will revise the abstract to include key details such as the number of trials conducted across tasks, the specific baselines used for comparison, and references to the statistical analysis and variance reported in the body of the paper. This will make the central claim more self-contained while preserving the abstract's brevity. revision: yes

  2. Referee: [Hardware and data pipeline sections] Hardware and data pipeline sections: the load-bearing assumption that the cross-morphology wearable and dual-modal (mocap/VR) streams deliver sufficiently authentic contact forces and slip events for policy training is not supported by any quantitative validation such as force RMSE against calibrated robot sensors, latency measurements, or cross-gripper calibration residuals.

    Authors: We acknowledge that the manuscript does not include explicit quantitative validation metrics (force RMSE, latency, or cross-gripper residuals) in the hardware and pipeline sections to directly support the authenticity of captured contact forces and slip events. This is a fair observation. We will add these validations in the revised manuscript, including force sensing accuracy comparisons, system latency measurements for both modes, and calibration results across gripper morphologies, to better substantiate the data quality assumptions. revision: yes

Circularity Check

0 steps flagged

No derivation chain present; empirical system description with measured outcomes

full rationale

The manuscript describes a hardware-software pipeline for visuo-tactile data collection and reports empirical task success rates (34% to 75%). No equations, first-principles derivations, fitted parameters renamed as predictions, or load-bearing self-citations appear in the provided text. The central claim is an observed performance delta from controlled experiments, not a mathematical reduction to prior inputs. Per the hard rules, absence of any quotable derivation step that reduces to its own inputs yields score 0.

Axiom & Free-Parameter Ledger

0 free parameters · 0 axioms · 0 invented entities

No free parameters, axioms, or invented physical entities are present; the paper describes an applied robotics system.

pith-pipeline@v0.9.0 · 5613 in / 1052 out tokens · 67184 ms · 2026-05-10T17:35:36.776211+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

60 extracted references · 35 canonical work pages · 6 internal anchors

  1. [1]

    Vitamin-b: A reliable and efficient visuo-tactile bimanual manipulation interface,

    C. Li, C. Liu, D. Wang, S. Zhang, L. Li, Z. Zeng, F. Liu, J. Xu, and R. Chen, “Vitamin-b: A reliable and efficient visuo-tactile bimanual manipulation interface,”arXiv preprint arXiv:2511.05858, 2025

  2. [2]

    Occlusion-robust autonomous robotic manipulation of human soft tissues with 3-d surface feedback,

    J. Hu, D. Jones, M. R. Dogar, and P. Valdastri, “Occlusion-robust autonomous robotic manipulation of human soft tissues with 3-d surface feedback,”TRO, 2023

  3. [3]

    Goal-conditioned dual-action imitation learning for dexterous dual-arm robot manipulation,

    H. Kim, Y . Ohmura, and Y . Kuniyoshi, “Goal-conditioned dual-action imitation learning for dexterous dual-arm robot manipulation,”TRO, 2024

  4. [4]

    Evetac: An event-based optical tactile sensor for robotic manipulation,

    N. Funk, E. Helmut, G. Chalvatzaki, R. Calandra, and J. Peters, “Evetac: An event-based optical tactile sensor for robotic manipulation,”TRO, 2024

  5. [5]

    Skillvla: Tackling combinatorial diversity in dual-arm manipulation via skill reuse,

    X. Zhai, Z. Huang, L. Wu, Q. Zhao, Q. Yu, J. Ren, C. Hao, and H. Soh, “Skillvla: Tackling combinatorial diversity in dual-arm manipulation via skill reuse,”arXiv preprint arXiv:2603.03836, 2026

  6. [6]

    RISE: Self-Improving Robot Policy with Compositional World Model

    J. Yang, K. Lin, J. Li, W. Zhang, T. Lin, L. Wu, Z. Su, H. Zhao, Y .-Q. Zhang, L. Chen, P. Luo, X. Yue, and H. Li, “Rise: Self- improving robot policy with compositional world model,”arXiv preprint arXiv:2602.11075, 2026

  7. [7]

    π0.5: a vision-language-action model with open-world generalization,

    P. Intelligence, K. Black, N. Brown, J. Darpinian, K. Dhabalia, D. Driess, A. Esmail, M. Equi, C. Finn, N. Fusai, M. Y . Galliker, D. Ghosh, L. Groom, K. Hausman, B. Ichter, S. Jakubczak, T. Jones, L. Ke, D. LeBlanc, S. Levine, A. Li-Bell, M. Mothukuri, S. Nair, K. Pertsch, A. Z. Ren, L. X. Shi, L. Smith, J. T. Springenberg, K. Stachowicz, J. Tanner, Q. V...

  8. [8]

    Rotipbot: Robotic handling of thin and flexible objects using rotatable tactile sensors,

    J. Jiang, X. Zhang, D. F. Gomes, T.-T. Do, and S. Luo, “Rotipbot: Robotic handling of thin and flexible objects using rotatable tactile sensors,”TRO, 2025

  9. [9]

    AnyTouch 2: General optical tactile rep- resentation learning for dynamic tactile perception, 2026

    R. Feng, Y . Zhou, S. Mei, D. Zhou, P. Wang, S. Cui, B. Fang, G. Yao, and D. Hu, “Anytouch 2: General optical tactile representation learning for dynamic tactile perception,”arXiv preprint arXiv:2602.09617, 2026

  10. [10]

    Egomimic: Scaling imitation learning via egocentric video,

    S. Kareer, D. Patel, R. Punamiya, P. Mathur, S. Cheng, C. Wang, J. Hoffman, and D. Xu, “Egomimic: Scaling imitation learning via egocentric video,” inICRA, 2025

  11. [11]

    EgoDex: Learning Dexterous Manipulation from Large-Scale Egocentric Video

    R. Hoque, P. Huang, D. J. Yoon, M. Sivapurapu, and J. Zhang, “Egodex: Learning dexterous manipulation from large-scale egocentric video,” arXiv preprint arXiv:2505.11709, 2026

  12. [12]

    arXiv preprint arXiv:2602.10106 (2026) 5

    M. Shi, S. Peng, J. Chen, H. Jiang, Y . Li, D. Huang, P. Luo, H. Li, and L. Chen, “Egohumanoid: Unlocking in-the-wild loco-manipulation with robot-free egocentric demonstration,”arXiv preprint arXiv:2602.10106, 2026

  13. [13]

    Re- active diffusion policy: Slow-fast visual-tactile policy learning for contact-rich manipulation.arXiv preprint arXiv:2503.02881, 2025

    H. Xue, J. Ren, W. Chen, G. Zhang, Y . Fang, G. Gu, H. Xu, and C. Lu, “Reactive diffusion policy: Slow-fast visual-tactile policy learning for contact-rich manipulation,”arXiv preprint arXiv:2503.02881, 2025

  14. [14]

    Unibidex: A unified teleoperation framework for robotic bimanual dexterous manipulation,

    Z. Li, Z. Guo, J. Hu, D. Navarro-Alarcon, J. Pan, H. Wu, and P. Zhou, “Unibidex: A unified teleoperation framework for robotic bimanual dexterous manipulation,”arXiv preprint arXiv:2601.04629, 2026

  15. [15]

    Dart: Dexterous augmented reality teleoperation platform for large-scale robot data collection in simulation,

    Y . Park, J. S. Bhatia, L. Ankile, and P. Agrawal, “Dart: Dexterous augmented reality teleoperation platform for large-scale robot data collection in simulation,” inICRA, 2025

  16. [16]

    Universal Manipulation Interface: In-The-Wild Robot Teaching Without In-The-Wild Robots

    C. Chi, Z. Xu, C. Pan, E. Cousineau, B. Burchfiel, S. Feng, R. Tedrake, and S. Song, “Universal manipulation interface: In-the-wild robot teaching without in-the-wild robots,”arXiv preprint arXiv:2402.10329, 2024

  17. [17]

    Umi on legs: Making manipulation policies mobile with manipulation-centric whole-body controllers, 2024

    H. Ha, Y . Gao, Z. Fu, J. Tan, and S. Song, “Umi on legs: Making manipulation policies mobile with manipulation-centric whole-body controllers,”arXiv preprint arXiv:2407.10353, 2024

  18. [18]

    Mv-umi: A scalable multi- view interface for cross-embodiment learning, 2025

    O. Rayyan, J. Abanes, M. Hafez, A. Tzes, and F. Abu-Dakka, “Mv-umi: A scalable multi-view interface for cross-embodiment learning,”arXiv preprint arXiv:2509.18757, 2025

  19. [19]

    arXiv preprint arXiv:2511.00153 , year=

    J. Yu, Y . Shentu, D. Wu, P. Abbeel, K. Goldberg, and P. Wu, “Egomi: Learning active vision and whole-body manipulation from egocentric human demonstrations,”arXiv preprint arXiv:2511.00153, 2026

  20. [20]

    Fastumi: A scalable and hardware- independent universal manipulation interface.arXiv preprint arXiv:2409.19499, 2024

    Zhaxizhuoma, K. Liu, C. Guan, Z. Jia, Z. Wu, X. Liu, T. Wang, S. Liang, P. Chen, P. Zhang, H. Song, D. Qu, D. Wang, Z. Wang, N. Cao, Y . Ding, B. Zhao, and X. Li, “Fastumi: A scalable and hardware- independent universal manipulation interface with dataset,”arXiv preprint arXiv:2409.19499, 2025

  21. [21]

    Legato: Cross- embodiment imitation using a grasping tool,

    M. Seo, H. A. Park, S. Yuan, Y . Zhu, and L. Sentis, “Legato: Cross- embodiment imitation using a grasping tool,”RAL, 2025

  22. [22]

    Umigen: A unified framework for egocentric point cloud generation and cross-embodiment robotic imitation learn- ing.arXiv preprint arXiv:2511.09302, 2025

    Y . Huang, S. Li, X. Li, and W. Ding, “Umigen: A unified framework for egocentric point cloud generation and cross-embodiment robotic imitation learning,”arXiv preprint arXiv:2511.09302, 2025

  23. [23]

    Tactile-conditioned diffusion policy for force-aware robotic manipulation, 2025

    E. Helmut, N. Funk, T. Schneider, C. de Farias, and J. Peters, “Tactile- conditioned diffusion policy for force-aware robotic manipulation,”arXiv preprint arXiv:2510.13324, 2025

  24. [24]

    Can vision feel touch? tactile- aware visual grasping for transparent objects,

    L. Tong, K. Qian, Z. Yue, and S. Luo, “Can vision feel touch? tactile- aware visual grasping for transparent objects,”TCSVT, 2026

  25. [25]

    Manip- force: Force-guided policy learning with frequency-aware representa- tion for contact-rich manipulation,

    G. Lee, Y . Lee, K. Kim, S. Lee, S. Noh, S. Back, and K. Lee, “Manipforce: Force-guided policy learning with frequency-aware representation for contact-rich manipulation,”arXiv preprint arXiv:2509.19047, 2025

  26. [26]

    Simultaneous tactile-visual perception for learning multimodal robot manipulation.arXiv preprint arXiv:2512.09851, 2025

    Y . Li, Y . Chen, Z. Zhao, P. Li, T. Liu, S. Huang, and Y . Zhu, 12 “Simultaneous tactile-visual perception for learning multimodal robot manipulation,”arXiv preprint arXiv:2512.09851, 2026

  27. [27]

    Dexgrasp-zero: A morphology-aligned policy for zero-shot cross- embodiment dexterous grasping,

    Y . Wu, Y . Lin, W. Lao, Y . Lin, Y .-L. Wei, W.-S. Zheng, and A. Wu, “Dexgrasp-zero: A morphology-aligned policy for zero-shot cross- embodiment dexterous grasping,”arXiv preprint arXiv:2603.16806, 2026

  28. [28]

    Xgrasp: Gripper-aware grasp detection with multi-gripper data generation,

    Y . Lee, J. Mun, H. Shin, G. Hwang, J. Nam, T. Lee, and S. Jo, “Xgrasp: Gripper-aware grasp detection with multi-gripper data generation,”arXiv preprint arXiv:2510.11036, 2026

  29. [29]

    ARCap: Collecting high-quality human demonstrations for robot learning with augmented reality feedback,

    S. Chen, C. Wang, K. Nguyen, L. Fei-Fei, and C. K. Liu, “ARCap: Collecting high-quality human demonstrations for robot learning with augmented reality feedback,” inICRA, 2025

  30. [30]

    Robopocket: Improve robot policies instantly with your phone.arXiv preprint arXiv:2603.05504, 2026

    J. Fang, W. Chen, H. Xue, F. Zhou, T. Le, Y . Wang, Y . Zhang, J. Lv, C. Wen, and C. Lu, “Robopocket: Improve robot policies instantly with your phone,”arXiv preprint arXiv:2603.05504, 2026

  31. [31]

    Clear- mp: Clearance learning-based efficient motion planning for dual-arm robots under end-effector orientation constraints,

    B. Chen, H. Zhang, K. Li, Y . Fan, Y . Jiang, C. Yang, and Y . Wang, “Clear- mp: Clearance learning-based efficient motion planning for dual-arm robots under end-effector orientation constraints,”TASE, 2026

  32. [32]

    Compliant residual dagger: Im- proving real-world contact-rich manipulation with human corrections,

    X. Xu, Y . Hou, C. Xin, Z. Liu, and S. Song, “Compliant residual dagger: Improving real-world contact-rich manipulation with human corrections,” arXiv preprint arXiv:2506.16685, 2025

  33. [33]

    Racer: Rich language-guided failure recovery policies for imitation learning,

    Y . Dai, J. Lee, N. Fazeli, and J. Chai, “Racer: Rich language-guided failure recovery policies for imitation learning,” inICRA, 2025

  34. [34]

    Rac: Robot learning for long-horizon tasks by scaling recovery and correction.arXiv preprint arXiv:2509.07953, 2025

    Z. Hu, R. Wu, N. Enock, J. Li, R. Kadakia, Z. Erickson, and A. Kumar, “Rac: Robot learning for long-horizon tasks by scaling recovery and correction,”arXiv preprint arXiv:2509.07953, 2025

  35. [35]

    Dexhil: A human-in-the-loop framework for vision-language- action model post-training in dexterous manipulation,

    Y . Han, Z. Chen, Y . Zhao, C. Xu, Y . Shao, Y . Peng, Y . Mu, and W. Lian, “Dexhil: A human-in-the-loop framework for vision-language- action model post-training in dexterous manipulation,”arXiv preprint arXiv:2603.09121, 2026

  36. [36]

    AgiBot World Colosseo: A Large-scale Manipulation Platform for Scalable and Intelligent Embodied Systems

    Q. Bu, J. Cai, L. Chen, X. Cui, Y . Ding, S. Feng, S. Gao, X. He, X. Hu, X. Huanget al., “Agibot world colosseo: A large-scale manipulation platform for scalable and intelligent embodied systems,”arXiv preprint arXiv:2503.06669, 2025

  37. [37]

    A humanoid visual-tactile-action dataset for contact-rich manipulation,

    E. Kwon, S. Oh, I.-C. Baek, Y . Park, G. Kim, J. Moon, Y . Choi, and K.-J. Kim, “A humanoid visual-tactile-action dataset for contact-rich manipulation,” 2025

  38. [38]

    Mobile aloha: Learning bimanual mobile manipulation with low-cost whole-body teleoperation,

    Z. Fu, T. Z. Zhao, and C. Finn, “Mobile aloha: Learning bimanual mobile manipulation with low-cost whole-body teleoperation,” 2024

  39. [39]

    Lemmo-plan: Llm-enhanced learning from multi-modal demonstration for planning sequential contact-rich manipulation tasks,

    K. Chen, Z. Shen, Y . Zhang, L. Chen, F. Wu, Z. Bing, S. Haddadin, and A. Knoll, “Lemmo-plan: Llm-enhanced learning from multi-modal demonstration for planning sequential contact-rich manipulation tasks,” 2025

  40. [40]

    GEN-0: Embodied foundation models that scale with physical interaction,

    Generalist AI Team, “GEN-0: Embodied foundation models that scale with physical interaction,” https://generalistai.com/blog/ preview-uqlxvb-bb.html, 2025

  41. [41]

    Freetacman: Robot-free visuo-tactile data collection system for contact-rich manipulation.arXiv preprint arXiv:2506.01941,

    L. Wu, C. Yu, J. Ren, L. Chen, Y . Jiang, R. Huang, G. Gu, and H. Li, “Freetacman: Robot-free visuo-tactile data collection system for contact- rich manipulation,”arXiv preprint arXiv:2506.01941, 2025

  42. [42]

    Fastumi-100k: Advancing data-driven robotic manipulation with a large-scale umi-style dataset,

    K. Liu, Z. Jia, Y . Li, Zhaxizhuoma, P. Chen, S. Liu, X. Liu, P. Zhang, H. Song, X. Ye, N. Cao, Z. Wang, J. Zeng, D. Wang, Y . Ding, B. Zhao, and X. Li, “Fastumi-100k: Advancing data-driven robotic manipulation with a large-scale umi-style dataset,” 2025

  43. [43]

    Touch in the wild: Learning fine-grained manipulation with a portable visuo-tactile gripper,

    X. Zhu, B. Huang, and Y . Li, “Touch in the wild: Learning fine-grained manipulation with a portable visuo-tactile gripper,” 2025

  44. [44]

    In-the-wild compliant manipulation with umi-ft.arXiv preprint arXiv:2601.09988, 2026

    H. Choi, Y . Hou, C. Pan, S. Hong, A. Patel, X. Xu, M. R. Cutkosky, and S. Song, “In-the-wild compliant manipulation with umi-ft,”arXiv preprint arXiv:2601.09988, 2026

  45. [45]

    Tacumi: A multi-modal universal manipulation interface for contact-rich tasks.arXiv preprint arXiv:2601.14550, 2026

    T. Cheng, K. Chen, L. Chen, L. Zhang, Y . Zhang, Y . Ling, M. Hamad, Z. Bing, F. Wu, K. Sharmaet al., “Tacumi: A multi-modal uni- versal manipulation interface for contact-rich tasks,”arXiv preprint arXiv:2601.14550, 2026

  46. [46]

    MC-Tac: Modular camera-based tactile sensor for robot gripper,

    J. Ren, J. Zou, and G. Gu, “MC-Tac: Modular camera-based tactile sensor for robot gripper,” inICIRA, 2023

  47. [47]

    Omnivta: Visuo-tactile world modeling for contact- rich robotic manipulation, 2026

    Y . Zheng, S. Gu, W. Li, Y . Zheng, Y . Zang, S. Tian, X. Li, C. Hao, C. Gao, S. Liu, H. Li, Y . Chen, S. Yan, and W. Ding, “Omnivta: Visuo-tactile world modeling for contact-rich robotic manipulation,”arXiv preprint arXiv:2603.19201, 2026

  48. [48]

    TouchGuide: Inference-Time Steering of Visuomotor Policies via Touch Guidance

    Z. Zhang, J. Ma, X. Yang, X. Wen, Y . Zhang, B. Li, Y . Qin, J. Liu, C. Zhao, L. Kanget al., “Touchguide: Inference-time steering of visuo- motor policies via touch guidance,”arXiv preprint arXiv:2601.20239, 2026

  49. [49]

    Vitamin: Learning contact-rich tasks through robot-free visuo-tactile manipulation interface,

    F. Liu, C. Li, Y . Qin, J. Xu, P. Abbeel, and R. Chen, “Vitamin: Learning contact-rich tasks through robot-free visuo-tactile manipulation interface,” 2025

  50. [50]

    exumi: Extensible robot teaching system with action-aware task-agnostic tactile representation, 2025

    Y . Xu, L. Wei, P. An, Q. Zhang, and Y .-L. Li, “exumi: Extensible robot teaching system with action-aware task-agnostic tactile representation,” arXiv preprint arXiv:2509.14688, 2025

  51. [51]

    Force-Aware Residual DAgger via Trajectory Editing for Precision Insertion with Impedance Control

    Y . Huang, M. Ning, W. Zhao, Z. Liu, J. Sun, Q. Wang, and Y . Chen, “Force-aware residual dagger via trajectory editing for precision insertion with impedance control,”arXiv preprint arXiv:2603.04038, 2026

  52. [52]

    A reduction of imitation learning and structured prediction to no-regret online learning,

    S. Ross, G. Gordon, and D. Bagnell, “A reduction of imitation learning and structured prediction to no-regret online learning,” inAISTATS, 2011

  53. [53]

    Diffusion policy: Visuomotor policy learning via action diffusion,

    C. Chi, Z. Xu, S. Feng, E. Cousineau, Y . Du, B. Burchfiel, R. Tedrake, and S. Song, “Diffusion policy: Visuomotor policy learning via action diffusion,” 2024

  54. [54]

    Learning from interventions,

    J. Spencer, S. Choudhury, M. Barnes, M. Schmittle, M. Chiang, P. Ramadge, and S. Srinivasa, “Learning from interventions,” inRSS, 2020

  55. [55]

    χ0: Resource-aware robust manipulation via taming distributional inconsistencies,

    C. Yu, C. Sima, G. Jiang, H. Zhang, H. Mai, H. Li, H. Wang, J. Chen, K. Wu, L. Chen, L. Zhao, M. Shi, P. Luo, Q. Bu, S. Peng, T. Li, and Y . Yuan, “ χ0: Resource-aware robust manipulation via taming distributional inconsistencies,”arXiv preprint arXiv:2602.09021, 2026

  56. [56]

    Robocopilot: Human-in-the-loop interactive imitation learning for robot manipulation.arXiv preprint arXiv:2503.07771, 2025

    P. Wu, Y . Shentu, Q. Liao, D. Jin, M. Guo, K. Sreenath, X. Lin, and P. Abbeel, “Robocopilot: Human-in-the-loop interactive imitation learning for robot manipulation,”arXiv preprint arXiv:2503.07771, 2025

  57. [57]

    Conrft: A reinforced fine-tuning method for vla models via consistency policy,

    Y . Chen, S. Tian, S. Liu, Y . Zhou, H. Li, and D. Zhao, “Conrft: A reinforced fine-tuning method for vla models via consistency policy,” 2025

  58. [58]

    Openmocap: Rethinking optical motion capture under real-world occlusion,

    C. Qian, D. Li, X. Yu, Z. Yang, and Q. Ma, “Openmocap: Rethinking optical motion capture under real-world occlusion,”arXiv preprint arXiv:2508.12610, 2025

  59. [59]

    Rapid: Reconfigurable, adaptive platform for iterative design,

    Z. Yin, F. Li, S. Zheng, and J. Liu, “Rapid: Reconfigurable, adaptive platform for iterative design,”arXiv preprint arXiv:2602.06653, 2026

  60. [60]

    Learning fine-grained bimanual manipulation with low-cost hardware,

    T. Z. Zhao, V . Kumar, S. Levine, and C. Finn, “Learning fine-grained bimanual manipulation with low-cost hardware,” inRSS, 2023. 13 APPENDIX A. Hardware Design | Supplement to Sec. III-B in the Main paper. Manufacturing and assembly details.Figure 10 shows the exploded view of TAMEn. The structural components are fabricated using 3D printing, enabling ra...