pith. machine review for the scientific record. sign in

arxiv: 1802.01557 · v1 · submitted 2018-02-05 · 💻 cs.LG · cs.AI· cs.CV· cs.RO

Recognition: unknown

One-Shot Imitation from Observing Humans via Domain-Adaptive Meta-Learning

Authors on Pith no claims yet
classification 💻 cs.LG cs.AIcs.CVcs.RO
keywords humanrobotvideolearningmeta-learningpriordemonstrationhumans
0
0 comments X
read the original abstract

Humans and animals are capable of learning a new behavior by observing others perform the skill just once. We consider the problem of allowing a robot to do the same -- learning from a raw video pixels of a human, even when there is substantial domain shift in the perspective, environment, and embodiment between the robot and the observed human. Prior approaches to this problem have hand-specified how human and robot actions correspond and often relied on explicit human pose detection systems. In this work, we present an approach for one-shot learning from a video of a human by using human and robot demonstration data from a variety of previous tasks to build up prior knowledge through meta-learning. Then, combining this prior knowledge and only a single video demonstration from a human, the robot can perform the task that the human demonstrated. We show experiments on both a PR2 arm and a Sawyer arm, demonstrating that after meta-learning, the robot can learn to place, push, and pick-and-place new objects using just one video of a human performing the manipulation.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 2 Pith papers

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Mobile ALOHA: Learning Bimanual Mobile Manipulation with Low-Cost Whole-Body Teleoperation

    cs.RO 2024-01 conditional novelty 6.0

    A low-cost whole-body teleoperation system enables effective imitation learning for complex bimanual mobile manipulation by co-training on mobile and static demonstration datasets.

  2. Bridge Data: Boosting Generalization of Robotic Skills with Cross-Domain Datasets

    cs.RO 2021-09 accept novelty 6.0

    A large multi-task multi-domain robot dataset combined with 50 new demonstrations yields 2x higher success rates on never-before-seen tasks in new domains.