pith. machine review for the scientific record. sign in

arxiv: 1710.00489 · v1 · submitted 2017-10-02 · 💻 cs.RO · cs.AI· cs.CV· cs.NE· cs.SY

Recognition: unknown

SE3-Pose-Nets: Structured Deep Dynamics Models for Visuomotor Planning and Control

Authors on Pith no claims yet
classification 💻 cs.RO cs.AIcs.CVcs.NEcs.SY
keywords controldeepdynamicsmodelposescenespacestructured
0
0 comments X
read the original abstract

In this work, we present an approach to deep visuomotor control using structured deep dynamics models. Our deep dynamics model, a variant of SE3-Nets, learns a low-dimensional pose embedding for visuomotor control via an encoder-decoder structure. Unlike prior work, our dynamics model is structured: given an input scene, our network explicitly learns to segment salient parts and predict their pose-embedding along with their motion modeled as a change in the pose space due to the applied actions. We train our model using a pair of point clouds separated by an action and show that given supervision only in the form of point-wise data associations between the frames our network is able to learn a meaningful segmentation of the scene along with consistent poses. We further show that our model can be used for closed-loop control directly in the learned low-dimensional pose space, where the actions are computed by minimizing error in the pose space using gradient-based methods, similar to traditional model-based control. We present results on controlling a Baxter robot from raw depth data in simulation and in the real world and compare against two baseline deep networks. Our method runs in real-time, achieves good prediction of scene dynamics and outperforms the baseline methods on multiple control runs. Video results can be found at: https://rse-lab.cs.washington.edu/se3-structured-deep-ctrl/

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.