Learning to Fill the Seam by Vision: Sub-millimeter Peg-in-hole on Unseen Shapes in Real World

04/16/2022
by   Liang Xie, et al.
0

In the peg insertion task, human pays attention to the seam between the peg and the hole and tries to fill it continuously with visual feedback. By imitating the human behavior, we design architectures with position and orientation estimators based on the seam representation for pose alignment, which proves to be general to the unseen peg geometries. By putting the estimators into the closed-loop control with reinforcement learning, we further achieve a higher or comparable success rate, efficiency, and robustness compared with the baseline methods. The policy is trained totally in simulation without any manual intervention. To achieve sim-to-real, a learnable segmentation module with automatic data collecting and labeling can be easily trained to decouple the perception and the policy, which helps the model trained in simulation quickly adapt to the real world with negligible effort. Results are presented in simulation and on a physical robot. Code, videos, and supplemental material are available at https://github.com/xieliang555/SFN.git

READ FULL TEXT

page 3

page 5

page 6

research
05/09/2022

Learning A Simulation-based Visual Policy for Real-world Peg In Unseen Holes

This paper proposes a learning-based visual peg-in-hole that enables tra...
research
03/07/2023

End-to-End Deep Visual Control for Mastering Needle-Picking Skills With World Models and Behavior Cloning

Needle picking is a challenging surgical task in robot-assisted surgery ...
research
08/22/2021

Transferring Dexterous Manipulation from GPU Simulation to a Remote Real-World TriFinger

We present a system for learning a challenging dexterous manipulation ta...
research
10/29/2020

A Framework for Learning Predator-prey Agents from Simulation to Real World

In this paper, we propose an evolutionary predatorprey robot system whic...
research
02/01/2018

Virtual-to-Real: Learning to Control in Visual Semantic Segmentation

Collecting training data from the physical world is usually time-consumi...
research
02/18/2019

DIViS: Domain Invariant Visual Servoing for Collision-Free Goal Reaching

Robots should understand both semantics and physics to be functional in ...

Please sign up or login with your details

Forgot password? Click here to reset