Coarse-to-Fine Q-attention: Efficient Learning for Visual Robotic Manipulation via Discretisation

06/23/2021
by   Stephen James, et al.
0

Reflecting on the last few years, the biggest breakthroughs in deep reinforcement learning (RL) have been in the discrete action domain. Robotic manipulation, however, is inherently a continuous control environment, but these continuous control reinforcement learning algorithms often depend on actor-critic methods that are sample-inefficient and inherently difficult to train, due to the joint optimisation of the actor and critic. To that end, we explore how we can bring the stability of discrete action RL algorithms to the robot manipulation domain. We extend the recently released ARM algorithm, by replacing the continuous next-best pose agent with a discrete next-best pose agent. Discretisation of rotation is trivial given its bounded nature, while translation is inherently unbounded, making discretisation difficult. We formulate the translation prediction as the voxel prediction problem by discretising the 3D space; however, voxelisation of a large workspace is memory intensive and would not work with a high density of voxels, crucial to obtaining the resolution needed for robotic manipulation. We therefore propose to apply this voxel prediction in a coarse-to-fine manner by gradually increasing the resolution. In each step, we extract the highest valued voxel as the predicted location, which is then used as the centre of the higher-resolution voxelisation in the next step. This coarse-to-fine prediction is applied over several steps, giving a near-lossless prediction of the translation. We show that our new coarse-to-fine algorithm is able to accomplish RLBench tasks much more efficiently than the continuous control equivalent, and even train some real-world tasks, tabular rasa, in less than 7 minutes, with only 3 demonstrations. Moreover, we show that by moving to a voxel representation, we are able to easily incorporate observations from multiple cameras.

READ FULL TEXT

page 2

page 6

page 8

research
08/08/2023

Actor-Critic with variable time discretization via sustained actions

Reinforcement learning (RL) methods work in discrete time. In order to a...
research
05/31/2021

Q-attention: Enabling Efficient Learning for Vision-based Robotic Manipulation

Despite the success of reinforcement learning methods, they have yet to ...
research
04/26/2022

Coarse-to-fine Q-attention with Tree Expansion

Coarse-to-fine Q-attention enables sample-efficient robot manipulation b...
research
06/30/2023

Act3D: Infinite Resolution Action Detection Transformer for Robotic Manipulation

3D perceptual representations are well suited for robot manipulation as ...
research
11/27/2020

Adaptable Automation with Modular Deep Reinforcement Learning and Policy Transfer

Recent advances in deep Reinforcement Learning (RL) have created unprece...
research
08/23/2019

A Comparison of Action Spaces for Learning Manipulation Tasks

Designing reinforcement learning (RL) problems that can produce delicate...
research
09/12/2022

Perceiver-Actor: A Multi-Task Transformer for Robotic Manipulation

Transformers have revolutionized vision and natural language processing ...

Please sign up or login with your details

Forgot password? Click here to reset