CoBERL: Contrastive BERT for Reinforcement Learning

07/12/2021
by   Andrea Banino, et al.
10

Many reinforcement learning (RL) agents require a large amount of experience to solve tasks. We propose Contrastive BERT for RL (CoBERL), an agent that combines a new contrastive loss and a hybrid LSTM-transformer architecture to tackle the challenge of improving data efficiency. CoBERL enables efficient, robust learning from pixels across a wide range of domains. We use bidirectional masked prediction in combination with a generalization of recent contrastive methods to learn better representations for transformers in RL, without the need of hand engineered data augmentations. We find that CoBERL consistently improves performance across the full Atari suite, a set of control tasks and a challenging 3D environment.

READ FULL TEXT

page 4

page 6

page 8

page 9

page 10

page 11

page 13

page 17

research
03/15/2021

Sample-efficient Reinforcement Learning Representation Learning with Curiosity Contrastive Forward Dynamics Model

Developing an agent in reinforcement learning (RL) that is capable of pe...
research
06/19/2023

Enhancing Generalization and Plasticity for Sample Efficient Reinforcement Learning

In Reinforcement Learning (RL), enhancing sample efficiency is crucial, ...
research
10/15/2020

Masked Contrastive Representation Learning for Reinforcement Learning

Improving sample efficiency is a key research problem in reinforcement l...
research
05/02/2022

CCLF: A Contrastive-Curiosity-Driven Learning Framework for Sample-Efficient Reinforcement Learning

In reinforcement learning (RL), it is challenging to learn directly from...
research
06/13/2019

Contrastive Bidirectional Transformer for Temporal Representation Learning

This paper aims at learning representations for long sequences of contin...
research
09/15/2022

Human-level Atari 200x faster

The task of building general agents that perform well over a wide range ...
research
04/08/2020

CURL: Contrastive Unsupervised Representations for Reinforcement Learning

We present CURL: Contrastive Unsupervised Representations for Reinforcem...

Please sign up or login with your details

Forgot password? Click here to reset