Masked Trajectory Models for Prediction, Representation, and Control

05/04/2023
by   Philipp Wu, et al.
0

We introduce Masked Trajectory Models (MTM) as a generic abstraction for sequential decision making. MTM takes a trajectory, such as a state-action sequence, and aims to reconstruct the trajectory conditioned on random subsets of the same trajectory. By training with a highly randomized masking pattern, MTM learns versatile networks that can take on different roles or capabilities, by simply choosing appropriate masks at inference time. For example, the same MTM network can be used as a forward dynamics model, inverse dynamics model, or even an offline RL agent. Through extensive experiments in several continuous control tasks, we show that the same MTM network – i.e. same weights – can match or outperform specialized networks trained for the aforementioned capabilities. Additionally, we find that state representations learned by MTM can significantly accelerate the learning speed of traditional RL algorithms. Finally, in offline RL benchmarks, we find that MTM is competitive with specialized offline RL algorithms, despite MTM being a generic self-supervised learning method without any explicit RL components. Code is available at https://github.com/facebookresearch/mtm

READ FULL TEXT

page 7

page 13

page 16

research
06/17/2022

Bootstrapped Transformer for Offline Reinforcement Learning

Offline reinforcement learning (RL) aims at learning policies from previ...
research
07/07/2023

Goal-Conditioned Predictive Coding as an Implicit Planner for Offline Reinforcement Learning

Recent work has demonstrated the effectiveness of formulating decision m...
research
11/20/2022

UniMASK: Unified Inference in Sequential Decision Problems

Randomly masking and predicting word tokens has been a successful approa...
research
04/01/2023

On Context Distribution Shift in Task Representation Learning for Offline Meta RL

Offline meta reinforcement learning (OMRL) aims to learn transferrable k...
research
12/20/2021

RvS: What is Essential for Offline RL via Supervised Learning?

Recent work has shown that supervised learning alone, without temporal d...
research
07/21/2023

Model-based Offline Reinforcement Learning with Count-based Conservatism

In this paper, we propose a model-based offline reinforcement learning m...
research
06/08/2023

Offline Prioritized Experience Replay

Offline reinforcement learning (RL) is challenged by the distributional ...

Please sign up or login with your details

Forgot password? Click here to reset