Using World Models for Pseudo-Rehearsal in Continual Learning

03/06/2019
by   Nicholas Ketz, et al.
0

The utility of learning a dynamics/world model of the environment in reinforcement learning has been shown in a many ways. When using neural networks, however, these models suffer catastrophic forgetting when learned in a lifelong or continual fashion. Current solutions to the continual learning problem require experience to be segmented and labeled as discrete tasks, however, in continuous experience it is generally unclear what a sufficient segmentation of tasks would be. Here we propose a method to continually learn these internal world models through the interleaving of internally generated rollouts from past experiences (i.e., pseudo-rehearsal). We show this method can sequentially learn unsupervised temporal prediction, without task labels, in a disparate set of Atari games. Empirically, this interleaving of the internally generated rollouts with the external environment's observations leads to an average 4.5x reduction in temporal prediction loss compared to non-interleaved learning. Similarly, we show that the representations of this internal model remain stable across learned environments. Here, an agent trained using an initial version of the internal model can perform equally well when using a subsequent version that has successfully incorporated experience from multiple new environments.

READ FULL TEXT

page 5

page 6

research
11/06/2018

Towards continual learning in medical imaging

This work investigates continual learning of two segmentation tasks in b...
research
10/09/2019

Continual Learning Using Bayesian Neural Networks

Continual learning models allow to learn and adapt to new changes and ta...
research
04/12/2022

Continual Predictive Learning from Videos

Predictive learning ideally builds the world model of physical processes...
research
03/23/2018

Accelerating Learning in Constructive Predictive Frameworks with the Successor Representation

Here we propose using the successor representation (SR) to accelerate le...
research
10/09/2021

Cognitively Inspired Learning of Incremental Drifting Concepts

Humans continually expand their learned knowledge to new domains and lea...
research
09/04/2018

Recurrent World Models Facilitate Policy Evolution

A generative recurrent neural network is quickly trained in an unsupervi...
research
05/24/2022

Thalamus: a brain-inspired algorithm for biologically-plausible continual learning and disentangled representations

Animals thrive in a constantly changing environment and leverage the tem...

Please sign up or login with your details

Forgot password? Click here to reset