Neural Distillation as a State Representation Bottleneck in Reinforcement Learning

10/05/2022
by   Valentin Guillet, et al.
0

Learning a good state representation is a critical skill when dealing with multiple tasks in Reinforcement Learning as it allows for transfer and better generalization between tasks. However, defining what constitute a useful representation is far from simple and there is so far no standard method to find such an encoding. In this paper, we argue that distillation – a process that aims at imitating a set of given policies with a single neural network – can be used to learn a state representation displaying favorable characteristics. In this regard, we define three criteria that measure desirable features of a state encoding: the ability to select important variables in the input space, the ability to efficiently separate states according to their corresponding optimal action, and the robustness of the state encoding on new tasks. We first evaluate these criteria and verify the contribution of distillation on state representation on a toy environment based on the standard inverted pendulum problem, before extending our analysis on more complex visual tasks from the Atari and Procgen benchmarks.

READ FULL TEXT

page 4

page 6

page 9

page 17

page 18

page 19

page 20

page 21

research
10/05/2022

On Neural Consolidation for Transfer in Reinforcement Learning

Although transfer learning is considered to be a milestone in deep reinf...
research
06/11/2019

Continual Reinforcement Learning deployed in Real-life using Policy Distillation and Sim2Real Transfer

We focus on the problem of teaching a robot to solve tasks presented seq...
research
02/13/2018

Progressive Reinforcement Learning with Distillation for Multi-Skilled Motion Control

Deep reinforcement learning has demonstrated increasing capabilities for...
research
07/11/2019

DisCoRL: Continual Reinforcement Learning via Policy Distillation

In multi-task reinforcement learning there are two main challenges: at t...
research
12/31/2021

SimSR: Simple Distance-based State Representation for Deep Reinforcement Learning

This work explores how to learn robust and generalizable state represent...
research
03/23/2021

Drop-Bottleneck: Learning Discrete Compressed Representation for Noise-Robust Exploration

We propose a novel information bottleneck (IB) method named Drop-Bottlen...
research
02/06/2021

Study on the simulation control of neural network algorithm in thermally coupled distillation

Thermally coupled distillation is a new energy-saving method, but the tr...

Please sign up or login with your details

Forgot password? Click here to reset