Unifying Count-Based Exploration and Intrinsic Motivation

06/06/2016
by   Marc G. Bellemare, et al.
0

We consider an agent's uncertainty about its environment and the problem of generalizing this uncertainty across observations. Specifically, we focus on the problem of exploration in non-tabular reinforcement learning. Drawing inspiration from the intrinsic motivation literature, we use density models to measure uncertainty, and propose a novel algorithm for deriving a pseudo-count from an arbitrary density model. This technique enables us to generalize count-based exploration algorithms to the non-tabular case. We apply our ideas to Atari 2600 games, providing sensible pseudo-counts from raw pixels. We transform these pseudo-counts into intrinsic rewards and obtain significantly improved exploration in a number of hard games, including the infamously difficult Montezuma's Revenge.

READ FULL TEXT

page 15

page 17

research
09/14/2021

Focus on Impact: Indoor Exploration with Intrinsic Motivation

Exploration of indoor environments has recently experienced a significan...
research
07/31/2018

Count-Based Exploration with the Successor Representation

The problem of exploration in reinforcement learning is well-understood ...
research
06/05/2023

Flipping Coins to Estimate Pseudocounts for Exploration in Reinforcement Learning

We propose a new method for count-based exploration in high-dimensional ...
research
08/29/2018

Approximate Exploration through State Abstraction

Although exploration in reinforcement learning is well understood from a...
research
05/02/2023

Unlocking the Power of Representations in Long-term Novelty-based Exploration

We introduce Robust Exploration via Clustering-based Online Density Esti...
research
01/06/2019

A copula based approach for electoral quick counts

An electoral quick count is a statistical procedure whose main objective...

Please sign up or login with your details

Forgot password? Click here to reset