Will it Blend? Composing Value Functions in Reinforcement Learning

07/12/2018
by   Benjamin van Niekerk, et al.
0

An important property for lifelong-learning agents is the ability to combine existing skills to solve unseen tasks. In general, however, it is unclear how to compose skills in a principled way. We provide a "recipe" for optimal value function composition in entropy-regularised reinforcement learning (RL) and then extend this to the standard RL setting. Composition is demonstrated in a video game environment, where an agent with an existing library of policies is able to solve new tasks without the need for further learning.

READ FULL TEXT

page 7

page 8

research
03/05/2023

Bounding the Optimal Value Function in Compositional Reinforcement Learning

In the field of reinforcement learning (RL), agents are often tasked wit...
research
06/09/2021

Self-Paced Context Evaluation for Contextual Reinforcement Learning

Reinforcement learning (RL) has made a lot of advances for solving a sin...
research
12/18/2018

Universal Successor Features Approximators

The ability of a reinforcement learning (RL) agent to learn about many r...
research
01/06/2020

A Boolean Task Algebra for Reinforcement Learning

We propose a framework for defining a Boolean algebra over the space of ...
research
12/05/2018

Entropic Policy Composition with Generalized Policy Improvement and Divergence Correction

Deep reinforcement learning (RL) algorithms have made great strides in r...
research
05/23/2019

Hierarchical Reinforcement Learning for Concurrent Discovery of Compound and Composable Policies

A common strategy to deal with the expensive reinforcement learning (RL)...
research
06/29/2023

Safety-Aware Task Composition for Discrete and Continuous Reinforcement Learning

Compositionality is a critical aspect of scalable system design. Reinfor...

Please sign up or login with your details

Forgot password? Click here to reset