Leveraging Human Guidance for Deep Reinforcement Learning Tasks

by   Ruohan Zhang, et al.
The University of Texas at Austin

Reinforcement learning agents can learn to solve sequential decision tasks by interacting with the environment. Human knowledge of how to solve these tasks can be incorporated using imitation learning, where the agent learns to imitate human demonstrated decisions. However, human guidance is not limited to the demonstrations. Other types of guidance could be more suitable for certain tasks and require less human effort. This survey provides a high-level overview of five recent learning frameworks that primarily rely on human guidance other than conventional, step-by-step action demonstrations. We review the motivation, assumption, and implementation of each framework. We then discuss possible future research directions.


page 1

page 2

page 3

page 4


Recent Advances in Leveraging Human Guidance for Sequential Decision-Making Tasks

A longstanding goal of artificial intelligence is to create artificial a...

The MineRL Competition on Sample-Efficient Reinforcement Learning Using Human Priors: A Retrospective

To facilitate research in the direction of sample-efficient reinforcemen...

Learning to Request Guidance in Emergent Communication

Previous research into agent communication has shown that a pre-trained ...

MineRL: A Large-Scale Dataset of Minecraft Demonstrations

The sample inefficiency of standard deep reinforcement learning methods ...

Improving Deep Reinforcement Learning in Minecraft with Action Advice

Training deep reinforcement learning agents complex behaviors in 3D virt...

Self-directed Machine Learning

Conventional machine learning (ML) relies heavily on manual design from ...

Interactive Reinforcement Learning with Dynamic Reuse of Prior Knowledge from Human/Agent's Demonstration

Reinforcement learning has enjoyed multiple successes in recent years. H...

Please sign up or login with your details

Forgot password? Click here to reset