Delay-Aware Model-Based Reinforcement Learning for Continuous Control

05/11/2020
by   Baiming Chen, et al.
0

Action delays degrade the performance of reinforcement learning in many real-world systems. This paper proposes a formal definition of delay-aware Markov Decision Process and proves it can be transformed into standard MDP with augmented states using the Markov reward process. We develop a delay-aware model-based reinforcement learning framework that can incorporate the multi-step delay into the learned system models without learning effort. Experiments with the Gym and MuJoCo platforms show that the proposed delay-aware model-based algorithm is more efficient in training and transferable between systems with various durations of delay compared with off-policy model-free reinforcement learning methods. Codes available at: https://github.com/baimingc/dambrl.

READ FULL TEXT
research
05/11/2020

Delay-Aware Multi-Agent Reinforcement Learning for Cooperative and Competitive Environments

Action and observation delays exist prevalently in the real-world cyber-...
research
08/30/2022

An Analysis of Abstracted Model-Based Reinforcement Learning

Many methods for Model-based Reinforcement learning (MBRL) provide guara...
research
01/28/2021

Acting in Delayed Environments with Non-Stationary Markov Policies

The standard Markov Decision Process (MDP) formulation hinges on the ass...
research
01/27/2023

SNeRL: Semantic-aware Neural Radiance Fields for Reinforcement Learning

As previous representations for reinforcement learning cannot effectivel...
research
03/31/2020

Learning to Ask Medical Questions using Reinforcement Learning

We propose a novel reinforcement learning-based approach for adaptive an...
research
05/03/2023

Predictive Wand: a mathematical interface design for operations with delays

Action-feedback delay during operation reduces both task performance and...

Please sign up or login with your details

Forgot password? Click here to reset