Comparing Model-free and Model-based Algorithms for Offline Reinforcement Learning

01/14/2022
by   Phillip Swazinna, et al.
0

Offline reinforcement learning (RL) Algorithms are often designed with environments such as MuJoCo in mind, in which the planning horizon is extremely long and no noise exists. We compare model-free, model-based, as well as hybrid offline RL approaches on various industrial benchmark (IB) datasets to test the algorithms in settings closer to real world problems, including complex noise and partially observable states. We find that on the IB, hybrid approaches face severe difficulties and that simpler algorithms, such as rollout based algorithms or model-free algorithms with simpler regularizers perform best on the datasets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/12/2020

Overcoming Model Bias for Robust Offline Deep Reinforcement Learning

State-of-the-art reinforcement learning algorithms mostly rely on being ...
research
10/07/2019

Multi-step Greedy Policies in Model-Free Deep Reinforcement Learning

Multi-step greedy policies have been extensively used in model-based Rei...
research
07/03/2019

Benchmarking Model-Based Reinforcement Learning

Model-based reinforcement learning (MBRL) is widely seen as having the p...
research
09/20/2022

Relational Reasoning via Set Transformers: Provable Efficiency and Applications to MARL

The cooperative Multi-A gent R einforcement Learning (MARL) with permuta...
research
08/12/2020

Model-Based Offline Planning

Offline learning is a key part of making reinforcement learning (RL) use...
research
06/09/2021

Planning for Novelty: Width-Based Algorithms for Common Problems in Control, Planning and Reinforcement Learning

Width-based algorithms search for solutions through a general definition...
research
06/14/2023

Off-policy Evaluation in Doubly Inhomogeneous Environments

This work aims to study off-policy evaluation (OPE) under scenarios wher...

Please sign up or login with your details

Forgot password? Click here to reset