Improving Training Result of Partially Observable Markov Decision Process by Filtering Beliefs

01/05/2021
by   Oscar LiJen Hsu, et al.
0

In this study I proposed a filtering beliefs method for improving performance of Partially Observable Markov Decision Processes(POMDPs), which is a method wildly used in autonomous robot and many other domains concerning control policy. My method search and compare every similar belief pair. Because a similar belief have insignificant influence on control policy, the belief is filtered out for reducing training time. The empirical results show that the proposed method outperforms the point-based approximate POMDPs in terms of the quality of training results as well as the efficiency of the method.

READ FULL TEXT
research
01/11/2020

Point-Based Methods for Model Checking in Partially Observable Markov Decision Processes

Autonomous systems are often required to operate in partially observable...
research
01/15/2014

Monte Carlo Sampling Methods for Approximating Interactive POMDPs

Partially observable Markov decision processes (POMDPs) provide a princi...
research
10/19/2020

Belief-Grounded Networks for Accelerated Robot Learning under Partial Observability

Many important robotics problems are partially observable in the sense t...
research
04/02/2022

Hierarchical Reinforcement Learning under Mixed Observability

The framework of mixed observable Markov decision processes (MOMDP) mode...
research
09/30/2011

Anytime Point-Based Approximations for Large POMDPs

The Partially Observable Markov Decision Process has long been recognize...
research
05/06/2020

Multi-Resolution POMDP Planning for Multi-Object Search in 3D

Robots operating in household environments must find objects on shelves,...
research
01/10/2013

Value-Directed Sampling Methods for POMDPs

We consider the problem of approximate belief-state monitoring using par...

Please sign up or login with your details

Forgot password? Click here to reset