SPUDD: Stochastic Planning using Decision Diagrams

01/23/2013
by   Jesse Hoey, et al.
0

Markov decisions processes (MDPs) are becoming increasing popular as models of decision theoretic planning. While traditional dynamic programming methods perform well for problems with small state spaces, structured methods are needed for large problems. We propose and examine a value iteration algorithm for MDPs that uses algebraic decision diagrams(ADDs) to represent value functions and policies. An MDP is represented using Bayesian networks and ADDs and dynamic programming is applied directly to these ADDs. We demonstrate our method on large MDPs (up to 63 million states) and show that significant gains can be had when compared to tree-structured representations (with up to a thirty-fold reduction in the number of nodes required to represent optimal value functions).

READ FULL TEXT

page 1

page 2

page 3

page 5

page 7

page 8

page 9

page 10

research
06/09/2011

Efficient Solution Algorithms for Factored MDPs

This paper addresses the problem of planning under uncertainty in large ...
research
01/16/2014

Probabilistic Relational Planning with First Order Decision Diagrams

Dynamic programming algorithms have been successfully applied to proposi...
research
02/27/2023

Optimistic Planning by Regularized Dynamic Programming

We propose a new method for optimistic planning in infinite-horizon disc...
research
01/03/2023

Faster Approximate Dynamic Programming by Freezing Slow States

We consider infinite horizon Markov decision processes (MDPs) with fast-...
research
02/08/2020

Inferential Induction: Joint Bayesian Estimation of MDPs and Value Functions

Bayesian reinforcement learning (BRL) offers a decision-theoretic soluti...
research
02/14/2012

Symbolic Dynamic Programming for Discrete and Continuous State MDPs

Many real-world decision-theoretic planning problems can be naturally mo...
research
02/06/2013

Correlated Action Effects in Decision Theoretic Regression

Much recent research in decision theoretic planning has adopted Markov d...

Please sign up or login with your details

Forgot password? Click here to reset