Speeding Up the Convergence of Value Iteration in Partially Observable Markov Decision Processes
Journal of Artificial Intelligence Research2001Vol. 14, pp. 29–51
Citations Over TimeTop 10% of 2001 papers
Abstract
Partially observable Markov decision processes (POMDPs) have recently become popular among many AI researchers because they serve as a natural model for planning under uncertainty. Value iteration is a well-known algorithm for finding optimal policies for POMDPs. It typically takes a large number of iterations to converge. This paper proposes a method for accelerating the convergence of value iteration. The method has been evaluated on an array of benchmark problems and was found to be very effective: It enabled value iteration to converge after only a few iterations on all the test problems.
Related Papers
- → Partially Observable Markov Decision Processes(2012)293 cited
- → Information state for Markov decision processes with network delays(2008)9 cited
- → A Method for Speeding Up Value Iteration in Partially Observable Markov Decision Processes(2013)11 cited
- → Markov Decision Processes(2014)6 cited
- → Point-Based Methods for Model Checking in Partially Observable Markov Decision Processes(2020)1 cited