State of the Art—A Survey of Partially Observable Markov Decision Processes: Theory, Models, and Algorithms
Citations Over TimeTop 10% of 1982 papers
Abstract
This paper surveys models and algorithms dealing with partially observable Markov decision processes. A partially observable Markov decision process (POMDP) is a generalization of a Markov decision process which permits uncertainty regarding the state of a Markov process and allows for state information acquisition. A general framework for finite state and action POMDP's is presented. Next, there is a brief discussion of the development of POMDP's and their relationship with other decision processes. A wide range of models in such areas as quality control, machine maintenance, internal auditing, learning, and optimal stopping are discussed within the POMDP-framework. Lastly, algorithms for computing optimal solutions to POMDP's are presented.
Related Papers
- → Partially Observable Markov Decision Processes(2012)293 cited
- → Information state for Markov decision processes with network delays(2008)9 cited
- → A Method for Speeding Up Value Iteration in Partially Observable Markov Decision Processes(2013)11 cited
- → Markov Decision Processes(2014)6 cited
- → Point-Based Methods for Model Checking in Partially Observable Markov Decision Processes(2020)1 cited