Abstract
A partially-observable Markov decision process (POMDP) is a generalization of a Markov decision process that allows for incomplete information regarding the state of the system. We consider several flavors of finite-horizon POMDPs. Our results concern the complexity of the policy evaluation and policy existence problems, which are characterized in terms of completeness for complexity classes. We prove a new upper bound for the policy evaluation problem for POMDPs, showing it is complete for Probabilistic Logspace. From this, we prove policy existence problems for several variants of unobservable, succinctly represented MDPs to be complete for NPPP , a class for which not many natural problems are known to be complete.
Original language | English |
---|---|
Title of host publication | Mathematical Foundations of Computer Science 1997 - 22nd International Symposium, MFCS 1997, Proceedings |
Editors | Igor Privara, Peter Ruzicka |
Pages | 129-138 |
Number of pages | 10 |
DOIs | |
State | Published - 1997 |
Event | 22nd International Symposium on Mathematical Foundations of Computer Science, MFCS 1997 - Bratislava, Slovakia Duration: Aug 25 1997 → Aug 29 1997 |
Publication series
Name | Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) |
---|---|
Volume | 1295 |
ISSN (Print) | 0302-9743 |
ISSN (Electronic) | 1611-3349 |
Conference
Conference | 22nd International Symposium on Mathematical Foundations of Computer Science, MFCS 1997 |
---|---|
Country/Territory | Slovakia |
City | Bratislava |
Period | 8/25/97 → 8/29/97 |
Bibliographical note
Publisher Copyright:© Springer-Verlag Berlin Heidelberg 1997.
ASJC Scopus subject areas
- Theoretical Computer Science
- General Computer Science