Nonapproximability results for partially observable Markov decision processes

Christopher Lusena, Judy Goldsmith, Martin Mundhenk

Research output: Contribution to journalArticlepeer-review

66 Scopus citations

Abstract

We show that for several variations of partially observable Markov decision processes, polynomial-time algorithms for finding control policies are unlikely to or simply don't have guarantees of finding policies within a constant factor or a constant summand of optimal. Here "unlikely" means "unless some complexity classes collapse," where the collapses considered are P = NP, P = PSPACE, or P = EXP. Until or unless these collapses are shown to hold, any control-policy designer must choose between such performance guarantees and efficient computation.

Original languageEnglish
Pages (from-to)83-103
Number of pages21
JournalJournal of Artificial Intelligence Research
Volume14
DOIs
StatePublished - 2001

Funding

FundersFunder number
National Science Foundation Arctic Social Science Program
Directorate for Computer and Information Science and Engineering9315354, 9610348
Directorate for Computer and Information Science and Engineering

    ASJC Scopus subject areas

    • Artificial Intelligence

    Fingerprint

    Dive into the research topics of 'Nonapproximability results for partially observable Markov decision processes'. Together they form a unique fingerprint.

    Cite this