Expediting RL by using graphical structures

Peng Dai, Alexander L. Strehi, Judy Goldsmith

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

2 Scopus citations

Abstract

The goal of Reinforcement learning (RL) is to maximize reward (minimize cost) in a Markov decision process (MDP) without knowing the underlying model a priori. RL algorithms tend to be much slower than planning algorithms, which require the model as input. Recent results demonstrate that MDP planning can be expedited, by exploiting the graphical structure of the MDP. We present extensions to two popular RL algorithms, Q-learning and RMax, that learn and exploit the graphical structure of problems to improve overall learning speed. Use of the graphical structure of the under- lying MDP can greatly improve the speed of planning algorithms, if the underlying MDP has a nontrivial topological structure. Our experiments show that use of the apparent topological structure of an MDP speeds up reinforcement learning, even if the MDP is simply connected.

Original languageEnglish
Title of host publication7th International Joint Conference on Autonomous Agents and Multiagent Systems, AAMAS 2008
Pages1297-1300
Number of pages4
StatePublished - 2008
Event7th International Joint Conference on Autonomous Agents and Multiagent Systems, AAMAS 2008 - Estoril, Portugal
Duration: May 12 2008May 16 2008

Publication series

NameProceedings of the International Joint Conference on Autonomous Agents and Multiagent Systems, AAMAS
Volume3
ISSN (Print)1548-8403
ISSN (Electronic)1558-2914

Conference

Conference7th International Joint Conference on Autonomous Agents and Multiagent Systems, AAMAS 2008
Country/TerritoryPortugal
CityEstoril
Period5/12/085/16/08

ASJC Scopus subject areas

  • Artificial Intelligence
  • Software
  • Control and Systems Engineering

Fingerprint

Dive into the research topics of 'Expediting RL by using graphical structures'. Together they form a unique fingerprint.

Cite this