Market-based reinforcement learning in partially observable worlds

Ivo Kwee, Marcus Hutter, Jürgen Schmidhuber

Research output: Chapter in Book/Report/Conference proceedingConference contribution

11 Scopus citations

Abstract

Unlike traditional reinforcement learning (RL), marketbased RL is in principle applicable to worlds described by partially observable Markov Decision Processes (POMDPs), where an agent needs to learn short-term memories of relevant previous events in order to execute optimal actions. Most previous work, however, has focused on reactive settings (MDPs) instead of POMDPs. Here we reimplement a recent approach to market-based RL and for the first time evaluate it in a toy POMDP setting.
Original languageEnglish (US)
Title of host publicationLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
PublisherSpringer [email protected]
Pages865-873
Number of pages9
ISBN (Print)3540424865
DOIs
StatePublished - Jan 1 2001
Externally publishedYes

Fingerprint

Dive into the research topics of 'Market-based reinforcement learning in partially observable worlds'. Together they form a unique fingerprint.

Cite this