Publications• Sorted by Date • Classified by Publication Type • Classified by Research Category • Scalable Bayesian Reinforcement Learning for Multiagent POMDPsChristopher Amato, Frans A. Oliehoek, and Eric Shyu. Scalable Bayesian Reinforcement Learning for Multiagent POMDPs. In Proc. of the First Multidisciplinary Conference on Reinforcement Learning and Decision Making (RLDM2013), 2013. DownloadAbstractBayesian methods for reinforcement learning (RL) allow model uncertainty to be considered explicitly and offer a principled way of dealing with the exploration/exploitation tradeoff. However, for multiagent systems there have been few such approaches, and none of them apply to problems with state uncertainty. In this paper, we fill this gap by proposing a Bayesian RL framework for multiagent partially observable Markov decision processes that is able to take advantage of structure present in many problems. In this framework, a team of agents operates in a centralized fashion, but has uncertainty about the model of the environment. Fitting many real-world situations, we consider the case where agents learn the appropriate models while acting in an online fashion. Because it can quickly become intractable to choose the optimal action in na Ìve versions of this online learning problem, we propose a more scalable approach based on sample-based search and factored value functions for the set of agents. Experimental results show that we are able to provide high quality solutions to large problems even with a large amount of initial model uncertainty. BibTeX Entry@inproceedings{Amato13RLDM,
author = {Christopher Amato and Frans A. Oliehoek and Eric Shyu},
booktitle = {Proc. of the First Multidisciplinary Conference on
Reinforcement Learning and Decision Making (RLDM2013)},
title = {Scalable Bayesian Reinforcement Learning for Multiagent {POMDPs}},
year = 2013,
note = {},
abstract = {
Bayesian methods for reinforcement learning (RL) allow model
uncertainty to be considered explicitly and offer a principled way
of dealing with the exploration/exploitation tradeoff. However, for
multiagent systems there have been few such approaches, and none of
them apply to problems with state uncertainty. In this paper, we fill
this gap by proposing a Bayesian RL framework for multiagent partially
observable Markov decision processes that is able to take advantage of
structure present in many problems. In this framework, a team of
agents operates in a centralized fashion, but has uncertainty about
the model of the environment. Fitting many real-world situations, we
consider the case where agents learn the appropriate models while
acting in an online fashion. Because it can quickly become
intractable to choose the optimal action in na Ìve versions of this
online learning problem, we propose a more scalable approach based on
sample-based search and factored value functions for the set of
agents. Experimental results show that we are able to provide high
quality solutions to large problems even with a large amount of
initial model uncertainty.
}
}
Generated by
bib2html.pl
(written by Patrick Riley) on
Wed Nov 06, 2013 16:37:07 UTC
|