Scalable Planning and Learning for Multiagent POMDPs

Similar documents
Best-response play in partially observable card games

The MultiAgent Decision Process toolbox: software for decision-theoretic planning in multiagent systems

An Environment Model for N onstationary Reinforcement Learning

Online Planning Algorithms for POMDPs

Solving Hybrid Markov Decision Processes

Computing Near Optimal Strategies for Stochastic Investment Planning Problems

Measuring the Performance of an Agent

Chi-square Tests Driven Method for Learning the Structure of Factored MDPs

Using Markov Decision Processes to Solve a Portfolio Allocation Problem

An Introduction to Markov Decision Processes. MDP Tutorial - 1

A Sarsa based Autonomous Stock Trading Agent

Lecture 1: Introduction to Reinforcement Learning

Learning the Structure of Factored Markov Decision Processes in Reinforcement Learning Problems

An Empirical Study of Two MIS Algorithms

Feature Selection with Monte-Carlo Tree Search

Agent-based decentralised coordination for sensor networks using the max-sum algorithm

An Application of Inverse Reinforcement Learning to Medical Records of Diabetes Treatment

Learning to Explore and Exploit in POMDPs

Cooperative Multi-Agent Systems from the Reinforcement Learning Perspective Challenges, Algorithms, and an Application

Compact Representations and Approximations for Compuation in Games

Learning is planning: near Bayes-optimal reinforcement learning via Monte-Carlo tree search

Compact Mathematical Programs For DEC-MDPs With Structured Agent Interactions

Goal-Directed Online Learning of Predictive Models

What mathematical optimization can, and cannot, do for biologists. Steven Kelk Department of Knowledge Engineering (DKE) Maastricht University, NL

Preliminaries: Problem Definition Agent model, POMDP, Bayesian RL

Goal-Directed Online Learning of Predictive Models

Markov Decision Processes for Ad Network Optimization

Agent-Based Decentralised Coordination for Sensor Networks using the Max-Sum Algorithm

Motivation. Motivation. Can a software agent learn to play Backgammon by itself? Machine Learning. Reinforcement Learning

Eligibility Traces. Suggested reading: Contents: Chapter 7 in R. S. Sutton, A. G. Barto: Reinforcement Learning: An Introduction MIT Press, 1998.

Predictive Representations of State

A survey of point-based POMDP solvers

POMPDs Make Better Hackers: Accounting for Uncertainty in Penetration Testing. By: Chris Abbott

Learning Tetris Using the Noisy Cross-Entropy Method

Options with exceptions

Rafael Witten Yuze Huang Haithem Turki. Playing Strong Poker. 1. Why Poker?

We employed reinforcement learning, with a goal of maximizing the expected value. Our bot learns to play better by repeated training against itself.

Scheduling Home Health Care with Separating Benders Cuts in Decision Diagrams

Inductive QoS Packet Scheduling for Adaptive Dynamic Networks

A Behavior Based Kernel for Policy Search via Bayesian Optimization

Compression algorithm for Bayesian network modeling of binary systems

The Adaptive k-meteorologists Problem and Its Application to Structure Learning and Feature Selection in Reinforcement Learning

Bayesian Statistics: Indian Buffet Process

Branch-and-Price Approach to the Vehicle Routing Problem with Time Windows

TD(0) Leads to Better Policies than Approximate Value Iteration

Beyond Reward: The Problem of Knowledge and Data

SYSM 6304: Risk and Decision Analysis Lecture 5: Methods of Risk Analysis

Reinforcement Learning of Task Plans for Real Robot Systems

Bidding Dynamics of Rational Advertisers in Sponsored Search Auctions on the Web

A web marketing system with automatic pricing. Introduction. Abstract: Keywords:

Similarity Search and Mining in Uncertain Spatial and Spatio Temporal Databases. Andreas Züfle

Safe robot motion planning in dynamic, uncertain environments

Neuro-Dynamic Programming An Overview

Reinforcement Learning for Vulnerability Assessment in Peer-to-Peer Networks

Multiagent Reputation Management to Achieve Robust Software Using Redundancy

strategic white paper

Structure Learning in Ergodic Factored MDPs without Knowledge of the Transition Function s In-Degree

6.852: Distributed Algorithms Fall, Class 2

MapReduce and Distributed Data Analysis. Sergei Vassilvitskii Google Research

Reinforcement Learning with Factored States and Actions

Moral Hazard. Itay Goldstein. Wharton School, University of Pennsylvania

Deterministic Sampling-based Switching Kalman Filtering for Vehicle Tracking

Course: Model, Learning, and Inference: Lecture 5

Real Time Traffic Monitoring With Bayesian Belief Networks

Local optimization in cooperative agents networks

Monte Carlo Simulations for Patient Recruitment: A Better Way to Forecast Enrollment

8.1 Min Degree Spanning Tree

The Advantages and Disadvantages of Online Linear Optimization

Load Balancing in MapReduce Based on Scalable Cardinality Estimates

Simulation and Risk Analysis

Sampling via Moment Sharing: A New Framework for Distributed Bayesian Inference for Big Data

Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization

Efficient Statistical Methods for Evaluating Trading Agent Performance

A Learning Based Method for Super-Resolution of Low Resolution Images

Decision Theory Rational prospecting

Parallel Computing for Option Pricing Based on the Backward Stochastic Differential Equation

Bargaining Solutions in a Social Network

A Network Flow Approach in Cloud Computing

Urban Traffic Control Based on Learning Agents

Online Learning and Exploiting Relational Models in Reinforcement Learning

Point-Based Value Iteration for Continuous POMDPs

CMSC 858T: Randomized Algorithms Spring 2003 Handout 8: The Local Lemma

CNFSAT: Predictive Models, Dimensional Reduction, and Phase Transition

Intelligent Agents Serving Based On The Society Information

Point-based value iteration: An anytime algorithm for POMDPs

Bayesian networks - Time-series models - Apache Spark & Scala

IMPROVING PERFORMANCE OF RANDOMIZED SIGNATURE SORT USING HASHING AND BITWISE OPERATORS

Master s thesis tutorial: part III

Notes from Week 1: Algorithms for sequential prediction

CHAPTER 1 INTRODUCTION

R-max A General Polynomial Time Algorithm for Near-Optimal Reinforcement Learning

Bootstrapping Big Data

Factoring & Primality

A Game Theoretical Framework for Adversarial Learning

A Spectral Clustering Approach to Validating Sensors via Their Peers in Distributed Sensor Networks

Approximate Solutions to Factored Markov Decision Processes via Greedy Search in the Space of Finite State Controllers

The Optimality of Naive Bayes

THE LOGIC OF ADAPTIVE BEHAVIOR

Using simulation to calculate the NPV of a project

Game theory and AI: a unified approach to poker games

Transcription:

Scalable Planning and Learning for Multiagent POMDPs Christopher Amato CSAIL MIT camato@csail.mit.edu Frans A. Oliehoek Informatics Institute, University of Amsterdam DKE, Maastricht University f.a.oliehoek@uva.nl Abstract Bayesian methods for reinforcement learning (BRL) allow model uncertainty to be considered explicitly and offer a principled way of dealing with the exploration/exploitation tradeoff. However, for multiagent systems there have been few such approaches, and none of them apply to problems with state uncertainty. In this paper, we fill this gap by proposing a BRL framework for multiagent partially observable Markov decision processes. It considers a team of agents that operates in a centralized fashion, but has uncertainty about both the state and the model of the environment, essentially transforming the learning problem to a planning problem. To deal with the complexity of this planning problem as well as other planning problems with a large number of actions and observations, we propose a novel scalable approach based on sample-based planning and factored value functions that exploits structure present in many multiagent settings. Experimental results show that we are able to provide high quality solutions to large problems even with a large amount of initial model uncertainty. We also show that our approach applies in the (traditional) planning setting, demonstrating significantly more efficient planning in factored settings. Introduction One of the main challenges in Reinforcement learning (RL) involves determining the proper balance between exploiting knowledge gained and exploring new possibilities. Bayesian RL (BRL) methods are promising in that, in principle, they provide an optimal exploration/exploitation trade-off with respect to the prior belief. In many real-world situations, the true model may not be known, but a prior can be expressed over a class of possible models. The uncertainty over models can be explicitly considered to choose actions that will maximize expected value over models, reducing uncertainty as needed to improve performance. In multiagent systems (MASs), BRL has been used in stochastic games (Chalkiadakis and Boutilier 2003) and factored Markov decision processes (MDPs) (Teacy et al. 2012). Both of these approaches assume the state of the problem is fully observable (or can be decomposed into fully observable components). While this is a common assumption, many real-world problems are only partially observable due to noisy or inadequate sensors. Unfortunately, no approaches have been proposed that can model and solve problems with partial observability. In fact, while planning in partially observable multiagent domains has had some success (e.g., (Becker et al. 2004; Gmytrasiewicz and Doshi 2005; Nair et al. 2005; Oliehoek 2012; Amato et al. 2013)), very few multiagent RL approaches of any kind consider partially observable domains. 1 A particular difficulty is that under partial observability the usual state is replaced by a state estimate or belief state. In learning settings, however, it is not even possible to compute this belief state. In addition, many advances in exploration strategies, such as R-Max (Brafman and Tennenholtz 2003), are built on the assumption of a discrete state space, making them incompatible with or difficult to use in partially observable settings. This paper presents a novel approach to BRL in MASs with state uncertainty that addresses these difficult problems. BRL methods, by assuming a prior over models, do allow the calculation of a (more complex) belief, and simultaneously address the problem of exploration. Our approach is based on multiagent partially observable Markov decision processes (MPOMDPs), which assume all agents share the same partially observable view of the world and can coordinate on their actions, but have uncertainty about the underlying environment. To deal with uncertainty about the MPOMDP model, we extend the Bayes-Adaptive POMDP (BA-POMDP) model (Ross, Chaib-draa, and Pineau 2007; Ross et al. 2011) to the multiagent setting. The resulting framework is a Bayesian approach for online learning which represents the initial model using priors and updates probability distributions over possible models as the agents act in the real world. Methods for solving BA-POMDPs have been developed which transform the Bayesian learning problem into a POMDP planning problem where states of the system include possible models of the environment. In theory, these approaches could also be used when learning in MPOMDPs, but current solution methods quickly become intractable as the number of joint actions and observations scales exponentially in the number of agents. To combat this intractability, this paper also provides a novel sample-based online planning algorithm that exploits structure present in many MASs. Our method, called 1 Notable exceptions, e.g., (Abdallah and Lesser 2007; Chang, Ho, and Kaelbling 2004; Peshkin et al. 2000; Zhang, Lesser, and Abdallah 2010)).

factored-value partially observable Monte Carlo planning (FV-POMCP), is based on a Monte Carlo tree search (MCTS) variant called POMCP which has shown promise for planning in large POMDPs (Silver and Veness 2010) and BRL in large MDPs (Guez, Silver, and Dayan 2012). FV-POMCP is the first MCTS method that exploits locality of interaction: in many MASs, agents interact directly with a subset of other agents. This structure enables a decomposition of the value function into a set of overlapping factors, which can be used to produce high quality solutions (Guestrin, Koller, and Parr 2001; Nair et al. 2005; Kok and Vlassis 2006). We present two variants of FV- POMCP that use different amounts of factorization of the value function thereby mitigating the additional challenges for scalability imposed by the exponential number of joint actions and observations. Because FV-POMCP exploits locality of interaction in MPOMDPs it is applicable in any large planning problem, whether generated from BRL or not. As such we show experimentally that our approach allows both planning and learning to be significantly more efficient in factored settings. Background We first provide an overview of the multiagent POMDP model and previous work on BRL for (single-agent) POMDPs (Kaelbling, Littman, and Cassandra 1998). Multiagent POMDPs MPOMDPs form a framework for multiagent planning under uncertainty for a team of agents (Messias, Spaan, and Lima 2011). At every stage, agents take individual actions and receive individual observations. However, in an MPOMDP, all individual observations are shared via communication, allowing the team of agents to act in a centralized manner. We will restrict ourselves to the setting where such communication is free of noise, costs and delays. An MPOMDP is a tuple I, S, {A i }, T, R, {Z i }, O, H with: I, a finite set of agents; S, a finite set of states with designated initial state distribution b 0 ; A = i A i, the set of joint actions, using action sets for each agent, i; T, a set of state transition probabilities: T s as = Pr(s s, a), the probability of transitioning from state s to s when actions a are taken by the agents; R, a reward function: R(s, a), the immediate reward for being in state s and taking actions a; Z = i Z i, the set of joint observations, using observation sets for each agent, i; O, a set of observation probabilities: O as z = Pr( z a, s ), the probability of seeing observations o given actions a were taken and resulting state s ; H, horizon. An MPOMDP can be reduced to a POMDP with a single centralized controller that takes joint actions and receives joint observations (Pynadath and Tambe 2002). Therefore, MPOMDPs can be solved with POMDP solution methods. However, such approaches do not exploit the particular structure inherent to many MASs. In Section, we present a first online planning method that overcomes this deficiency. This method also has the potential to overcome the strict requirements on communication that the MPOMDP model poses by limiting communication to a subset of agents. Monte Carlo Tree Search for POMDPs Learning in POMDPs is very difficult due to the lack of strong learning signal (i.e., a Markovian state). While there has been progress (Shani, Brafman, and Shimony 2005; Poupart and Vlassis 2008; Vlassis and Toussaint 2009; Cai, Liao, and Carin 2009; Doshi-Velez et al. 2010), and some amount of success can be achieved with feature-based RL methods (Sutton and Barto 1998) most research has considered the task of planning: given a full specification of the model, determine an optimal policy, π, mapping past observation histories (which can be summarized by distributions b(s) over states called beliefs) to actions. Such an optimal policy can be extracted from an optimal Q-value function, Q(b, a) = s R(s, a) + z P (z b, a) max a Q(b, a ), by acting greedily, in a way similar to regular MDPs. Computing Q(b, a), however, is complicated by the fact that the space of beliefs is continuous. A successful recent online planning method, called POMCP (Silver and Veness 2010), extends Monte Carlo tree search (MCTS) to solve POMDPs. At every stage, the algorithm performs online planning, given the current belief, by incrementally building a lookahead tree that contains (statistics that represent) Q(b, a). The algorithm, however, avoids expensive belief updates by creating nodes not for each belief, but simply for each action-observation history h. In particular, it samples hidden states s only at the root node (called root sampling ) and uses that state to sample a trajectory that first traverses the lookahead tree and then performs a (random) rollout. The return of this trajectory is used to update the statistics for all visited nodes. Even though the size of this search tree can be enormous, not all parts of it are relevant. In order to direct search to the relevant parts, actions are selected to maximize the upper confidence bounds : U(h, a) = Q(h, a) + c log(n + 1)/n. Here, N is the number of times the history has been reached and n is the number of times that action a has been taken in that history. When the exploration constant c is set correctly, POMCP can be shown to converge to an epsilonoptimal value function. Moreover, the method has demonstrated good performance in large domains with a limited numbers of simulations. BRL for POMDPs In contrast to what is assumed in planning, for many realworld applications, the model is not (perfectly) known in advance, which means that the agents have to learn about their environment during execution. This is the task considered in (multiagent) reinforcement learning (RL). BRL methods have become popular because they can provide a principled solution to this exploration/exploitation trade-off (Duff 2002; Engel, Mannor, and Meir 2005; Poupart et al. 2006; Vlassis et al. 2012). POMCP, discussed above, has been extended (Guez, Silver, and Dayan 2012) to solve Bayes-Adaptive MDPs (BA-MDPs) (Duff 2002), a framework that represents the uncertainty about the transi-

tion model using Dirichlet distributions (typically assuming the reward function is known). However, this model does not consider partially observable environments, which is the focus of this paper (nor does it consider multiple agents). Therefore, we build on the framework of Bayes- Adaptive POMDPs (Ross, Chaib-draa, and Pineau 2007; Ross et al. 2011), which extends the BA-MDP to partially observable settings. It utilizes Dirichlet distributions to model uncertainty over transitions and observations. Intuitively, if the agent could observe both states and observations, it could maintain vectors φ and ψ of counts for transitions and observations respectively. That is, φ a ss is the transition count representing the number times state s resulted from taking action a in state s and ψs a z is the observation count representing the number of times observation z was seen after taking action a and transitioning to state s. These counts together then induce a (product of Dirichlets) probability distribution over the possible transition and observation models. Even though the agent cannot observe the states and has uncertainty about the actual count vectors, this uncertainty can be represented using the POMDP formalism. That is, the count vectors are included as part of the hidden state of a special POMDP, called BA-POMDP. In general, the number of count vectors, thus states, is infinite, making belief updating and solving the BA- POMDP impossible without approximation. While it is possible to make a quality-bounded reduction to a finite state space (Ross et al. 2011), this finite space is still very large. This necessitates sample-based planning approaches to provide solutions. For example, (Ross et al. 2011) use an online planning approach that is similar in spirit to POMCP. BA-MPOMDPs We now discuss the extension of the BA-POMDP to the multiagent setting as the Bayes-Adaptive multiagent POMDP (BA-MPOMDP). The BA-MPOMDP model allows a team of agents to learn about its environment while acting in a Bayesian fashion and is applicable in any multiagent RL setting where there is instantaneous communication. Since a BA-MPOMDP can be simply seen as a BA-POMDP where the actions are joint actions and the observations are joint observations, the theoretical results related to BA-POMDPs also apply to the BA-MPOMDP model. This section is necessarily concise, for more details see (Ross et al. 2011). The Model A BA-MPOMDP can be represented as a tuple I, S BM, {A i }, T BM, R BM, {Z i }, O BM, H where I, {A i }, {Z i }, h are as before. As mentioned, the state of the BA-MPOMDP includes the Dirichlet parameters (i.e., count vectors): s BM = s, φ, ψ. As such, the set of states is given by S BM = S T O where T and O are the spaces of possible transition and observation counts respectively. To define T BM, O BM, the transition and observation probabilities for the BA-MPOMDP, we use the expected transition and observation probabilities induced by (the count vectors of) a state: Tφ s as = E[T s as φ] = φ a s a ss /Nφ, O as z ψ = E[O as z ψ] = ψ a s a s z /N as ψ, where Nφ = s φ a ss, and N as ψ = z ψ a s z. Remember that these count vectors are not observed by the agent, since that would require observation of the state. The agent can only maintain beliefs over count vectors. The transition probability is defined as T BM ((s, φ, ψ), a, (s, φ, ψ )) = Tφ s as O as z ψ if φ = φ + δ a ss and ψ = ψ + δ a s z, but 0 otherwise. Here, vector δ a ss is 1 at the index of a, s and s and 0 otherwise and vector δ a s z has value 1 at the index a, s and z and 0 otherwise. Because the observations are included in the transition probabilities, observations in the BA-MPOMDP can be defined to be deterministic: O BM ((s, φ, ψ), a, (s, φ, ψ ), z) = 1 if the count vectors add up, and 0 otherwise. The reward model remains the same (since it is assumed to be known), R BM ((s, φ, ψ), a) = R(s, a). Finally, we assume that the initial state distribution b 0 and initial count vectors φ 0 and ψ 0 are given. Solution Methods Because BA-MPOMDPs can be solved as POMDPs, the formulation of the optimal value function follows trivially. However, the formalism suffers from an infinite state space since there can be infinitely many count vectors. Fortunately, many pairs of φ, ψ (especially ones with large counts) will lead to very similar distributions over T, O, allowing a finite approximate model to be constructed in the multiagent case (Ross et al. 2011). That is, it can be shown that: Given any BA-MPOMDP, ɛ > 0 and horizon H, it is possible to construct a finite POMDP by removing states with count vectors φ, ψ that have N s,a φ > NS ɛ as or Nψ > NZ ɛ for suitable thresholds NS ɛ, N Z ɛ that depend linearly on the number of states and joint observations, respectively. While this result follows from the single-agent setting directly, it is worth noting that while NZ ɛ does depend on the number of joint observations, the thresholds do not depend on the number of joint actions. Therefore, for problems with few observations, constructing this approximation might be feasible, even if there are many actions. In general even a finite approximation of a BA-MPOMDP is intractable to solve optimally. Fortunately, online samplebased planning the approach that allows solutions to be found in a single-agent setting in principle applies to this setting too. That is, the team of agents could perform online planning over a small lookahead horizon, perform the resulting joint action and plan again. Since a BA-MPOMDP is a POMDP, it is possible to apply, e.g., the POMCP algorithm to this setting. Doing so means that during online planning, a lookahead tree will be constructed that has nodes corresponding to joint action observation histories h, which each incrementally construct a particle filter, and where statistics are stored that represent the expected values Q( h, a) and upper confidence bounds U( h, a). Factored-Value POMCP The above description reveals a shortcoming of POMCP and MCTS in general: they are not directly suitable for multia-

(a) Figure 1: (a) Illustration of an fire fighting. (b) The coordination graph with houses h1,...,h4 and agents a1,...,a3. a2 a3 h2 a1 h3 (b) h1 a4 h4 gent problems due to the large number of joint actions and observations, which are exponential in the number of agents. The large number of joint observations is problematic since it will lead to a lookahead tree with very high branching factor. Even though this is theoretically not a problem in MDPs (it can be shown that sample-based planning is independent of the number of states in such settings (Kearns, Mansour, and Ng 2002)), in partially observable settings that use particle filters it does lead to severe practical problems. In particular, in order to have a good particle representation at the next real time step, the actual joint observation received must be sampled often enough during planning for the previous stage. If the actual experienced joint observation had not been sampled frequently enough (or not at all), the particle filter will be a bad approximation (or collapse). This results in a need to sample starting from the initial belief again, or alternatively, to fall back to acting using a separate (history independent) policy such as a random one. The issue of large numbers of joint actions is also problematic: the standard POMCP algorithm will, at each node, maintain separate statistics, and thus separate upper confidence bounds, for each of the exponentially many joint actions. This means that each of the exponentially many joint actions will have to be selected at least a few times to reduce their confidence bounds (i.e., exploration bonus). Clearly this is intractable for all but the smallest problems. The problem of joint actions is a principled one: in cases where it is truly the case that each combination of individual actions may lead to completely different effects, there may be no better thing to do than try all of them at least a few times. In many cases, however, the effect of a joint action is factorizable as the effects of the action of individual agents or small groups of agents. For instance, consider a team of agents that is tasked with fighting fire at a number of burning houses, as illustrated in Figure 1(a). In such a setting, the overall transition probabilities can be factored as a product of transition probabilities for each house (Oliehoek et al. 2008), and the transitions of a particular house may depend only on the amount of water deposited on that house (rather than the exact joint action). While this problem lends itself to a natural factorization, other problems may also be factorized to permit approximation. We propose to exploit this type of structure inside MCTS for MASs by introducing Factored-Value POMCP (FV- POMCP), which uses factored representations of (the statistics representing) the value function inside POMCP. In particular, we introduce variants of FV-POMCP based on two techniques to overcome exponentially large action and observation spaces. The first technique that we call factored statistics, only addresses the complexity introduced by joint actions. The second technique, factored trees, additionally addresses the problem of many joint observations. FV- POMCP is the first MCTS method to exploit structure in MASs, achieving better sample complexity and improving value function generalization by factorization. FV-POMCP applies to BRL for MPOMDPs, but it also directly applies to planning in MPOMDPs and may be beneficial in other multiagent models and other types of factored POMDPs. Coordination Graphs In the multiagent case, we can consider agents interactions in the form of a coordination graph (Guestrin, Koller, and Parr 2001). This coordination graph represents interactions between subsets of agents and permits factored linear value functions as an approximation to the joint value function. We follow the cited literature in assuming that a suitable factorization is easily identifiable by the designer, but it may also be learnable. For the moment assuming a stateless problem, an action-value function can be approximated by Q( a) = e Q e( a e ), where each component e is specified over only a subset of agents (leading to an interaction hypergraph (Nair et al. 2005)). 2 Using a factored representation, the maximization max a e Q e( a e ) can be performed efficiently via variable elimination (VE) (Guestrin, Koller, and Parr 2001), or maxsum (Kok and Vlassis 2006). These algorithms are not exponential in the number of agents, and therefore enable significant speed-ups for larger number of agents. The VE algorithm is still exponential in the induced width of the coordination graph. Max-sum, which is approximate, has been shown to be more effective for densely connected graphs. In order to avoid introducing unnecessary approximations, we focus on using VE in this paper. Factored Statistics The first technique we introduce, called Factored Statistics directly applies the idea of coordination graphs inside MCTS. Rather than maintaining one set of statistics in each node representing the expected value for each joint action Q( h, a), we maintain several sets of statistics, each one expressing the value for a set of agents Q e ( h, a e ). The Q-value function is then approximated by Q( h, a) e Q e( h, a e ). However, joint actions are selected according to the maximum upper confidence bound: max a e U e( h, a e ), which also follows directly from the factored form. Here U e ( h, a e ) = Q e ( h, a e ) + c log(n h + 1)/n ae using the Q- value and the exploration bonus added for that factor. In order to implement this approach, at each node for a joint history h, we store the count for the full history N h as well as the Q-values, Q e, and the counts for actions, n ae, separately for each edge in the coordination graph e. 2 Note, a normalization term of 1/ e can scale the Q-values to the range of the original problem, but the maximizing actions remain the same.

This approach allows the value function to generalize across joint actions using the coordination graph, providing speedups in learning and little or no loss in value if the factorization is correct. This approach can be seen as using linear function approximation for the action component only. Since h is Markov (i.e., contains enough information to predict the value of any future policy) and each Q e includes the full h as an argument, the expected return for a ( h, a e )-pair is only affected by the future policy, but not the policy followed up to this point. As such, an inductive argument easily establishes convergence of this method, even though the solution to which it converges may be arbitrarily far from the optimal solution if the action factorization is poor. Since this method retains fewer statistics and performs joint action selection more efficiently via VE, we expect that it will be more efficient than plain application of POMCP to the BA-MPOMDP. However, the complexity due to joint observations is not directly addressed: because joint histories are used, reuse of nodes and creation of new nodes for all possible histories (including the one that will be realized) may be limited if the number of joint observations is large. Factored Trees The second technique, called Factored Trees, additionally tries to overcome the burden of the large number of joint observations. It further decomposes the local Q e s by splitting joint histories into local histories and distributing them over the factors. In this case, the Q-values are approximated by Q( h, a) e Q e( h e, a e ), and action selection is conducted by maximizing over the sum of the upper confidence bounds: max a e U e( h e, a e ), where U e ( h e, a e ) = Q e ( h e, a e ) + c log(n he + 1)/n ae. We assume that the set of agents with relevant actions and histories for component Q e are the same, but this can be generalized. This approach further reduces the number of statistics maintained and increases the reuse of nodes in MCTS and the chance that nodes in the trees will exist for observations that are seen during execution. As such, it aims to increase performance by utilizing more generalization (also over local histories), as well as producing more robust particle filters. Because histories for other agents outside the factor are not included, the approximation quality may suffer: where h is Markov, this is not the case for the local history he. As such, the expected return for such a local history does not only depend on the future policy, but also on the past one (via the distribution over histories of agents not included in e). This means that convergence is no longer guaranteed, similar to using linear function approximation in, e.g., Q- learning (Sutton and Barto 1998). Still, the latter approach often gives good results in practice, and therefore we expect that if the problem exhibits enough locality, the factored trees approximation may allow good quality policies to be found. This type of factorization has a major effect on the implementation of the approach: rather than constructing a single tree, we now construct a number of trees in parallel, one for each factor e. A node of the tree for component e now stores the required statistics: N he, the count for the local history, n ae, the counts for actions taken in the local tree and Q e for the tree. Finally, we point out that this decentralization of statistics has a major implication on the communication requirements: where regular MPOMDP methods require full synchronization of the observations of all agents, the FT variant of FV-POMCP only requires the agents to know about the local histories. As such, the method only requires local communication between agents connected in the coordination graph. Experimental Results Here we present an empirical evaluation of FV-POMCP. First, we investigate performance on regular MPOMDP online planning tasks. Next, we test on BA-MPOMDPs in order to determine the performance on the considerably harder problem of BRL for MPOMDPs. Experimental Setup. In order to test the empirical performance of the proposed methods, we performed an evaluation on versions of the firefighting problem from Section. 3 Fires are suppressed more quickly if a larger number of agents choose that particular house. Fires also spread to neighbor s houses and can start at any house with a small probability. Each experiment was run for a given number of simulations, the number of samples used at each step to choose an action, and averaged over a number of episodes. We report undiscounted return with error bars corresponding to the standard error. Experiments were run on a single core of a 2.5 GHz machine with 8GB of memory. MPOMDPs. Here, we investigate the performance of the factored statistics (FS) and factored tree (FT) versions of FV-POMCP in multiagent planning problems formalized as MPOMDPs. That is, the agents are given the true MPOMDP model of the environment (in the form of a simulator) and use it to plan. For this setting, we compare to two baseline methods: POMCP: regular POMCP applied to the MPOMDP, random: uniform random action selection. The results for 4-agent and 10-agent instances of the horizon 10 firefighting problem are shown in Figure 2(a). For the 4-agent problem, POMCP performs poorly with a few simulations, but as the number of simulations increases it outperforms the other methods (presumably converging to an optimal solution). FT is able to provide a high-quality solution with a very small number of simulations, but the resulting value plateaus due to approximation error. FS also provides a high-quality solution with a very small number of simulations, but is then able to converge to a solution that is near POMCP. In the 10-agent problem, we see that POMCP is only able to generate a solution that is slightly better than random (-99.5) while the FV-POMCP methods are able to perform much better. In fact, using FS leads to much better policies using 500 simulations, and improves gradually with more simulations. Moreover, using FT a very good solution is generated using only 100 samples (the steep part of the 3 While these problems were introduced as Dec-POMDPs (i.e., without communication), we treat them as MPOMDPs (i.e., with communication), making the rewards not directly comparable.

(a) MPOMDP results (b) BA-MPOMDP results Figure 2: Results for (a) the planning (MPOMDP) case with 4 and 10 agents (log scale x-axis) and (b) the learning (BA- MPOMDP) case with 4 agents for horizons 10 and 50. improvement curves is not visible in 2(a)) and then small improvements are made until the value converges to approximately -66. These results clearly illustrate the benefit of FV-POMCP by exploiting structure for planning in MASs. BA-MPOMDPs. Here we investigate performance in the learning setting, i.e., when the agents are only given the BA- POMDP model. In this setting, at the end of each episode, both the state and count vectors are reset to their initial values. As discussed, learning in partially observable environments is extremely hard, and there may be many equivalence classes of transition and observation models that are indistinguishable when learning. Therefore, it may be necessary to have an estimate of the transition or observation model. In the following we will assume that we have a reasonably good guess of the transitions (e.g., because the designer might have a good idea of the dynamics of the environment), but only a poor estimate of the observation model (because the sensors of an agent may be harder to model). For the BRL setting, we additionally compare to the following baseline methods: POMCP-T: regular POMCP applied to the true model using 100,000 simulations to provide an upper bound, BA-POMCP: regular POMCP applied to the BA-POMDP, No-learning: like BA-POMCP but never updates the count vectors (which causes it to retain the same uncertain distribution over models). While demonstrating the additional benefit of (approximate) Bayesian exploration would be enlightening, the focus of this paper is on improving Bayesian RL for MPOMDPs. Additionally, other common (non-bayesian) baseline exploration methods are not applicable in partially observable settings. 4 Results for a four agent instance of fire fighting are shown in Figure 2(b), for H = 10, 50. The figure illustrates that, in both cases, the FS and FT variants approach the POMCP-T value. For a small number of simulations FT learns very quickly, providing significantly better values than the flat methods and better than the FS method for the increased horizon. Using FS learns more slowly, but the value function is closer to optimal (as seen in the horizon 10 case) as the number of simulations increases due to the use of the full history (rather than the factored history). After more simula- 4 For instance, it is not clear how to apply methods such as R- Max (Brafman and Tennenholtz 2003), or epsilon-greedy exploration (Sutton and Barto 1998) in an efficient way in this setting. tions in the horizon 10 problem, the performance of the flat model (BA-MPOMDP) improves, but the factored methods still outperform it and this increase is less visible for the longer horizon problem. In both cases, the BA-MPOMDP performance is very similar to the case without learning. This surprisingly strong behavior of no-learning can be explained by realizing that even though it does not update the count vectors, in performing the rollouts and constructing the tree, it still implicitly takes into account the effect of learning. These experiments show that even in challenging multiagent settings with state uncertainty, BRL methods have the potential to learn, as long as we effectively exploit structure. Conclusions We presented the first method to utilize multiagent structure to produce a scalable method for Monte Carlo tree search for POMDPs. This approach formalizes a team of agents as a multiagent POMDP, thus allowing planning and BRL techniques from the POMDP literature to be applied. However, since the number of joint actions and observations grows exponentially with the number of agents, naïve extensions of single agent methods will not generally scale well. To combat this problem, we introduced FV-POMCP, which is an online planner based on POMCP (Silver and Veness 2010) that exploits multiagent structure using two novel techniques factored statistics and factored trees to reduce 1) the number of joint actions and 2) the number of joint histories considered. Our empirical results demonstrate that FV-POMCP greatly increases scalability of online planning for MPOMDPs by scaling to problems with 10 agents. Further investigation showed that the increase in performance is large enough to even tackle the more complex learning problem in a four-agent problem. These approaches may serve as the basis for many future directions in multiagent learning. Further improvements could utilize additional domain or policy assumptions to improve scalability. For instance, with transition and observation independence (Becker et al. 2004), and other models of locality of interaction, such as ND-POMDPs (Nair et al. 2005), further factoring of statistics is possible. The methods from this paper could also be used to solve POMDPs and BA-POMDPs with large action and observation spaces

as well the recent Bayes-Adaptive extension (Ng et al. 2012) of the self interested I-POMDP model (Gmytrasiewicz and Doshi 2005). References Abdallah, S., and Lesser, V. 2007. Multiagent reinforcement learning and self-organization in a network of agents. In AAMAS, 172 179. Amato, C.; Chowdhary, G.; Geramifard, A.; Ure, N. K.; and Kochenderfer, M. J. 2013. Decentralized control of partially observable Markov decision processes. In CDC. Becker, R.; Zilberstein, S.; Lesser, V.; and Goldman, C. V. 2004. Solving transition-independent decentralized Markov decision processes. JAIR 22:423 455. Brafman, R. I., and Tennenholtz, M. 2003. R-max-A general polynomial time algorithm for near-optimal reinforcement learning. JMLR 3:213 231. Cai, C.; Liao, X.; and Carin, L. 2009. Learning to explore and exploit in POMDPs. In NIPS 22, 198 206. Chalkiadakis, G., and Boutilier, C. 2003. Coordination in multiagent reinforcement learning: A Bayesian approach. In AAMAS, 709 716. Chang, Y.-H.; Ho, T.; and Kaelbling, L. P. 2004. All learning is local: Multi-agent learning in global reward games. In NIPS 16. Doshi-Velez, F.; Wingate, D.; Roy, N.; and Tenenbaum, J. 2010. Nonparametric Bayesian policy priors for reinforcement learning. In NIPS 23. Duff, M. 2002. Optimal learning: Computational procedures for Bayes-adaptive Markov decision processes. Ph.D. Dissertation, University of Massachusetts Amherst. Engel, Y.; Mannor, S.; and Meir, R. 2005. Reinforcement learning with gaussian processes. In ICML, 201 208. Gmytrasiewicz, P. J., and Doshi, P. 2005. A framework for sequential planning in multi-agent settings. JAIR 24:24 49. Guestrin, C.; Koller, D.; and Parr, R. 2001. Multiagent planning with factored MDPs. In NIPS, 15, 1523 1530. Guez, A.; Silver, D.; and Dayan, P. 2012. Efficient Bayesadaptive reinforcement learning using sample-based search. In NIPS 12. Kaelbling, L. P.; Littman, M. L.; and Cassandra, A. R. 1998. Planning and acting in partially observable stochastic domains. AIJ 101:1 45. Kearns, M.; Mansour, Y.; and Ng, A. Y. 2002. A sparse sampling algorithm for near-optimal planning in large Markov decision processes. MLJ 49(2-3):193 208. Kok, J. R., and Vlassis, N. 2006. Collaborative multiagent reinforcement learning by payoff propagation. JMLR 7:1789 1828. Messias, J. V.; Spaan, M.; and Lima, P. U. 2011. Efficient offline communication policies for factored multiagent POMDPs. In NIPS 24. 1917 1925. Nair, R.; Varakantham, P.; Tambe, M.; and Yokoo, M. 2005. Networked distributed POMDPs: a synthesis of distributed constraint optimization and POMDPs. In AAAI. Ng, B.; Boakye, K.; Meyers, C.; and Wang, A. 2012. Bayesadaptive interactive POMDPs. In AAAI. Oliehoek, F. A.; Spaan, M. T. J.; Whiteson, S.; and Vlassis, N. 2008. Exploiting locality of interaction in factored Dec- POMDPs. In AAMAS, 517 524. Oliehoek, F. A. 2012. Decentralized POMDPs. In Wiering, M., and van Otterlo, M., eds., Reinforcement Learning: State of the Art. Springer. 471 503. Peshkin, L.; Kim, K.-E.; Meuleau, N.; and Kaelbling, L. P. 2000. Learning to cooperate via policy search. In UAI, 489 496. Poupart, P., and Vlassis, N. 2008. Model-based Bayesian reinforcement learning in partially observable domains. In Proceedings of the Eleventh International Symposium on Artificial Intelligence and Mathematics. Poupart, P.; Vlassis, N.; Hoey, J.; and Regan, K. 2006. An analytic solution to discrete Bayesian reinforcement learning. In ICML, 697 704. Pynadath, D. V., and Tambe, M. 2002. The communicative multiagent team decision problem: Analyzing teamwork theories and models. JAIR 16:389 423. Ross, S.; Pineau, J.; Chaib-draa, B.; and Kreitmann, P. 2011. A Bayesian approach for learning and planning in partially observable Markov decision processes. JAIR 12:1729 1770. Ross, S.; Chaib-draa, B.; and Pineau, J. 2007. Bayesadaptive POMDPs. In NIPS 19. Shani, G.; Brafman, R. I.; and Shimony, S. E. 2005. Modelbased online learning of POMDPs. In ECML. Silver, D., and Veness, J. 2010. Monte-carlo planning in large POMDPs. In NIPS 23, 2164 2172. Sutton, R. S., and Barto, A. G. 1998. Reinforcement Learning: An Introduction. MIT Press. Teacy, W. T. L.; Chalkiadakis, G.; Farinelli, A.; Rogers, A.; Jennings, N. R.; McClean, S.; and Parr, G. 2012. Decentralized Bayesian reinforcement learning for online agent collaboration. In AAMAS. Vlassis, N., and Toussaint, M. 2009. Model-free reinforcement learning as mixture learning. In ICML, 1081 1088. Vlassis, N.; Ghavamzadeh, M.; Mannor, S.; and Poupart, P. 2012. Bayesian reinforcement learning. In Wiering, M., and van Otterlo, M., eds., Reinforcement Learning: State of the Art. Springer. Zhang, C.; Lesser, V.; and Abdallah, S. 2010. Selforganization for coordinating decentralized reinforcement learning. In AAMAS, 739 746.