models of coordination in multiagent decision...

104
MODELS OF COORDINATION IN MULTIAGENT DECISION MAKING Chris Amato Many slides from a larger tutorial with Matthijs Spaan and Shlomo Zilberstein Some other slides courtesy of: Dan Bernstein, Frans Oliehoek and Alan Carlin 1 May 21, 2012 CTS Conference 2012

Upload: others

Post on 04-Oct-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

MODELS OF COORDINATION IN MULTIAGENT DECISION MAKING

Chris Amato Many slides from a larger tutorial with Matthijs Spaan and Shlomo Zilberstein Some other slides courtesy of: Dan Bernstein, Frans Oliehoek and Alan Carlin

1 May 21, 2012 CTS Conference 2012

Page 2: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

General overview • Agents situated in a world, receiving information and

choosing actions •  Uncertainty about outcomes and sensors •  Sequential domains •  Cooperative multi-agent •  Decision-theoretic approach

• Developing approaches to scale to real-world domains

May 21, 2012 CTS Conference 2012 2

Page 3: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Outline •  Background on decision-theoretic planning •  Models for cooperative sequential decision-making

•  Dec-POMDPs and MTDPs •  Notable subclasses (Dec-MDPs, ND-POMDPs) •  Related competitive models (POSGs, I-POMDPs)

•  Optimal algorithms •  Finite and infinite horizon •  Top down and bottom up

•  Approximate algorithms •  Finite and infinite horizon

•  Communication •  Applications being targeted •  Conclusion

May 21, 2012 CTS Conference 2012 3

Page 4: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Decision-theoretic planning •  Tackles uncertainty in sensing and acting in a principled

way • Popular for single-agent planning under uncertainty (MDPs,

POMDPs) • We need to model:

•  Each agent's actions •  Their sensors •  Their environment •  Their task

May 21, 2012 CTS Conference 2012 4

????

?

Page 5: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Modeling assumptions • Sequential decisions: problems are formulated as a

sequence of discrete “independent” decisions • Markovian environment: the state at time t depends only

on the events at time t-1 • Stochastic models: the uncertainty about the outcome of

actions and sensing can be accurately captured • Objective encoding: the overall objective can be encoded

using cumulative (discounted) rewards over time steps

May 21, 2012 CTS Conference 2012 5

Page 6: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Problem aspects • On-line vs. off-line • Centralized vs. distributed

•  Planning •  Execution

• Cooperative vs. self-interested • Observability • Communication

May 21, 2012 CTS Conference 2012 6

Page 7: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

A toy example (decentralized tiger) •  2 agents with 2 doors •  A tiger is behind one and a treasure is

behind the other •  Can listen for a noisy signal about the

location of the tiger •  If either agent opens the door with the

tiger, they both get mauled •  If the door with treasure is opened,

they share the reward •  Don’t know each other’s actions and

observations

May 21, 2012 CTS Conference 2012 7

?

Page 8: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Single agent/centralized models • Markov decision processes (MDPs)

•  Stochastic actions, but fully observe state at each step •  P complexity

• Maximize value

• Can be multiagent, but centralized execution (and large size)

May 21, 2012 CTS Conference 2012 8

V (s) = R(s,a)+γmaxa

P(s ' | s,a)V (s ')s '∑

Page 9: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Single agent/centralized models • Partially observable Markov decision processes (POMDPs)

•  Receive an observation rather than true state •  PSPACE complexity (for finite horizons)

• Maximize value in a similar way (but over distributions over states or beliefs)

• Can also be (centralized) multiagent

May 21, 2012 CTS Conference 2012 9

Page 10: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Decentralized domains • Cooperative multiagent problems • Each agent’s choice affects all others, but must be made

using only local information • Properties

•  Often a decentralized solution is required •  Multiple agents making choices independently of the others •  Does not require communication on each step (may be impossible

or too costly) •  But now agents must also reason about the previous and future

choices of the others (more difficult)

May 21, 2012 CTS Conference 2012 10

Page 11: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Example cooperative multiagent problems •  Multi-agent planning (examples below, reconnaissance, etc.) •  Human-robot coordination (combat, industry, etc.) •  Sensor networks (e.g. target tracking from multiple viewpoints) •  E-commerce (e.g. decentralized web agents, stock markets)

May 21, 2012 CTS Conference 2012 11

Page 12: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Sensor network problems • Sensor networks for

•  Target tracking (Nair et al., 05, Kumar and Zilberstein 09-AAMAS)

•  Weather phenomena (Kumar and Zilberstein 09-IJCAI)

•  Two or more cooperating sensors

May 21, 2012 CTS Conference 2012 12

Page 13: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Other possible application domains • Multi-robot coordination

•  Space exploration rovers (Zilberstein et al., 02)

•  Helicopter flights (Pynadath and Tambe, 02)

•  Navigation (Emery-Montemerlo et al., 05; Spaan and Melo 08)

•  Load balancing for decentralized queues (Cogill et al., 04)

• Multi-access broadcast channels (Ooi and Wornell, 96)

• Network routing (Peshkin and Savova, 02)

• Sensor network management (Nair et al., 05)

May 21, 2012 CTS Conference 2012 13

Page 14: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Multiple cooperating agents •  Decentralized partially observable Markov decision process (Dec-

POMDP) also called multiagent team decision problem (MTDP) •  Extension of the single agent POMDP •  Multiagent sequential decision-making under uncertainty

•  At each stage, each agent takes an action and receives: •  A local observation •  A joint immediate reward

Environment

a1

o1 an

on

r

May 21, 2012 CTS Conference 2012 14

Page 15: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Dec-POMDP definition • A Dec-POMDP can be defined with the tuple: M = <I, S, {Ai}, P, R, {Ωi}, O>

•  I, a finite set of agents •  S, a finite set of states with designated initial state distribution b0 •  Ai, each agent’s finite set of actions •  P, the state transition model: •  R, the reward model: •  Ωi, each agent’s finite set of observations •  O, the observation model:

Note: Functions depend on all agents

May 21, 2012 CTS Conference 2012 15

P(s ' | s, a)R(s, a)

P(o | s, a)

Page 16: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Dec-POMDP solutions • A local policy for each agent is a mapping from its

observation sequences to actions, Ω* A •  State is unknown, so beneficial to remember history

• A joint policy is a local policy for each agent • Goal is to maximize expected cumulative reward over a

finite or infinite horizon •  For infinite-horizon cannot remember the full observation history •  In infinite case, a discount factor, γ, is used

May 21, 2012 CTS Conference 2012 16

Page 17: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Example: 2-Agent Navigation

States: grid cell pairs Actions: move , , , , stay Transitions: noisy Observations: red lines Rewards: negative unless sharing the same square

May 21, 2012 CTS Conference 2012 17

Page 18: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Challenges in solving Dec-POMDPs • Partial observability makes the problem difficult to solve • No common state estimate (centralized belief state)

•  Each agent depends on the others •  This requires a belief over the possible policies of the other agents •  Can’t transform Dec-POMDPs into a continuous state MDP (how

POMDPs are typically solved)

•  Therefore, Dec-POMDPs cannot be solved by centralized (POMDP) algorithms

May 21, 2012 CTS Conference 2012 18

Page 19: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

General complexity results

subclasses and finite horizon complexity results

P PSPACE NEXP

NEXP

May 21, 2012 CTS Conference 2012 19

Page 20: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Relationship with other models

Ovals represent complexity, while colors represent number of agents and cooperative or competitive models

Relationships among the models

MMDP

DEC−MDP

POSG

MDPI−POMDP(finitely nested) POMDP

MTDPDEC−POMDPDEC−POMDP−COM

38/142

May 21, 2012 CTS Conference 2012 20

Page 21: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

POSGs • A partially observable stochastic game (POSG) is a tuple

M = <I, S, {Ai}, P, {Ri}, {Ωi}, O> where •  All the components except the reward function are the same as in a

DEC-POMDP •  Each agent has an individual reward function:

•  Ri(s,ai) denotes the reward obtained after action ai was taken by agent i and a state transition to s’ occurred

•  This is the self-interested version of the DEC-POMDP model

May 21, 2012 CTS Conference 2012 21

Ri : Ai × S→ℜ

Page 22: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

I-POMDPs •  Interactive POMDPs (I-POMDPs) extend state space with

behavioral models of other agents (Gmytrasiewicz and Doshi 05)

• Agents maintain beliefs over physical and models of others

• Recursive modeling • When assuming a finite nesting, beliefs and value

functions can be computed (approximately). •  Finitely nested I-POMDPs can be solved as a set of

POMDPs

May 21, 2012 CTS Conference 2012 22

Page 23: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Dec-MDPs •  Joint full observability = collective observability • A Dec-POMDP is jointly fully observable if the n-tuple of

observations made by all the agents uniquely determine the current global state. •  That is, if then

• A decentralized Markov decision process (Dec-MDP) is a Dec-POMDP with joint full observability.

• A stronger result: The problem is NEXP-hard even when the state is jointly observed! •  That is, two-agent finite-horizon DEC-MDPs are NEXP-hard.

May 21, 2012 CTS Conference 2012 23

O(o | s ', a)> 0 P(s ' | o) =1

Page 24: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Classes of Dec-POMDPs •  A factored n-agent Dec-MDP is a Dec-MDP for which the world

state can be factored into n+1 components, S= S0 ×S1 ×…× Sn •  A factored, n-agent Dec-MDP is said to be locally fully

observable if each agent observes its own state component

•  Local state/observation/action is referred to as the local state, as the local action, and as the local observation for agent I

•  Full observability = individual observability •  A Dec-POMDP is fully observable if there exists a mapping for

each agent i, such that whenever is non-zero then

May 21, 2012 CTS Conference 2012 24

∀oi∃si :P(si | oi ) =1si ∈ Si × S0

ai ∈ Ai oi ∈Oi

fi :Ωi → S O(o | s ', a)fi (oi ) = s '

Page 25: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Classes of Dec-POMDPs • A multi-agent Markov decision process (MMDP) is a Dec-

POMDP with full observability (Boutilier, 96)

• A factored, n-agent Dec-MDP is said to be transition independent if there exists P0 through Pn such that

• A factored, n-agent Dec-MDP is said to be observation independent if there exists O1 through On such that:

May 21, 2012 CTS Conference 2012 25

P(s 'i | (so,…, sn ),a, (s 'o,…, s 'i−1, s 'i+1,…, s 'n )) =

P(oi |,a, (s 'o,…, s 'n ), (oo,…,oi−1,oi+1,…,on )) = P(oi | ai, s 'i )

P0 (s '0 | s0 ) i = 0

Pi (s 'i | si,a, s '0 ) 1≤ i ≤ n

Page 26: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Classes of Dec-POMDPs • A factored, n-agent Dec-MDP is said to be reward

independent if there exist ƒ and R1 through Rn such that

and •  For example, additive rewards •  If a Dec-MDP has independent observations and

transitions, then the Dec-MDP is locally fully observable.

May 21, 2012 CTS Conference 2012 26

R((s0,…, sn ),a) = f (R1(s1,a1),...,Rn (sn,an ))

Ri (si,ai ) ≤ Ri (s 'i,a 'i )⇔f (R1...Ri (si,ai )...Rn ) ≤ f (R1...Ri (s 'i,a 'i )...Rn )

Page 27: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Some complexity results

May 21, 2012 CTS Conference 2012 27

Observability General communication Free communication

Full MMDP (P-complete) MMDP (P-complete)

Joint full Dec-MDP (NEXP) MMDP (P-complete)

Partial Dec-POMDP (NEXP) MPOMDP (PSPACE)

Page 28: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

ND-POMDPs (Nair et al., 05)

• A network distributed POMDP (ND-POMDP) is a factored n-agent Dec-POMDP with independent transitions and observations as well as a decomposable reward function

•  That is,

Where l is the subgroup of agents of size k •  Model locality of interaction •  Doesn’t make joint full observability assumption of Dec-MDPs •  Complexity depends on k (but still NEXP in worst case)

May 21, 2012 CTS Conference 2012 28

Ag1 Ag2 Ag3

Ag5 Ag4

R((s0, s1,…, sn ),a) = Rl (s0, sl1,.., slk,a1k,...,alk )

l∑

Page 29: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Factored Dec-POMDPs (Oliehoek et al., 08)

• Motivation: exploit locality of interaction, but no strict independence assumptions

• More general and powerful than ND-POMDPs

•  Less scalable (non-stationary interaction graph)

May 21, 2012 CTS Conference 2012 29

t

FL1’

FL2’

FL3’

FL4’

a1

a2

a3

FL1

FL2

FL3

FL4

t ’

o1

o2

o3

R1

R2

R3

R4

Page 30: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Algorithms • How do we produce a solution for these models? • Will discuss

•  Solution representations •  Optimal methods •  Approximate methods •  Subclasses •  Communication

May 21, 2012 CTS Conference 2012 30

Page 31: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Policy representations – finite horizon •  Policy trees, one for each agent •  Nodes define actions, links define observations •  One leaf for each history for the given horizon

•  Evaluation: •  Starting from a node q, taking the associated action and transition

May 21, 2012 CTS Conference 2012 31

V (q, s) = R(s, aq )+γ P(s ' | s, aq ) O(o | s ', a)V (qo, s ')o∑

s '∑

li

o 1 o 2

o1 o2

o r li

o 1 o2

li

li

li

o 1 o 2

o1 o2

li

li li

o 1 o 2

li

li

li ol

li

o 1 o 2

li

li

o 1 o 2

o1 o2

o r li

o 1 o2

li

li

li

o 1 o 2

o1 o2

li

li li

o 1 o 2

li

li

li ol

li

o 1 o 2

li

Page 32: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Policy representations – infinite horizon

•  Designated initial node •  Nodes define actions •  Transitions based on

observations seen •  Inherently infinite-horizon •  With fixed memory,

randomness can help •  One controller for each agent

•  Value for node and state s:

Actions: move in direction or stop Observations: wall left, wall right

Action selection, P(a|q): Q ΔA Transitions, P(q’|q,o): Q × O ΔQ!

V (q, s) = P(a | q) R(s, a)+ γ P(s ' | s, a) O(o | s ', a) P(q ' | q,o)V (q ', s ')q '∑

o∑

s '∑

⎣⎢

⎦⎥

a∑

May 21, 2012 CTS Conference 2012 32

q

Page 33: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Optimal methods • Want to produce the optimal solution for a problem •  That is, optimal horizon h tree or ε-optimal controller • Do this in a bottom up or a top down manner

May 21, 2012 CTS Conference 2012 33

Page 34: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Bottom up methods • Build the policies up for each agent simultaneously • Begin on the last step (single action) and continue until

the first • When done, choose highest value set of trees for any

initial state

May 21, 2012 CTS Conference 2012 34

Page 35: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Exhaustive search • Construct all possible policies for the set of agents • Do this in a bottom up fashion • Stop when the desired horizon is reached •  Trivially includes an optimal set of trees when finished • Choose the set of policies with the highest value at the

initial belief

May 21, 2012 CTS Conference 2012 35

Page 36: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Exhaustive search example (2 agents)

a1 a2 a1 a2

May 21, 2012 CTS Conference 2012 36

Page 37: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Exhaustive search continued

a1

a1 a2

o1 o2

a2

a1 a2

o1 o2

a1

a2 a1

o1 o2

a2

a1 a1

o1 o2

a2

a2 a1

o1 o2

a1

a2 a2

o1 o2

a2

a2 a2

o1 o2

a1

a1 a1

o1 o2

a1

a1 a2

o1 o2

a2

a1 a2

o1 o2

a1

a2 a1

o1 o2

a2

a1 a1

o1 o2

a2

a2 a1

o1 o2

a1

a2 a2

o1 o2

a2

a2 a2

o1 o2

a1

a1 a1

o1 o2

May 21, 2012 CTS Conference 2012 37

Page 38: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Exhaustive search continued

May 21, 2012 CTS Conference 2012 38

Page 39: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Exhaustive search summary • Can find an optimal set of trees • Number of each agent's trees grows exponentially at each

step • Many trees will not contribute to an optimal solution • Can we reduce the number of trees we consider?

May 21, 2012 CTS Conference 2012 39

Page 40: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Finite horizon dynamic programming (DP) •  Build policy tree sets simultaneously •  Returns optimal policy for any initial state •  Prune using a generalized belief space (with linear program) •  This becomes iterative elimination of dominated strategies in POSGs •  For two agents:

s1

s2

×

agent 2 state space

×

agent 1 state space

a3

a1 a1

o1 o2

a2

a3 a1

o1 o2 p1 p2

a2

a2 a2

o1 o2

a3

a1 a2

o1 o2 q1 q2

q1

q2

Q× Sagent 1 value (4D)

(Hansen et al. 2004)

May 21, 2012 CTS Conference 2012 40

Page 41: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

DP for DEC-POMDPs example (2 agents)

a1 a2 a1 a2

May 21, 2012 CTS Conference 2012 41

Page 42: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

DP for DEC-POMDPs continued

a1

a1 a2

o1 o2

a2

a1 a2

o1 o2

a1

a2 a1

o1 o2

a2

a1 a1

o1 o2

a2

a2 a1

o1 o2

a1

a2 a2

o1 o2

a2

a2 a2

o1 o2

a1

a1 a1

o1 o2

a1

a1 a2

o1 o2

a2

a1 a2

o1 o2

a1

a2 a1

o1 o2

a2

a1 a1

o1 o2

a2

a2 a1

o1 o2

a1

a2 a2

o1 o2

a2

a2 a2

o1 o2

a1

a1 a1

o1 o2

May 21, 2012 CTS Conference 2012 42

Page 43: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

DP for DEC-POMDPs continued

a1

a1 a2

o1 o2

a1

a2 a1

o1 o2

a2

a2 a1

o1 o2

a1

a2 a2

o1 o2

a2

a2 a2

o1 o2

a1

a1 a1

o1 o2

a1

a1 a2

o1 o2

a2

a1 a2

o1 o2

a1

a2 a1

o1 o2

a2

a1 a1

o1 o2

a2

a2 a1

o1 o2

a1

a2 a2

o1 o2

a2

a2 a2

o1 o2

a1

a1 a1

o1 o2

May 21, 2012 CTS Conference 2012 43

Page 44: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

DP for DEC-POMDPs continued

a1

a1 a2

o1 o2

a1

a2 a1

o1 o2

a2

a2 a1

o1 o2

a1

a2 a2

o1 o2

a2

a2 a2

o1 o2

a1

a1 a1

o1 o2

a1

a1 a2

o1 o2

a2

a1 a2

o1 o2

a1

a2 a1

o1 o2

a2

a2 a1

o1 o2

a2

a2 a2

o1 o2

May 21, 2012 CTS Conference 2012 44

Page 45: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

DP for DEC-POMDPs continued

a1

a2 a1

o1 o2

a2

a2 a1

o1 o2

a2

a2 a2

o1 o2

a1

a1 a1

o1 o2

a1

a1 a2

o1 o2

a2

a1 a2

o1 o2

a1

a2 a1

o1 o2

a2

a2 a1

o1 o2

a2

a2 a2

o1 o2

May 21, 2012 CTS Conference 2012 45

Page 46: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

DP for DEC-POMDPs continued

a1

a2 a1

o1 o2

a2

a2 a1

o1 o2

a2

a2 a2

o1 o2

a1

a1 a1

o1 o2

a1

a1 a2

o1 o2

a1

a2 a1

o1 o2

a2

a2 a1

o1 o2

a2

a2 a2

o1 o2

May 21, 2012 CTS Conference 2012 46

Page 47: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

DP for DEC-POMDPs continued

May 21, 2012 CTS Conference 2012 47

Page 48: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Infinite horizon version (Bernstein et al., 09)

• Remember we need to define a controller for each agent • How many nodes do you need and what should the

parameter values be for an optimal infinite-horizon policy? •  This may be infinite! •  First ε-optimal algorithm for infinite-horizon Dec-POMDPs:

Policy Iteration

May 21, 2012 CTS Conference 2012 48

Page 49: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Optimal DP: Policy Iteration •  Start with a given controller •  Exhaustive backup (for all agents):

generate all next step policies •  Evaluate: determine value of starting

at each node at each state and for each policy for the other agents

•  Prune: remove those that always have lower value (merge as needed)

•  Continue with backups and pruning until error is below ε

s1 s2

(backup for action 1) Q × S

o2

a1 o1

a1 a2

o1

o1

o2

o2

a1 a2

o1

o1

o2

o2

o1 a1

o2

a1 o1,o2

a1 o1,o2

= Initial controller for agent 1

May 21, 2012 CTS Conference 2012 49

Page 50: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Optimal DP: Policy Iteration •  Start with a given controller •  Exhaustive backup (for all agents):

generate all next step policies •  Evaluate: determine value of starting

at each node at each state and for each policy for the other agents

•  Prune: remove those that always have lower value (merge as needed)

•  Continue with backups and pruning until error is below ε

Key: Prune over not just states, but possible policies of the other agents! s1 s2

(backup for action 1) Q × S

o2

a1 o1

a1 a2

o1

o1

o2

o2

a1 a2

o1

o1

o2

o2

a1 o1,o2

= Initial controller for agent 1

May 21, 2012 CTS Conference 2012 50

Page 51: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

DP for for Dec-POMDPs • What happens in infinite or large horizons?

•  Number of trees is doubly exponential in the horizon •  Doesn’t consider start state •  Full backup wasteful (many trees pruned)

• Need more efficient backups • Or approximate approaches to increase scalability

May 21, 2012 CTS Conference 2012 51

Page 52: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Incremental policy generation (Amato et al., 09)

• Optimal dynamic programming for Dec-POMDPs requires a large amount of time and space

•  In POMDPs, methods have been developed to make optimal DP more efficient (e.g., incremental pruning)

•  These cannot be extended to Dec-POMDPs (due to the lack of a shared viewpoint by the agents)

• Makes the optimal approaches for both finite and infinite-horizon more efficient

May 21, 2012 CTS Conference 2012 52

Page 53: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Incremental policy generation (cont.) •  Can avoid exhaustively generating policies (backups) •  Can limit the number of states considered based on action and

observation (see a wall, other agent, etc.) •  This allows policies for an agent to be built up incrementally •  Add only subtrees (or subcontrollers) that are not dominated

Key: rune only over reachable subspaces

May 21, 2012 CTS Conference 2012 53

Page 54: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Benefits of IPG and results

•  Solve larger problems optimally •  Can make use of start state information as well •  Can be used in other dynamic programming algorithms

•  Optimal: Finite-, infinite- and indefinite horizon •  Approximate: PBDP, MBDP, IMBDP, MBDP-OC and PBIP

Increases horizon in optimal DP (finite or infinite-hor)

0 2 4 6 8

10 12 14 16 18 20

Grid 3x3 Box Pushing Mars Hotel

Hor

izon

Domain

DP

IPG

IPG with start

May 21, 2012 CTS Conference 2012 54

State-of-the-art in bottom up and infinite-horizon

Page 55: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Top down approaches • Perform a search starting from a known initial (belief)

state • Continue until the desired horizon is reached

May 21, 2012 CTS Conference 2012 55

a 1

o1 o2

o1 o2

a 1

a 1 a 2

o1 o2

a 2

a 2

a 1

o1 o2

o1 o2

a 1

a 1 a 2

o1 o2

a 2

a 2

Bottom upTop down

newold

a 2 a 2t = 2

t = 1

t = 0

Page 56: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Multiagent A* (Szer et al., 05)

•  Can build up policies for agents from the first step •  Use heuristic search over joint policies

•  Actions: and , observations: and •  History: and joint policies for a single step: •  Heuristic value for remaining steps (e.g., MDP or POMDP)

May 21, 2012 CTS Conference 2012 56

a a o o

a

a

a

aaa

a

oo

o

oo

o

t = 0

t = 1

t = 2

!2i

"0i

"1i

"2i

a

a

a

aaa

a

oo

o

oo

o

t = 0

t = 1

t = 2

!2i

"0i

"1i

"2i

... ... ...a1

a1

a1

a1a1a1

a1

o1o1

o1

o1o1

o1

a2

a2

a2

a2a2a2

a2

o2o2

o2

o2o2

o2t = 0

t = 1

t = 2

!21 !2

2

!2

"2

Page 57: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Multiagent A* continued

• Requires an admissible heuristic function • A*-like search over partially specified joint policies:

• Heuristic value:

•  If is admissible (overestimation), so is

May 21, 2012 CTS Conference 2012 57 Multiagent A*

• MAA!: top-down heuristic policy search (Szer et al., 2005).! Requires an admissable heuristic function.! A*-like search over partially specified joint policies:

!t = ("0, "1, . . . , "t"1),

"t = ("t0, . . . , "tn) "ti : #Ot

i ! Ai

• Heuristic value for !t:

!V (!t)" #$ %F

= V 0...t"1(!t)" #$ %G

+ !V t...h"1" #$ %

H

! If !V t...h"1 is admissible (overestimation), so is !V (!t).

Optimal, General, Finite-horizon– p. 79/147

Multiagent A*

• MAA!: top-down heuristic policy search (Szer et al., 2005).! Requires an admissable heuristic function.! A*-like search over partially specified joint policies:

!t = ("0, "1, . . . , "t"1),

"t = ("t0, . . . , "tn) "ti : #Ot

i ! Ai

• Heuristic value for !t:

!V (!t)" #$ %F

= V 0...t"1(!t)" #$ %G

+ !V t...h"1" #$ %

H

! If !V t...h"1 is admissible (overestimation), so is !V (!t).

Optimal, General, Finite-horizon– p. 79/147

Multiagent A*

• MAA!: top-down heuristic policy search (Szer et al., 2005).! Requires an admissable heuristic function.! A*-like search over partially specified joint policies:

!t = ("0, "1, . . . , "t"1),

"t = ("t0, . . . , "tn) "ti : #Ot

i ! Ai

• Heuristic value for !t:

!V (!t)" #$ %F

= V 0...t"1(!t)" #$ %G

+ !V t...h"1" #$ %

H

! If !V t...h"1 is admissible (overestimation), so is !V (!t).

Optimal, General, Finite-horizon– p. 79/147

Multiagent A*

• MAA!: top-down heuristic policy search (Szer et al., 2005).! Requires an admissable heuristic function.! A*-like search over partially specified joint policies:

!t = ("0, "1, . . . , "t"1),

"t = ("t0, . . . , "tn) "ti : #Ot

i ! Ai

• Heuristic value for !t:

!V (!t)" #$ %F

= V 0...t"1(!t)" #$ %G

+ !V t...h"1" #$ %

H

! If !V t...h"1 is admissible (overestimation), so is !V (!t).

Optimal, General, Finite-horizon– p. 79/147

Page 58: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Multiagent A* continued • Expand children by growing policy

to the next horizon • Choose nodes with the highest F

value • Continue until node with highest F

is a full horizon policy

May 21, 2012 CTS Conference 2012 58

Multiagent A*

• MAA!: top-down heuristic policy search (Szer et al., 2005).! Requires an admissable heuristic function.! A*-like search over partially specified joint policies:

!t = ("0, "1, . . . , "t"1),

"t = ("t0, . . . , "tn) "ti : #Ot

i ! Ai

• Heuristic value for !t:

!V (!t)" #$ %F

= V 0...t"1(!t)" #$ %G

+ !V t...h"1" #$ %

H

! If !V t...h"1 is admissible (overestimation), so is !V (!t).

Optimal, General, Finite-horizon– p. 79/147

Page 59: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Heuristic functions •  Defined by solving simplified problem settings •  QMDP: assume the underlying state is fully observable by the

agents from that step on •  Cheap to compute (solve an MDP) •  Often loose (strong assumptions: centralized and fully observable)

•  QPOMDP: assume the observations of all agents are shared from that step on •  More difficult to solve (exponential in h to solve POMDP)

•  QBG: assume the observations of all agents are shared on the next step on •  Still harder to solve

•  Hierarchy of upper bounds

May 21, 2012 CTS Conference 2012 59

Q*≤QBG≤QPODP≤QMDP

Page 60: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Heuristic functions: example tiger problem

May 21, 2012 CTS Conference 2012 60

0 10

10

20

30

40

50

60Q−heuristics for horizon=4 Dec−Tiger at t=0

P(sl | et )

Qm

ax =

max

a Q(e

t ,a)

QBGQPOMDPQMDP

Q*

0 1−10

0

10

20

30

40

50

60Q−heuristics for horizon=4 Dec−Tiger at t=1

P(sl | et )

Qm

ax =

max

a Q(e

t ,a)

QBGQPOMDPQMDP

Q*

0 1−20

−10

0

10

20

30

40Q−heuristics for horizon=4 Dec−Tiger at t=2

P(sl | et )

Qm

ax =

max

a Q(e

t ,a)

QBGQPOMDPQMDP

Q*

0 1−100

−80

−60

−40

−20

0

20Q−heuristics for horizon=4 Dec−Tiger at t=3

P(sl | et )

Qm

ax =

max

a Q(e

t ,a)

QBGQPOMDPQMDP

Q*

Page 61: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Generalized MAA* (Oliehoek et al., 08)

• Represent Dec-POMDPs as a series of Bayesian games • Also allowed different heuristic functions and solvers to be

used

May 21, 2012 CTS Conference 2012 61

DEC-POMDPs as series of Bayesian Games

t = 0

t = 1

joint actionsjoint observationsjoint act.-obs. history!a1, a2"

!a1, a2"

!a1, a2"

!a1, a2"

!o1, o2"

!o1, o2"!o1, o2"

!o1, o2"

– p. 87/147

Page 62: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Dec-POMDPs as a series of Bayesian games

May 21, 2012 CTS Conference 2012 62

DEC-POMDPs as series of Bayesian Games

!" t=02

()

!" t=01

a2 a2

() a1 +2.75 !4.1

a1 !0.9 +0.3

!" t=12

(a2, o2) (a2, o2) ...!" t=11

a2 a2 a2 a2

(a1, o1)a1 !0.3 +0.6 !0.6 +4.0 ...

a1 !0.6 +2.0 !1.3 +3.6 ...

(a1, o1)a1 +3.1 +4.4 !1.9 +1.0 ...

a1 +1.1 !2.9 +2.0 !0.4

(a1, o1)a1 !0.4 !0.9 !0.5 !1.0 ...

a1 !0.9 !4.5 !1.0 +3.5 ...

(a1, o1) ... ... ... ... ... ...

– p. 86/147

Page 63: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Lossless clustering of histories (Oliehoek et al, 09)

•  Idea: if two individual histories induce the same distribution over states and over other agents' histories, they are equivalent and can be clustered

•  Lossless clustering, independent of heuristic, but problem dependent

• Clustering is bootstrapped: algorithms only deal with clustered Bayesian games

•  Large increase in scalability of optimal solvers

May 21, 2012 CTS Conference 2012 63

Page 64: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Incremental expansion (Spaan et al., 11)

• Number of children expanded is doubly exponential in t • Many of these not useful for optimal policy • Use incremental expansion

•  Find next step policies as a cooperative Bayesian game (CBG) •  Keep pointer to unexpanded nodes in open list

May 21, 2012 CTS Conference 2012 64

-Create new CBG-Solution of CBG:

-Create new CBG-Solution of CBG:

...

-Next solution of CBG:

Values in open list

... ... ...

!t!t!t!t

!t+1!t+1!t+1 !t+1!

!t+2!t+2

""

""

"!

7

!V=7

6

!V=6!V=6

!V=6-!!V=6-!

6-!

6-!4

4

!V=4!V=4!V=4

!V=4-!!V=4-!

4-!4-!

5.5

!V=5.5

!V=5.5

!V=5.5-!

5.5-!

Page 65: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Hybrid heuristic •  Two standard representations for heuristics

•  Tree: values for all joint action-observation histories •  Vector: a potentially exponential number of vectors

• Key insight: exponential growth of these representations is in opposite directions

May 21, 2012 CTS Conference 2012 65

Page 66: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Optimal results

• Problem size (all 2 agent) •  Broadcast Channel S=4 A=2 Ω=2 •  Box Pushing S=100 A=4 Ω=5 •  Fire Fighting S=432 A=3 Ω=2

• Performance a combination of better search and better heuristics

May 21, 2012 CTS Conference 2012 66

0 200 400 600 800

1000

GMAA*-ICE previous best

Broadcast

0

2

4

6

8

GMAA*-ICE previous best

Fire Fighting

0 1 2 3 4 5

GMAA*-ICE previous best

Box Pushing

State-of-the-art in top down and finite-horizon

Page 67: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Approximate methods • Optimal methods are intractable for many problems • Want to produce the best solution possible with given

resources • Quality bounds are usually not possible, but these

approaches often perform well

May 21, 2012 CTS Conference 2012 67

Page 68: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Joint equilibrium search for policies (JESP) •  Instead of exhaustive search, find best response • Algorithm:

Start with (full) policy for each agent while not converged do for i=1 to n Fix other agent policies Find a best response policy for agent i

May 21, 2012 CTS Conference 2012 68

(Nair et al., 03)

Page 69: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

JESP summary •  Finds a locally optimal set of policies • Worst case complexity is the same as exhaustive search,

but in practice is much faster • Can also incorporate dynamic programming to speed up

finding best responses •  Fix policies of other agents •  Create a (augmented) POMDP using the fixed policies of others •  Generate reachable belief states from initial state b0 •  Build up policies from last step to first •  At each step, choose subtrees that maximize value at reachable

belief states

May 21, 2012 CTS Conference 2012 69

Page 70: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Memory bounded dynamic programming (MBDP)

•  Do not keep all policies at each step of dynamic programming •  Keep a fixed number for each agent: maxTrees •  Select these by using heuristic solutions from initial state •  Combines top down and bottom up approaches

May 21, 2012 CTS Conference 2012 70

(Seuken and Zilberstein., 07)

Page 71: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

MBDP algorithm start with a one-step policy for each agent for t=h to 1 do backup each agent's policy for k=1 to maxTrees do

compute heuristic policy and resulting belief state b choose best set of trees starting at b

select best set of trees for initial state b0

May 21, 2012 CTS Conference 2012 71

Page 72: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

MBDP summary •  Linear complexity in problem horizon • Exponential in the number of observations • Performs well in practice (often with very small maxTrees) • Can be difficult to choose correct maxTrees

May 21, 2012 CTS Conference 2012 72

Page 73: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Extensions to MBDP •  IMBDP: Limit the number of observations used based on •  probability at each belief (Seuken and Zilberstein 07)

• MBDP-OC: compress observations based on the value produced (Carlin and Zilberstein 08)

• PBIP: heuristic search to find best trees rather than exhaustive (Dibangoye et al., 09)

• Current state-of-the-art •  PBIP-IPG: extends PBIP by limiting the possible states (Amato et al.,

09-AAMAS) •  CBPB: uses constraint satisfaction solver for subtree selection

(Kumar and Zilberstein, 10) •  PBPG: approximate, linear programming method for subtree

selection (Wu et al, 10) – solves a problem with 3843 states and 11 obs to hor 20

May 21, 2012 CTS Conference 2012 73

Page 74: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Other finite-horizon approaches • Mixed integer linear programming (MILP) (Aras et al., 07)

•  Represent each agent's policy in sequence form (instead of as a tree)

•  Solve as a combinatorial optimization problem (MILP)

• Sampling methods •  Direct Cross-Entropy policy search (DICE) (Oliehoek et al., 08)

•  Randomized algorithm using combinatorial optimization •  Applies Cross-Entropy method to Dec-POMDPs •  Scales well wrt number of agents

•  Goal-directed sampling (Amato and Zilberstein, 09) •  Discussed later

May 21, 2012 CTS Conference 2012 74

Page 75: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Approximate infinite-horizon approaches • A large enough horizon can be used to approximate an

infinite-horizon solution, but this is neither efficient nor compact

• Specialized infinite-horizon solutions have also been developed: •  Best-First Search (BFS) •  Bounded Policy Iteration for Dec-POMDPs (Dec-BPI) •  Nonlinear Programming (NLP)

May 21, 2012 CTS Conference 2012 75

Page 76: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Memory-bounded solutions • Optimal approaches may be intractable • Can use fixed-size finite-state controllers as policies for

Dec-POMDPs • How do we set the parameters of these controllers to

maximize their value? •  Deterministic controllers - discrete methods such as branch and

bound and best-first search •  Stochastic controllers - continuous optimization

a? q

o2

o1 q?

q?

(deterministically) choosing an action and transitioning to the next node

May 21, 2012 CTS Conference 2012 76

Page 77: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Best-first search (Szer and Charpillet 05)

• Search through space of deterministic action selection and node transition parameters

• Produces optimal fixed-size deterministic controllers • High search time limits this to very small controllers (< 3

nodes)

May 21, 2012 CTS Conference 2012 77

Page 78: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Bounded policy iteration (BPI) (Bernstein et al., 05)

•  Improve the controller over a series of steps until value converges

•  Alternate between improvement and evaluation •  Improvement

•  Use a linear program to determine if a node's parameters can be changed, while fixing the rest of the controller and other agent policies

•  Improved nodes must have better value for all states and nodes of the other agents (multiagent belief space)

•  Evaluation: Update the value of all nodes in the agent's controller

•  Can solve much larger controller than BFS, but value is low due to lack of start state info and LP

May 21, 2012 CTS Conference 2012 78

Page 79: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Nonlinear programming approach (Amato et al., 07, 09b)

• Use a nonlinear program (NLP) to represent an optimal fixed-size set of controllers for Dec-POMDPs

• Consider node value as well as action and transition parameters as variables

• Maximize the value using a known start state • Constraints maintain valid values and probabilities

May 21, 2012 CTS Conference 2012 79

Page 80: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

NLP formulation Variables:

Objective: Maximize

Value Constraints:

Probability constraints ensure all probabilities must sum to 1 and be

greater than 0

∀s∈S, q ∈Q

z(q, s) = x(qi,ai ) R(s,a)+ γ P(s ' | s, a) O(o | s ', a) y(qi ',ai,qi,oi )

i∏

q '∑

o∑ z(q ', s ')

s '∑

⎣⎢

⎦⎥

i∏⎛

⎝⎜

⎠⎟

a∑

b0 (s)s∑ z(q0, s)

May 21, 2012 CTS Conference 2012 80

x(qi,ai ) = P(ai | qi ), y(qi,ai,oi,qi ') = P(qi ' | qi,ai,oi ), z(q, s) =V (q, s)

Page 81: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Mealy controllers (Amato et al, 10)

•  Controllers currently used are Moore controllers •  Mealy controllers are more powerful than Moore controllers

(can represent higher quality solutions with the same number of nodes)

• Key difference: action depends on node and observation

a1

o2

o1

o2 o1

a2

,a1

o2

o1 ,a2

o1 ,a2

,a1

o2 Moore= Mealy= Q A Q×O A

May 21, 2012 CTS Conference 2012 81

Page 82: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Mealy controllers continued

• More powerful • Provides extra structure that algorithms can use

•  Can automatically simplify representation based on informative observations

•  Can be done in controller or solution method • Can be used in place of Moore controllers in all

controller-based algorithms for POMDPs and DEC-POMDPs (not just NLP) •  Optimal infinite-horizon DP •  Approximate algorithms

o2

o1 ,a2

o1 ,a2

,a1

o2

,a1

May 21, 2012 CTS Conference 2012 82

Page 83: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Some infinite-horizon results •  Optimal algorithm can only solve very small problems •  Approximate algorithms are more scalable

•  GridSmall: 16 states, 4 actions, 2 obs •  Policy Iteration: 3.7 with 80 nodes in 821s before out of memory

May 21, 2012 CTS Conference 2012 83

Page 84: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Indefinite-horizon • Unclear how many steps are needed until termination • Many natural problems terminate after a goal is reached

•  Meeting or catching a target •  Cooperatively completing a task

May 21, 2012 CTS Conference 2012 84

Page 85: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Indefinite-horizon Dec-POMDPs (Amato et al, 09a)

• Extends indefinite-horizon POMDPs Patek 01 and Hansen 07

• Our assumptions •  Each agent possesses a set of terminal actions •  Negative rewards for non-terminal actions

• Can capture uncertainty about reaching goal • Many problems can be modeled this way

An optimal solution to this problem can be found using dynamic programming

May 21, 2012 CTS Conference 2012 85

Page 86: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Goal-directed Dec-POMDPs •  Relax assumptions, but still have goal •  Problem terminates when

•  A single agent or set of agents reach local or global goal states •  Any combination of actions and observations is taken or seen

•  More problems fall into this class (can terminate without agent knowledge)

•  Solve by sampling trajectories •  Produce only action and observation sequences that lead to goal •  This reduces the number of policies to consider

Can bound the number of samples required to approach optimality

g a1 a1 a1 o1 o3 o3 b0

May 21, 2012 CTS Conference 2012 86

Page 87: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Infinite and indefinite-horizon results

Algorithm Value Size TimeMeeting in a Grid: |S| = 16, |Ai| = 5, |Oi| = 2

Mealy 6.13 5 116HPI w/ NLP 6.04 7 16,763Moore 5.66 5 117

Goal-directed2 5.64 4 4

Box Pushing: |S| = 100, |Ai| = 4, |Oi| = 5

Mealy 143.14 4 774HPI w/ NLP 95.63 10 6,545Moore 50.64 4 5,176

Goal-directed2 149.85 5 199

Mars Rover: |S| = 256, |Ai| = 6, |Oi| = 8

Mealy 19.67 3 396HPI w/ NLP 9.29 4 111Moore 8.16 2 43

Goal-directed2 21.48 6 956

Table 3: Results for DEC-POMDP problems comparing Mealyand Moore machines and other algorithms. The size refers to thenumber of nodes in the controller and the time is in seconds.

Algorithm Value Size TimeTwo Agent Tiger: |S| = 2, |Ai| = 3, |Oi| = 2HPI w/ NLP 6.80 6 119Goal-directed 5.04 12 75

Moore -1.09 19 6,173Meeting in a Grid: |S| = 16, |Ai| = 5, |Oi| = 2

Mealy 6.13 5 116HPI w/ NLP 6.04 7 16,763

Moore 5.66 5 117Goal-directed 5.64 4 4

Box Pushing: |S| = 100, |Ai| = 4, |Oi| = 5Goal-directed 149.85 5 199

Mealy 143.14 4 774HPI w/ NLP 95.63 10 6,545

Moore 50.64 4 5,176Mars Rover: |S| = 256, |Ai| = 6, |Oi| = 8

Goal-directed 21.48 6 956Mealy 19.67 3 396

HPI w/ NLP 9.29 4 111Moore 8.16 2 43

Table 4: Results for DEC-POMDP problems comparing Mealyand Moore machines and other algorithms. The size refers to thenumber of nodes in the controller and the time is in seconds.

Number of nodesType 1 2 3 4 5

Meeting in a grid: |S| = 16, |Ai| = 5, |Oi| = 2

Mealy 5.50 6.00 5.87 6.05 6.13Moore 3.58 4.83 5.23 5.62 5.66

Box pushing: |S| = 100, |Ai| = 4, |Oi| = 5

Mealy 123.46 124.20 133.67 143.14Moore -1.58 31.97 46.28 50.64

Mars rover: |S| = 256, |Ai| = 6, |Oi| = 8

Mealy 18.92 19.17 19.67Moore 0.80 8.16

Table 5: Results for Mealy and Moore machines of different sizesfor DEC-POMDP benchmarks. A blank entry means that the con-troller of that size could not been computed given the resource re-strictions of the NEOS server.

§  Standard infinite-horizon benchmarks

§  Approximate solutions §  Can provide a very high-

quality solution very quickly in each problem

May 21, 2012 CTS Conference 2012 87

State-of-the-art in infinite-horizon approximate

Page 88: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Scalability to larger problems • Dec-MDP

•  Collective observations fully determine state

•  Independent transitions and observations •  P(s’| s, a)=P(s’| s, a1) P(s’| s, a2) •  O(o| s’, a) =P(o1| s’, a1) P(o2| s’, a2) •  Rewards still dependent

• Complexity drops to NP (from NEXP) •  Policy: nonstationary map from local states to actions si

t a • Many realistic problems have this type of independence

May 21, 2012 CTS Conference 2012 88

Page 89: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Sample-based heuristic search under submission

• Observation: can share policies before execution •  With Dec-MDP assumptions, can now estimate system state •  Don’t need history, so can use this estimate and single local obs

• Use branch and bound to search policy space •  Start from known initial state

•  Incorporate constraint optimization to more efficiently expand children

May 21, 2012 CTS Conference 2012 89

Page 90: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Dec-MDP results

•  2 agent problems

•  Scalable to 6 agents for small problems and horizons •  Note: ND-POMDP methods can scale to up to 15 agents by making

reward decomposability assumptions

May 21, 2012 CTS Conference 2012 90

T �0(⌘0) ICE IPG BLP LMPexh COP

Recylcing robot (|Z| = 4,|A| = 9)50 154.94 1.27 - 8848.7 0.016 0.560 185.71 6.00 - - 0.090 0.55570 216.47 28.6 - - 0.111 0.39590 278.01 - - - 0.151 0.373100 308.78 - - - 0.156 0.4381000 3078.0 - - - 1.440 5.374

Meeting Grid (|Z| = 81,|A| = 25)2 0.0 0.00 5 10.0 - 0.0303 0.13 0.02 17 34.7 - 0.1104 0.43 0.37 54 192.8 - 0.1145 0.89 4.38 600 571.2 - 0.1316 1.49 - - 1160.6 - 0.1597 2.19 - - 2262.9 - 0.197

10 4.68 - - 3938.5 - 0.309100 94.26 - - - - 13.121000 994.2 - - - - 33.59

T �0(⌘0) ICE IPG BLP COPMeeting on a 8x8 Grid (|Z| = 4096,|A| = 25)

5 0.0 12.55 - - 5.0516 0.0 - - - 6.207 0.71 - - - 13.168 1.67 - - - 13.769 2.68 - - - 16.94

10 3.68 - - - 18.6620 13.68 - - - 38.3730 23.68 - - - 52.3940 33.68 - - - 59.7050 43.68 - - - 74.28100 93.68 - - - 214.73

Navigation (MIT) (|Z| = 7225,|A| = 16)10 0.0 85.85 - - 47.32220 0.0 - - - 321.2640 34.97 - - - 400.4850 54.92 - - - 1061.94100 154.93 - - - 1236.11

T �0(⌘0) ICE IPG BLP COPNavigation (ISR) (|Z| = 8100,|A| = 16)

2 0.0 0.71 - 3225.8 2.393 0.0 6.50 - - 3.194 0.0 - - - 4.315 0.38 - - - 13.4310 6.47 - - - 54.1650 83.02 - - - 194.66

100 182.54 - - - 1294.28Navigation (PENTAGON) (|Z| = 9801,|A| = 16)

2 0.0 0.99 - 4915.1 1.313 0.0 6.01 - - 5.114 0.0 - - - 8.755 0.38 - - - 13.8110 4.82 - - - 62.8920 19.73 - - - 129.4830 39.18 - - - 209.1140 57.75 - - - 276.2050 76.39 - - - 1033.70

Table 1: Experimental results of the COP and exh. variants of the LMP as well as GMAA⇤-ICE (termed just ICE here), IPG,and BLP algorithms.

the best available version of the bilinear program approachwhich was the iterative best response version with standardparameters. We do not compare to the coverage set algo-rithm because the bilinear programming method has beenshown to be more efficient than it for all available test prob-lems.

We provide values for the exhaustive variant, exh, onsmall problems and constraint optimization formulation,COP, for all problems. We tested our algorithms on sixbenchmarks: recycling robot, meeting grid 3x3 and 8x8;and navigation problems2. These are the largest and hard-est benchmarks we could find in the literature. We com-pare our algorithms with: GMAA⇤-ICE, IPG, and BLP. TheGMAA⇤-ICE (Spaan, Oliehoek, and Amato 2011) heuristicsearch consistently outperforms other generic exact solverssuch as MAA⇤, GMAA⇤ . The IPG (Amato, Dibangoye, andZilberstein 2009) algorithm is a competitive alternative tothe GMAA⇤ approach and performs well on problems withreduced reachability.

In all benchmarks, the COP variant outperforms the otheralgorithms. The results, illustrated in Table 1, show that theCOP variant produces the optimal policies in much less run-time for all tested benchmarks. As an example consider themeeting in a 3x3 grid for T = 5: the COP variant com-puted the optimal policies in about 33, 4358 and 4580 timesfaster than the GMAA⇤-ICE, BLP and IPG algorithms, re-spectively. We also note that the COP variant is very usefulfor the medium and large domains. For example all large do-mains, the exh. variant ran out of memory while the COPvariant computed the optimal solutions over hundreds ofhorizons. Yet, the exh. variant can compute the optimal so-lution of small problem faster than the COP variant. For ex-ample in the recycling robot at horizon T = 1000, the exh.variant computed the optimal solution in about 5 times fasterthan the COP variant the of LMP algorithm.

While preliminary, we also note that the constraint for-mulation, allows us to deal with larger numbers of agents.Indeed, we ran the COP variant of LMP on decentralizedMDPs with independent transitions and observations where

2All problem definitions are available athttp://users.isr.ist.utl.pt/⇠ mtjspaan/decpomdp/

each agent’s local MDP is based on the recycling-robotproblem. The COP variant was able to scale up to 6 agents athorizon 10 in about 10, 000 seconds. Despite this high run-ning time, LMP is the first generic algorithm that scales toteams of more than two agents. For example the BLP algo-rithm as it currently stands can only solve two-agent prob-lems.

There are many different reasons that explain these re-sults. The LMP outperforms GMAA⇤-ICE and IPG mainlybecause they perform a policy search in the space of decen-tralized history-dependent policies. Instead, the LMP algo-rithm performs its policy search in the space of decentralizedMarkov policies, which is exponentially smaller than that ofthe decentralized history-dependent policies. The LMP out-performs the BLP algorithm mainly because of the dimen-sion of its sufficient statistic. Indeed, the BLP algorithm usesa sufficient statistic that is T times larger. More specifically,the number of bilinear terms in the BLP approach growspolynomially in the horizon of the problem, causing it tonot perform well for large problems and large horizons withtightly coupled reward values.

Conclusion and Future WorkThis paper explores new theory and algorithms for solvingindependent transition and observation Dec-MDPs. We pro-vide a new proof that optimal policies do not depend onagent histories in this subclass. We also describe a novel al-gorithm that combines heuristic search and constraint opti-mization to more efficiently produce solutions. This new al-gorithm, termed learning Markov policy or LMP, was shownto scale up to large problems and planning horizons, reduc-ing computation time by multiple orders of magnitude overprevious approaches.

In the future, we plan to explore extending the LMP al-gorithm to other classes of problems and larger teams ofagents. For instance, we may be able to produce an opti-mal solution to more general classes of Dec-MDPs or pro-vide approximate results for Dec-POMDPs. Furthermore,the scalability of our approach to larger numbers of agentsis encouraging and we will pursue methods to increase thiseven further.

State-of-the-art in IT-IO Dec-MDPs

Page 91: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Communication • Communication can be implicitly represented in Dec-

POMDP model •  Free and instantaneous communication is equivalent to

centralization • Otherwise, need to reason about what and when to

communicate

May 21, 2012 CTS Conference 2012 91

Page 92: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Dec-POMDP-COM

May 21, 2012 CTS Conference 2012 92

• A DEC-POMDP-COM can be defined with the tuple: M = <I, S, {Ai}, P, R, {Ωi}, O, Σ,CΣ>

•  The first parameters are the same as a Dec-POMDP •  Σ, the alphabet of atomic communication messages for each agent

(including a null message) •  CΣ, the cost of transmitting an atomic message: •  R, the reward model integrates this cost for the set of agents

sending message σ: • Explicitly models communication • Same complexity as Dec-POMDP

CΣ :Σ→ℜ

R(s, a,σ )

Page 93: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Algorithms using communication • Analysis of possible communication models and

complexity (Pynadath et al., 02)

• Myopic communication in transition independent Dec-MDPs (Becker et al., 09)

• Reasoning about run-time communication decisions (Nair et al., 04; Roth et al., 05)

• Stochastically delayed communication (Spaan et al., 08)

May 21, 2012 CTS Conference 2012 93

Page 94: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Summary • Optimal algorithms for Dec-POMDPs give performance

guarantees, but are often intractable •  Top down and bottom up methods provide similar performance

• Approximate Dec-POMDP algorithms are much more scalable, but (often) lack quality bounds •  Bounding memory and sampling are dominant approaches

• Using subclasses can significantly improve solution scalability (if assumptions hold)

• Communication can be helpful, but difficult to decide when and how to communicate

May 21, 2012 CTS Conference 2012 94

Page 95: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Application: Personal assistant agents (Amato et al.,11)

• People connected to many others and sources of info • Use software personal assistant agents to provide support

(Dec-MDP, Shared-MDP) •  Agents collaborate on your behalf to find resources, teams, etc. •  Goal: work more efficiently with others and discover helpful info

May 21, 2012 CTS Conference 2012 95

Page 96: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Application: Unmanned system control • Planning for autonomous vehicles in a simulation

•  Single or centralized team: factored POMDP •  Decentralized team: Dec-POMDP •  Team considering adaptive enemy: I-POMDP

May 21, 2012 CTS Conference 2012 96

Page 97: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Application: Mixed-initiative robotics •  Humans and robots collaborating for search and pursuit •  Determine what tasks robots should do (MMDP, Dec-POMDP)

and what tasks humans should do •  Adjustable autonomy based on preferences and situation

May 21, 2012 CTS Conference 2012 97

Page 98: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Conclusion • What problems Dec-POMDPs are good for

•  Sequential (not “one shot” or greedy) •  Cooperative (not single agent or competitive) •  Decentralized (not centralized execution or free, instantaneous

communication) •  Decision-theoretic (probabilities and values)

May 21, 2012 CTS Conference 2012 98

Page 99: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

Resources • My Dec-POMDP webpage

•  Papers, talks, domains, code, results •  http://rbr.cs.umass.edu/~camato/decpomdp/

• Matthijs Spaan’s Dec-POMDP page •  Domains, code, results •  http://users.isr.ist.utl.pt/~mtjspaan/decpomdp/index_en.html

• USC’s Distributed POMDP page •  Papers, some code and datasets •  http://teamcore.usc.edu/projects/dpomdp/

• Our full tutorial “Decision Making in Multiagent Systems” •  http://users.isr.ist.utl.pt/~mtjspaan/tutorialDMMS/

May 21, 2012 CTS Conference 2012 99

Page 100: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

References •  Scaling Up Optimal Heuristic Search in Dec-POMDPs via Incremental Expansion. Frans A. Oliehoek,

Matthijs T. J. Spaan and Christopher Amato. In Proceedings of the Twenty-Second International Joint Conference on Artificial Intelligence (IJCAI-11), Barcelona, Spain, July, 2011

•  Towards Realistic Decentralized Modelling for Use in a Real-World Personal Assistant Agent Scenario. Christopher Amato, Nathan Schurr and Paul Picciano. In Proceedings of the Workshop on Optimisation in Multi-Agent Systems (OptMAS) at the Tenth International Conference on Autonomous Agents and Multi-Agent Systems (AAMAS-11), Taipei, Taiwan, May 2011.

•  Finite-state controllers based on Mealy machines for centralized and decentralized POMDPs. Christopher Amato, Blai Bonet and Shlomo Zilberstein. Proceedings of the Twenty-Fourth National Conference on Artificial Intellgence (AAAI-10), Atlanta, GA, July, 2010.

•  Point-based Backup for Decentralized POMDPs: Complexity and New Algorithms. Akshat Kumar and Shlomo Zilberstein. In Proc. of the International Conference on Autonomous Agents and Multiagent Systems (AAMAS), pages 1315-1322, 2010.

•  Point-Based Policy Generation for Decentralized POMDPs, Feng Wu, Shlomo Zilberstein, and Xiaoping Chen, Proceedings of the 9th International Conference on Autonomous Agents and Multi-Agent Systems (AAMAS-10), Page 1307- 1314, Toronto, Canada, May 2010.

•  Optimizing Fixed-size Stochastic Controllers for POMDPs and Decentralized POMDPs. Christopher Amato, Daniel S. Bernstein and Shlomo Zilberstein. Journal of Autonomous Agents and Multi-Agent Systems (JAAMAS) 2009.

•  Incremental Policy Generation for Finite-Horizon DEC-POMDPs. Christopher Amato, Jilles Steeve Dibangoye and Shlomo Zilberstein. Proceedings of the Nineteenth International Conference on Automated Planning and Scheduling (ICAPS-09), Thessaloniki, Greece, September, 2009.

May 21, 2012 CTS Conference 2012 100

Page 101: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

References continued •  Event-detecting Multi-agent MDPs: Complexity and Constant-Factor Approximation.

Akshat Kumar and Shlomo Zilberstein. In Proc. of the International Joint Conference on Artificial Intelligence (IJCAI), pages 201-207, 2009.

•  Point-based incremental pruning heuristic for solving finite-horizon DEC-POMDPs. Jilles S. Dibangoye, Abdel-Illah Mouaddib, and Brahim Chaib-draa. In Proc. of the Joint International Conference on Autonomous Agents and Multi-Agent Systems , 2009.

•  Constraint-Based Dynamic Programming for Decentralized POMDPs with Structured Interactions. Akshat Kumar and Shlomo Zilberstein In Proc. of the International Conference on Autonomous Agents and Multiagent Systems (AAMAS), pages 561-568, 2009.

•  Achieving Goals in Decentralized POMDPs. Christopher Amato and Shlomo Zilberstein. Proceedings of the Eighth International Joint Conference on Autonomous Agents and Multi-Agent Systems (AAMAS-09), Budapest, Hungary, May, 2009.

•  Policy Iteration for Decentralized Control of Markov Decision Processes. Daniel S. Bernstein, Christopher Amato, Eric A. Hansen and Shlomo Zilberstein. Journal of AI Research (JAIR), vol. 34, pages 89-132, February, 2009.

•  Analyzing myopic approaches for multi-agent communications. Raphen Becker, Alan Carlin, Victor Lesser, and Shlomo Zilberstein. Computational Intelligence}, 25(1): 31—50, 2009.

•  Optimal and Approximate Q-value Functions for Decentralized POMDPs. Frans A. Oliehoek, Matthijs T. J Spaan and Nikos Vlassis. Journal of Artificial Intelligence Research, 32:289–353, 2008.

May 21, 2012 CTS Conference 2012 101

Page 102: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

References continued •  Formal Models and Algorithms for Decentralized Decision Making Under Uncertainty. Sven Seuken and

Shlomo Zilberstein. Journal of Autonomous Agents and Multi-Agent Systems (JAAMAS). 17:2, pages 190-250, 2008. •  Interaction-driven Markov games for decentralized multiagent planning under uncertainty. Matthijs T. J.

Spaan and Francisco S. Melo. Proceedings of the Joint Conference on Autonomous Agents and Multi Agent Systems, pages 525--532, 2008.

•  Mixed integer linear programming for exact finite-horizon planning in Decentralized POMDPs. Raghav Aras, Alain Dutech, and Francois Charpillet. In Int. Conf. on Automated Planning and Scheduling, 2007.

•  Value-based observation compression for DEC-POMDPs. Alan Carlin and Shlomo Zilberstein.In Proceedings of the Joint Conference on Autonomous Agents and Multi Agent Systems, 2008.

•  Multiagent planning under uncertainty with stochastic communication delays. Matthijs T. J. Spaan, Frans A. Oliehoek, and Nikos Vlassis. In Int. Conf. on Automated Planning and Scheduling, pages 338--345, 2008.

•  Improved memory-bounded dynamic programming for decentralized POMDPs. Sven Seuken and Shlomo Zilberstein. Proceedings of Uncertainty in Artificial Intelligence, July, 2007

•  Memory-Bounded Dynamic Programming for DEC-POMDPs. Sven Seuken and Shlomo Zilberstein. Proceedings of the Twentieth International Joint Conference on Artificial Intelligences (IJCAI-07), January 2007

•  Networked Distributed POMDPs: A Synthesis of Distributed Constraint Optimization and POMDPs. Ranjit Nair, Pradeep Varakantham, Milind Tambe and Makoto Yokoo. Proceedings of the Twentieth National Conference on Artificial Intelligence (AAAI-05), Pittsburgh, PA, July, 2005.

May 21, 2012 CTS Conference 2012 102

Page 103: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

References continued •  A framework for sequential planning in multi-agent settings. Piotr J. Gmytrasiewicz and

Prashant Doshi. Journal of Artificial Intelligence Research, 24:49--79, 2005. •  Bounded policy iteration for decentralized POMDPs. Daniel S. Bernstein, Eric A. Hansen, and

Shlomo Zilberstein. In Proc. Int. Joint Conf. on Artificial Intelligence, 2005. •  An optimal best-first search algorithm for solving infinite horizon DEC-POMDPs. Daniel Szer

and Francois Charpillet. In European Conference on Machine Learning, 2005. •  MAA*: A Heuristic Search Algorithm for Solving Decentralized POMDPS. Daniel Szer, Francois

Charpillet, Shlomo Zilberstein. Proceedings of the 21st International Conference on Uncertainty in Artificial Intelligence (UAI-05), Edinburgh, Scotland, July 2005

•  Reasoning about joint beliefs for execution-time communication decisions. Maayan Roth, Reid G. Simmons and Manuela M. Veloso. Proc. of Int. Joint Conference on Autonomous Agents and Multi Agent Systems, 2005

•  Dynamic Programming for Partially Observable Stochastic Games. Eric A. Hansen, Daniel S. Bernstein, and Shlomo Zilberstein. Proceedings of the 19th National Conference on Artificial Intelligence (AAAI-04), 709-715, San Jose, California, July 2004

•  Approximate solutions for partially observable stochastic games with common payoffs. Rosemary Emery-Montemerlo, Geoff Gordon, Jeff Schneider, and Sebastian Thrun. In Proceedings of the Joint Conference on Autonomous Agents and Multi Agent Systems, 2004.

•  An approximate dynamic programming approach to decentralized control of stochastic systems. Randy Cogill, Michael Rotkowitz, Benjamin Van Roy and Sanjay Lall. In Proceedings of the 2004 Allerton Conference on Communication, Control, and Computing, 2004.

•  Decentralized Control of Cooperative Systems: Categorization and Complexity Analysis. Claudia V. Goldman and Shlomo Zilberstein. Journal of Artificial Intelligence Research Volume 22, pages 143-174, 2004.

May 21, 2012 CTS Conference 2012 103

Page 104: MODELS OF COORDINATION IN MULTIAGENT DECISION …people.csail.mit.edu/camato/publications/CTStutorial.pdfOutline • Background on decision-theoretic planning • Models for cooperative

References continued •  Communication for improving policy computation in distributed POMDPs. Ranjit Nair, Milind Tambe,

Maayan Roth, and Makoto Yokoo. Proc. of Int. Joint Conference on Autonomous Agents and Multi Agent Systems, 2004.

•  Planning, Learning and Coordination in Multiagent Decision Processes. Craig Boutilier Sixth Conference on Theoretical Aspects of Rationality and Knowledge (TARK-96), Amsterdam, pp.195--210 (1996).

•  Taming Decentralized POMDPs: Towards Efficient Policy Computation for Multiagent Settings. Ranjit Nair, David Pynadath, Makoto Yokoo, Milind Tambe and Stacy Marsella. Proceedings of the Eighteenth International Joint Conference on Artificial Intelligence (IJCAI-03), Acapulco, Mexico, August, 2003.

•  The Complexity of Decentralized Control of Markov Decision Processes. Daniel S. Bernstein, Robert Givan, Neil Immerman, and Shlomo Zilberstein. Mathematics of Operations Research, 27(4):819-840, November 2002.

•  The Communicative Multiagent Team Decision Problem: Analyzing Teamwork Theories and Models. David V. Pynadath and Milind Tambe. Journal of Artificial Intelligence Research (JAIR), Volume 16, pp. 389-423. 2002.

•  Decision-theoretic control of planetary rovers. Shlomo Zilberstein, Richard Washington, Daniels S. Bernstein, and Abdel-Illah Mouaddib. In Plan-Based control of Robotic Agents, volume 2466 of LNAI, pages 270--289. Springer, 2002.

•  Reinforcement learning for adaptive routing. Leonid Peshkin and Virginia Savova. In Proceedings of the International Joint Conference on Neural Networks, 2002.

•  Decentralized control of a multiple access broadcast channel: Performance bounds. James M. Ooi and Gregory W. Wornell. In Proceedings of the 35th Conference on Decision and Control, 1996.

May 21, 2012 CTS Conference 2012 104