Constraint-Based Dynamic Programming for Decentralized...

8
Constraint-Based Dynamic Programming for Decentralized POMDPs with Structured Interactions Akshat Kumar and Shlomo Zilberstein Department of Computer Science University of Massachusetts Amherst, MA, 01003 {akshat, shlomo}@cs.umass.edu ABSTRACT Decentralized partially observable MDPs (DEC-POMDPs) provide a rich framework for modeling decision making by a team of agents. Despite rapid progress in this area, the limited scalability of solution techniques has restricted the applicability of the model. To overcome this computational barrier, research has focused on restricted classes of DEC- POMDPs, which are easier to solve yet rich enough to cap- ture many practical problems. We present CBDP, an ef- ficient and scalable point-based dynamic programming al- gorithm for one such model called ND-POMDP (Network Distributed POMDP). Specifically, CBDP provides magni- tudes of speedup in the policy computation and generates better quality solution for all test instances. It has linear complexity in the number of agents and horizon length. Fur- thermore, the complexity per horizon for the examined class of problems is exponential only in a small parameter that de- pends upon the interaction among the agents, achieving sig- nificant scalability for large, loosely coupled multi-agent sys- tems. The efficiency of CBDP lies in exploiting the structure of interactions using constraint networks. These results ex- tend significantly the effectiveness of decision-theoretic plan- ning in multi-agent settings. Categories and Subject Descriptors I.2 [Artificial Intelligence]: [Dynamic programming, Mul- tiagent systems, Intelligent agents] Keywords Multiagent planning, DEC-POMDPs 1. INTRODUCTION DEC-POMDPs have emerged in recent years as an impor- tant framework for modeling team decision problems under uncertainty. They can effectively capture situations when the environment state dynamics and the reward system de- pend upon the joint action of all agents, but agents must act based on different partial knowledge of the overall situa- tion. Many problems such as multi-robot coordination [14], Cite as: Constraint-Based Dynamic Programming for Decentralized POMDPs with Structured Interactions, Akshat Kumar, Shlomo Zilberstein, Proc. of 8th Int. Conf. on Autonomous Agents and Multia- gent Systems (AAMAS 2009), Decker, Sichman, Sierra and Castel- franchi (eds.), May, 10–15, 2009, Budapest, Hungary, pp. XXX-XXX. Copyright c 2009, International Foundation for Autonomous Agents and Multiagent Systems (www.ifaamas.org). All rights reserved. broadcast channel protocols [2], and target tracking in sen- sor networks [8] can be formulated as DEC-POMDPs. Solving DEC-POMDPs optimally is NEXP-hard [2], with optimal algorithms able to provide solution for very small problems involving two agents and small horizons [6, 17]. Approximate algorithms can scale up the horizon, but they remain limited to a few agents [14, 3]. An emerging ap- proach to improve scalability has been to consider restricted forms of interaction that arise frequently in practice [1, 11]. In particular, ND-POMDP [11] is one such general model which is inspired by a realistic sensor network coordination problem [8]. The key assumption in ND-POMDP is that of transition and observation independence and the locality of interaction, which help construct efficient algorithms that exploit the independence among the agents. A rich portfolio of algorithms has been developed for solv- ing ND-POMDPs, featuring locally optimal policy search [11], approximation schemes [18, 10] and a globally optimal algorithm [11]. Most of these algorithms are based on policy search. To the best of our knowledge, no bottom up dynamic programming (DP) algorithm exists. The advantage of the DP approach lies in its ability to focus planning on the reach- able part of the belief space using a forward heuristic search. Such approximate DP algorithms–often referred to as point- based dynamic programming techniques–have shown great success in solving POMDPs. The algorithm we introduce, Constraint-Based Dynamic Programing (CBDP), shares its motivation with such point-based approaches. The efficiency of CBDP lies in effectively capturing the independence and structure among agents using constraints. CBDP uses a factored MDP-based heuristic function to iden- tify relevant belief points. Solving a factored MDP is not easy, as the space of both the joint actions and states is expo- nential in the number of agents. We provide a novel way to approximate this value function efficiently using constraint optimization. The planning (backup operation in DP) for these beliefs is performed using a constraint-based formu- lation rather than a naive search through the exponential number of policies. Such formulation makes the complexity of CBDP exponential only in the induced width of the depth- first search (DFS) ordering of the agent interaction graph, while linear in the number of agents. This enhances the scalability of CBDP for our sensor network domain, allow- ing us to solve large loosely-coupled multi-agent planning problems. Our results show that CBDP can handle very large domains. It can solve efficiently problems with hun- dreds of horizon steps while providing orders of magnitude time savings over the previous best algorithm.

Transcript of Constraint-Based Dynamic Programming for Decentralized...

Page 1: Constraint-Based Dynamic Programming for Decentralized ...rbr.cs.umass.edu/shlomo/papers/KZaamas09.pdfof interactions using constraint networks. These results ex-tend significantly

Constraint-Based Dynamic Programming forDecentralized POMDPs with Structured Interactions

Akshat Kumar and Shlomo ZilbersteinDepartment of Computer Science

University of MassachusettsAmherst, MA, 01003

akshat, [email protected]

ABSTRACTDecentralized partially observable MDPs (DEC-POMDPs)provide a rich framework for modeling decision making bya team of agents. Despite rapid progress in this area, thelimited scalability of solution techniques has restricted theapplicability of the model. To overcome this computationalbarrier, research has focused on restricted classes of DEC-POMDPs, which are easier to solve yet rich enough to cap-ture many practical problems. We present CBDP, an ef-ficient and scalable point-based dynamic programming al-gorithm for one such model called ND-POMDP (NetworkDistributed POMDP). Specifically, CBDP provides magni-tudes of speedup in the policy computation and generatesbetter quality solution for all test instances. It has linearcomplexity in the number of agents and horizon length. Fur-thermore, the complexity per horizon for the examined classof problems is exponential only in a small parameter that de-pends upon the interaction among the agents, achieving sig-nificant scalability for large, loosely coupled multi-agent sys-tems. The efficiency of CBDP lies in exploiting the structureof interactions using constraint networks. These results ex-tend significantly the effectiveness of decision-theoretic plan-ning in multi-agent settings.

Categories and Subject DescriptorsI.2 [Artificial Intelligence]: [Dynamic programming, Mul-tiagent systems, Intelligent agents]

KeywordsMultiagent planning, DEC-POMDPs

1. INTRODUCTIONDEC-POMDPs have emerged in recent years as an impor-

tant framework for modeling team decision problems underuncertainty. They can effectively capture situations whenthe environment state dynamics and the reward system de-pend upon the joint action of all agents, but agents mustact based on different partial knowledge of the overall situa-tion. Many problems such as multi-robot coordination [14],

Cite as: Constraint-Based Dynamic Programming for DecentralizedPOMDPs with Structured Interactions, Akshat Kumar, Shlomo Zilberstein,Proc. of 8th Int. Conf. on Autonomous Agents and Multia-gent Systems (AAMAS 2009), Decker, Sichman, Sierra and Castel-franchi (eds.), May, 10–15, 2009, Budapest, Hungary, pp. XXX-XXX.Copyright c© 2009, International Foundation for Autonomous Agents andMultiagent Systems (www.ifaamas.org). All rights reserved.

broadcast channel protocols [2], and target tracking in sen-sor networks [8] can be formulated as DEC-POMDPs.

Solving DEC-POMDPs optimally is NEXP-hard [2], withoptimal algorithms able to provide solution for very smallproblems involving two agents and small horizons [6, 17].Approximate algorithms can scale up the horizon, but theyremain limited to a few agents [14, 3]. An emerging ap-proach to improve scalability has been to consider restrictedforms of interaction that arise frequently in practice [1, 11].In particular, ND-POMDP [11] is one such general modelwhich is inspired by a realistic sensor network coordinationproblem [8]. The key assumption in ND-POMDP is that oftransition and observation independence and the locality ofinteraction, which help construct efficient algorithms thatexploit the independence among the agents.

A rich portfolio of algorithms has been developed for solv-ing ND-POMDPs, featuring locally optimal policy search[11], approximation schemes [18, 10] and a globally optimalalgorithm [11]. Most of these algorithms are based on policysearch. To the best of our knowledge, no bottom up dynamicprogramming (DP) algorithm exists. The advantage of theDP approach lies in its ability to focus planning on the reach-able part of the belief space using a forward heuristic search.Such approximate DP algorithms–often referred to as point-based dynamic programming techniques–have shown greatsuccess in solving POMDPs. The algorithm we introduce,Constraint-Based Dynamic Programing (CBDP), shares itsmotivation with such point-based approaches.

The efficiency of CBDP lies in effectively capturing theindependence and structure among agents using constraints.CBDP uses a factored MDP-based heuristic function to iden-tify relevant belief points. Solving a factored MDP is noteasy, as the space of both the joint actions and states is expo-nential in the number of agents. We provide a novel way toapproximate this value function efficiently using constraintoptimization. The planning (backup operation in DP) forthese beliefs is performed using a constraint-based formu-lation rather than a naive search through the exponentialnumber of policies. Such formulation makes the complexityof CBDP exponential only in the induced width of the depth-first search (DFS) ordering of the agent interaction graph,while linear in the number of agents. This enhances thescalability of CBDP for our sensor network domain, allow-ing us to solve large loosely-coupled multi-agent planningproblems. Our results show that CBDP can handle verylarge domains. It can solve efficiently problems with hun-dreds of horizon steps while providing orders of magnitudetime savings over the previous best algorithm.

Page 2: Constraint-Based Dynamic Programming for Decentralized ...rbr.cs.umass.edu/shlomo/papers/KZaamas09.pdfof interactions using constraint networks. These results ex-tend significantly

N

SEW

N

SEW

N

SEWloc1 loc2

1 2 3

Figure 1: A 3-chain sensor network configuration.Each sensor can scan in 4 directions. Target movesstochastically between loc1 and loc2.

2. BACKGROUNDThis section briefly introduces the motivating sensor net-

work domain and the ND-POMDP model. More details canbe found in [8, 11].

2.1 Coordinated target trackingOur target application is illustrated in Figure 1, which

shows a 3-chain sensor network configuration. Each sensorhas scanning ability in four directions (North, South, East,West). The target moves stochastically between loc1 andloc2. To track the target and obtain the associated rewardthe sensors with the overlapping area must scan the areatogether. For example, if the target is present in loc1 thensensor 1 must scan East and sensor 2 must scan West totrack it. Thus, coordination plays an important role. Themovement of targets is unaffected by sensor agents. Sen-sors have imperfect observability of the target, so there canbe false positive and negative observations. On successfullytracking a target, sensors receive a reward. Otherwise theyincur a cost, when they scan an area in an uncoordinatedfashion or when the target is absent. The specifications ofthe model are taken from [11].

2.2 ND-POMDP modelThe problem of target tracking in sensor networks can be

easily represented as a DEC-POMDP [2]. However, explicitutilization of independence among the agents can help usconstruct more efficient algorithms. For example, sensors inour domain are mostly independent–their state transitionsgiven the target location and the observations are indepen-dent of the actions of the other agents. Agents are coupledonly through the joint reward function. The ND-POMDPmodel [11] can express such independence relations. It isdescribed below for a group of n sensor agents.

S = ×1≤i≤nSi × Su. Si refers to the local state of agenti. Su refers to a set of uncontrollable states that areindependent of the actions of the agents. In the sensornetwork example, Si is empty, while Su corresponds tothe set of locations where targets can be present.

A = ×1≤i≤nAi where Ai is the set of actions for agent i.For the sensor network example A1 = N, W, E, S,Off.

Ω = ×1≤i≤nΩi is the joint observation set.

P P (s′|s, a) = Pu(s′u|su) · Π1≤i≤nPi(s′i|si, su, ai), where

a = 〈a1, . . . , an〉 is the joint action performed in jointstate s = 〈su, s1, . . . , sn〉 resulting in joint state s′ =〈s′u, s′1, . . . , s

′n〉. This relies on transition independence.

O O(ω|s, a) = Π1≤i≤nOi(ωi|si, su, ai), where s is the jointstate resulting after taking joint action a and receivingjoint observation ω. This relies on observation inde-pendence.

R R(s, a) = ΣlRl(sl, su, al). The reward function is de-composable among sub groups of agents referred byl. If k agents i1, . . . , ik are involved in a particularsub group l, then sl denotes the state along the linki.e. 〈si1 , . . . , sik 〉. Similarly, al = 〈ai1 , . . . , aik 〉. Inthe sensor domain in Figure 1, the reward is decom-posed among two subgroups (sensor 1–sensor 2) and(sensor 2–sensor 3), hence R = R12(s1, s2, su, a1, a2)+R23(s2, s3, su, a2, a3). Based on the reward function,an interaction hypergraph can be constructed. A hy-perlink l connects the subset of agents which form thereward component Rl. This hypergraph visually de-picts the dependency among agents and is an impor-tant tool for the complexity analysis in later sections.

bo bo = (bu, b1, . . . , bn) is the initial belief for joint states = 〈su, s1, . . . , sn〉 and b(s) = b(su) ·Πn

i=1b(si).

Solving the above decision problem requires to compute thejoint policy π = 〈π1, . . . , πn〉 that maximizes the total ex-pected reward of all agents over a finite horizon T startingfrom the initial belief bo.

The following theorem illustrates an important advantageof transition and observation independence in ND-POMDP.

Theorem 1. Given transition and observation indepen-dence, and a decomposable reward function, the value of ajoint policy π is also decomposable. That is,

V tπ(st,−→ω t) =

Xl

V tπl

(stl , s

tu,−→ω t

l).

Hence, the value function of a policy in ND-POMDP can berepresented as a sum of local value functions, one for eachlink. For proof please refer to [11].

3. POINT-BASED DYNAMICPROGRAMMING

The intractability of optimal (DEC) POMDP algorithmscan be attributed to planning for the complete belief space.DEC-POMDPs are further disadvantaged as the joint pol-icy has to be represented explicitly and cannot be extracteddirectly from the value function. The idea of planning for afinite set of belief points has been explored by Lovejoy [9],leading to several successful point-based POMDP algorithmssuch as PBVI [13], Perseus [16], and FSVI [15] among others.For DEC-POMDPs, a similar point-based algorithm calledMBDP has been developed by Seuken and Zilberstein [14].However, extending MBDP to multiple agents (>2) is non-trivial because the computational cost of planning as well asthe MDP heuristic will require exponential effort. Even rep-resenting the problem parameters (reward, transition func-tion) will be exponential in the number of agents, clearly in-feasible. CBDP shares its motivation with these successfulpoint-based algorithms. However, it is significantly differentfrom MBDP as it exploits the structure of interactions inthe ND-POMDP, whereas MBDP is oblivious to such struc-ture. Next, we presents several novel techniques we employin CBDP to enable efficient planning for multiple agents.

Algorithm 1 shows the general point-based dynamic pro-gramming framework that CBDP follows. The input is afinite horizon problem with T decision steps. The algorithmfirst computes the heuristic function used for sampling be-lief points for each time step t. Then, the main bottom-up

Page 3: Constraint-Based Dynamic Programming for Decentralized ...rbr.cs.umass.edu/shlomo/papers/KZaamas09.pdfof interactions using constraint networks. These results ex-tend significantly

Algorithm 1: Point-based DP framework

T ← horizon of the problem1

Belief set B ← 2

Precompute the belief selection heuristic3−→QT ← all one step policy trees4

for t = T − 1 until 1 do5 eQt ← DoBackup(−→Q t+1)6

Populate B using heuristic function7−→Q t ← findBest( eQt,B)8

Output−→Q1 as the solution9

dynamic programming loop starts. For each step t, the set−→Q t = qt

i |1 ≤ i ≤ n contains the best policies for eachagent. The set qt

i includes policies for agent i that con-tributed to the best joint policy for some sampled belief.

The algorithm starts with the last time step, where thereare only immediate actions to take. These actions, which are

relatively few, are stored in the set−→QT without any belief

sampling. In each further iteration, the backup of the last

step’s best policies is stored in eQt. Then, a desired numberof beliefs are sampled using the precomputed heuristic func-tion. For each of the sampled beliefs, the best joint policy

is extracted from the set eQt and stored in−→Q t, which forms

the basis for the next iteration. Finally, at t = 1, the bestpolicy for the initial belief bo is found and returned as thesolution. Note that the space and time complexity is definedby the number of sampled beliefs as planning is restrictedonly to this set. This is the key factor in the efficiency andscalability of point-based approaches.

We now examine the main steps of Algorithm 1 as imple-mented in CBDP: 1) Computation of the heuristic function;2) Belief sampling using heuristic function; and 3) Findingthe best joint policy for a particular belief. Contrary to asingle-agent POMDP, each of these steps is non-trivial formultiple agents. First, the well known MDP heuristic forPOMDPs translates to a factored MDP, which is difficultto solve due to exponential state and action space [7]. Sec-ond, belief sampling requires reasoning about joint statesand actions, which again requires exponential effort if donenaively. Finally, computing the best joint policy for a givenbelief could require search through the exponential numberof candidates. We show in the following sections how weperform efficiently each one of these steps of CBDP.

3.1 Heuristic function computationWe use as a heuristic the factored MDP derived from the

original ND-POMDP model by ignoring the observations.The optimal policy of this MDP is used for belief samplingin CBDP. One would intuitively believe that the factoredstate and the reward dynamics of this MDP will make thevalue function structured, making it easier to compute theoptimal policy. Unfortunately, this is not the case. Thevalue function eventually depends on all the states as hasbeen shown by Koller and Parr [7]. This makes the policyharder to compute. To address this, we introduce a way toapproximate the value function such that it becomes struc-tured and efficiently computable.

To simplify the equations, we introduce some shorthandnotation. Let rt

l denote the immediate reward along the link

l at time step t when the involved agents perform the actional, i.e. rt

l = Rl(stl , s

tu, al). Let pt

u denote the transition ofthe unaffected state at time t, i.e. pt

u = P (st+1u |st

u). Letpt

i denote the transition probability of agent i, i.e. pti =

Pi(st+1i |st

i, stu, ai). Similar shorthand notation can be de-

fined for a link l as well: ptl is the product of the transition

probabilities of all the agents (say k) involved in the link l,i.e. pt

l = pl1 . . . plk . The context of states and actions willbecome clear from the usage.

The MDP value of any joint state st−1, V (st−1), can bedefined as follows:

V (st−1) = maxa

(R(st−1, a) +Xst

P (st|st−1, a)V (st)).

which can be simplified as follows using the above notationand the transition independence of the factored MDP.

V (st−1) = maxa

(X

l

rt−1l +

Xst

u,st1...st

n

pt−1u pt−1

1 . . . pt−1n V (st))

(1)The time complexity of solving Eq. 1 for all horizons

is O(T |Ai|n|Su|2|Si|2n). The presence of two exponentialterms, |Ai|n and |Si|2n, makes it harder to compute. We

inductively define an approximate function V , which is anupper bound on the optimal value function. V is decompos-able along the links and is thus easier to compute.

For the last time step T , V is defined as the sum of im-mediate rewards for each link. Thus, it is trivially decom-posable along the links. Let us assume inductively that fortime step t, V is decomposable, i.e. V (st) =

Pl V t

l (stust

l |st).The conditioning over the current state is necessary as thetwo different joint states s, s′ may share the state compo-nent susl along the link l, and the value V (susl) may bedifferent in states s and s′. For example, let s = sus1s2s3,s′ = sus1s2s4 and the state along a link be s1s2. The valueV (sus1s2) can be different for s and s′ although they sharethe state along the link.

The derivation of V for the previous step t−1 is as follows.Define a new function V ′ for the step t− 1:

V ′(st−1) = maxa

Xl

rt−1l +

Xst

pt−1u . . . pt−1

n V (st)

Clearly, V ′(st−1) is an upper bound on the exact value

V (st−1) because we assumed V (st) was an upper bound on

V (st). In the following we derive V from V ′

V ′(st−1) = maxa

Xl

rt−1l +

Xst

pt−1u . . . pt−1

n V (st)

= maxa

Xl

rt−1l +

Xst

pt−1u . . . pt−1

n

Xl

V tl (st

ustl |st)

≤ maxa

Xl

rt−1l +

Xst

u,st1...st

n

pt−1u . . . pt−1

n

Xl

maxst

V tl| z

The quantity under the brace is uniquely defined for each

of the link l and thus can be replaced with eVlt(st

ustl) =

maxst V tl (st

ustl |st). Intuitively, this approximation will per-

form better when the sensor agents along a link in a particu-lar state (sl) with a given external state (su) provide similarcontribution to the value function across all joint states (st).

Upon rearranging the terms of the previous equation we

Page 4: Constraint-Based Dynamic Programming for Decentralized ...rbr.cs.umass.edu/shlomo/papers/KZaamas09.pdfof interactions using constraint networks. These results ex-tend significantly

A1

A2

A4

A3

Q12

Q13

Q24

Q34

Figure 2: A coordination graph for 4 agents

get

= maxa

Xl

rt−1l +

Xl

Xst

u,st1...st

n

pt−1u . . . pt−1

neVl

t

= maxa

Xl

rt−1l +

Xl

Xst

u,stl,st−l

pt−1u pt−1

l pt−1−l

eVlt

st−1l refers to the state of agents along the link l, st

−l refers tothe state of the rest of the agents. Similarly, pt−1

−l abbreviatesthe transition probability product of agents except those onlink l. pt−1

−l is independent of the rest of the terms and will

sum up to 1. Finally, we obtain the decomposable V whichwas our goal.

V (st−1) = maxa

Xl

(rt−1l +

Xst

u,stl

pt−1u pt−1

leVl

t) (2)

3.1.1 Joint action selectionSolving Eq. 2 still requires iterating over exponential num-

ber of actions. We now show how it can be solved effi-ciently for each state st using coordination graphs. Thesegraphs, which were first introduced in the context of fac-tored MDPs [5], are essentially constraint networks whosesolution corresponds to the solution of the underlying con-straint optimization problem.

A coordination graph is very similar to the interactiongraph defined using reward functions. There is a node foreach agent and an edge (or an hyper-edge) connects nodesfor agents that are part of the same reward component Rl.Agents have Q functions defining the local rewards, whichalso depends upon some other agents–exactly those whichshare the same reward component. For example, in Figure 2agent A2 has two Q functions, Q12 shared with agent A1

along the link A1 − A2 and Q24. Eq. 2 can be solved forevery state st by formulating a joint task for the agents tofind the joint action a? that maximizes the sum of all Qfunctions

Pl Ql. The Q function is defined for each link l

as follows.

st : Ql(al) = rtl +

Xst+1

u ,st+1l

ptupt

leVl

t+1

The above optimization problem can be easily solved us-ing techniques from constraint optimization. In particular,the bucket elimination algorithm [4] can solve such problemsefficiently using dynamic programming. The complexity ofbucket elimination is exponential in the induced width ofthe chosen DFS ordering of the coordination graph and lin-ear in the number of nodes of the graph which is the sameas the number of agents. Note that the induced width ismuch smaller than the number of agents. For a tree-likenetwork, the induced width is 1 no matter how many agents

are involved. Therefore, coordination graphs can be solvedefficiently for large loosely coupled systems.

3.1.2 ComplexityComputing the policy for our factored MDP requires solv-

ing Eq. 2 for each state and time step t using coordina-tion graphs. The time complexity is O(nT |Su|2|Si|2n|Ai|d)where d is the induced width of the coordination graph. Thespace complexity is O(T |Su||Si|n + n|Ai|d). The term |Ai|drefers to the space requirement of the bucket elimination al-gorithm. The exponential factor corresponding to the statespace still remains. However, note that for the sensor net-work domain, the internal states are empty. This makesthe time complexity O(nT |Su|2||Ai|d) and space complexityO(T |Su| + n|Ai|d) which are polynomial in the state spaceand linear in the number of agents. Hence, for the sensornetwork domain we can effectively solve large problems withlow induced width.

3.2 Belief samplingThis section describes the belief sampling for any horizon

t. We use a portfolio of three heuristic functions: one basedpurely on the underlying MDP, another taking observationsinto account, and finally the random heuristic.

3.2.1 MDP-based samplingIn the pure MDP-based sampling, the underlying joint

state is revealed after each step. In POMDPs this is easyas we draw a state from the current belief b. However, in aDEC-POMDP the belief is over joint states, hence one hasto reason about the joint state space. But fortunately, ina ND-POMDP, b(s) = bu(su)b1(s1) . . . bn(sn). This allowsus to sample each factored state from its corresponding dis-tribution. Once we get the current underlying joint statest = (st

u . . . stn), we select the action a? using the precom-

puted MDP policy. Then, the next unaffected state is sam-pled from the distribution st+1

u = Sample(P (·|stu)). Each

state st+1i can be sampled as st+1

i = Sample(P (·|sti, s

tu, a?

i )).Repeated simulations using such a scheme gives us a beliefpoint for any horizon starting from the initial belief bo.

3.2.2 Direct belief space explorationOne shortcoming of the MDP heuristic is that it ignores

the effect of observations which are important to determinethe belief state accurately in a partially observable environ-ment. To overcome this limitation, Shani et al. proposeda new MDP-based heuristic that directly explores the be-lief space [15]. Our second heuristic is motivated by thisapproach, presented in the Algorithm 2. The algorithm re-turns a belief after applying a desired number of updates.The first few steps are similar to those of the pure MDPheuristic. First, the current underlying state is sampledfrom the current belief bt, the MDP action for it is found andthe next state is sampled using the corresponding transitionfunctions (steps 3 to 6 of the algorithm). Next, the mostprobable observation is sampled for each agent i (step 7).Then, the belief for each state component is updated usingthe action a? and the observation ω = (ω1 . . . , ωn). This up-dated belief is passed to the next iteration of BeliefExploreuntil the desired number of updates are performed (step 1).The call to BeliefExplore is initialized using the initial beliefbo. Further discussion of belief updates follows.

The belief update of b′ based on b, action a and obser-

Page 5: Constraint-Based Dynamic Programming for Decentralized ...rbr.cs.umass.edu/shlomo/papers/KZaamas09.pdfof interactions using constraint networks. These results ex-tend significantly

Algorithm 2: BeliefExplore(bt)

if t == updatesRequired then1

return bt2

Sample st from bt3

a? ← GetMDPAction(st)4

st+1u ← Sample(P (·|st

u))5

st+1i ← Sample(P (·|st

i, stu, a?

i )) ∀i = 1 to n6

ωi ← Sample(O(·|st+1i , st+1

u , a?i )) ∀i = 1 to n7

bt+1u ← τu(bt

u, a?, ω)8

bt+1i ← τi(b

ti, a

?, ω) ∀i = 1 to n9

return BeliefExplore(bt+1)10

vation ω is represented by b′ = τ(b, a, ω). Performing suchupdates on the joint belief space is costly. Therefore, we usean approximation scheme which updates the factored beliefspace individually, i.e. the current belief bt = bt

ubt1 . . . bt

n isupdated as bt+1 = bt+1

u bt+11 . . . bt+1

n . The approximation is ifwe are updating the belief over Si then the rest of the jointstate is fixed to the state sampled using MDP policy.

bt+1u (su)a,ω = α

nYi=1

Oi(ωi| st+1i|z, su, ai)

Xs′u

btu(s′u)P (su|s′u)

The above equation shows the update of the unaffected statespace. The part of the joint state other than the unaffectedstate is fixed (see the term under the brace). Each st+1

i isfound using the MDP policy (in step 6 of Algorithm 2). α isthe normalization constant. Similarly, other factored statespaces can be updated as

bt+1i (si)

a,ω = αOi(ωi|si, st+1u , ai)

n\iYj=1

Oj(ωj |st+1j , st+1

u , aj) ·Xs′i

bti(s

′i)P (si|s′i, st+1

u , ai)

3.3 Best joint policy computationThis section describes the last step of the DP framework–

how to find the best policy for a sampled belief point. Afterthe backup of policies from the last iteration, each agentcan have a maximum of |Ai||B||Ωi| policies where |B| refersto the number of sampled belief points, which can be fixeda priori to a threshold maxBelief. Thus, the total numberof joint policies can be O((|Aj |maxBelief |Ωk|)n) where Aj

is the maximum action space of any agent and Ωk is themaximum observation space. Searching for the best jointpolicy will require exponential time in the number of agentsif done naively. We formulate below a constraint optimiza-tion problem for this task, which performs much better thannaive search.

The value for a joint belief b upon executing the jointpolicy π is given by

Vπ(b) =X

s

b(s)Vπ(s, 〈〉)

Using the independence relations in ND-POMDP and The-

orem 1, we get

Vπ(b) =X

su,s1,...,sn

bu(su)b1(s1) . . . bn(sn)X

l

Vπl(sl, su, 〈〉)

=X

l

Xsu,sl,s−l

bu(su)bl(sl)b−l(s−l)Vπl(sl, su, 〈〉)

where bl(sl) refers to the product of beliefs for agents onthe link l and b−l refers to the rest of the agent. b−l isindependent of Vπl and will sum up to 1. So we are left with

Vπ(b) =X

l

Xsu,sl

bu(su)bl(sl)Vπl(sl, su, 〈〉) (3)

The above equation shows that the value of a policy for abelief can be decomposed into contributions along the inter-action links. The best joint policy for a given belief is thusthe solution of the following optimization problem:

Vπ?(b) = maxπ

Xl

Xsu,sl

bu(su)bl(sl)Vπl(sl, su, 〈〉)

This problem can be solved using a constraint graph similarto the coordination graphs defined in Section 3.1.1. There isa policy variable Πi for each agent i. The domain D(Πi) =πi is the set of all backed up policies of agent i. VariablesΠi and Πj are connected by an edge if their correspondingagents i and j share a reward component Rl. The struc-ture of this graph is similar to the interaction graph, exceptthat the agents are replaced by their corresponding policyvariables. A constraint is defined for each link l with thevaluation Cl(πl) =

Psu,sl

bu(su)bl(sl)Vπl(sl, su, 〈〉). For ex-ample, consider the interaction structure of Figure 2. A con-straint is defined for each of the 4 links and C12(π1, π2) =P

su,s1,s2bu(su)b1(s1)b2(s2)Vπ1π2(s1, s2, su, 〈〉).

The optimization problem is to find the complete assign-ment π? to the variables which maximizes the sum

Pl Cl,

which is also the best joint policy for the particular belief.Using the bucket elimination algorithm, this problem can besolved in O(n(|Aj |maxBelief |Ωk|)d) time and space where dis the induced width of the interaction graph. Thus, wehave reduced the complexity from being exponential in thenumber of agents to being exponential in the induced width,which typically will be much smaller and linear in the num-ber of agents.

4. COMPLEXITY

Theorem 2. CBDP has a linear time and space complex-ity with respect to the problem horizon.

Proof. CBDP plans only for the sampled belief set whosesize can be fixed to maxBelief. Hence, the number of pol-icy trees for each agent is between QLB = maxBelief (as-suming each sampled belief requires a unique policy of theagent) and QUB = |Aj |maxBelief|Ωk|. For each horizon, onlyQLB policy trees are retained. To construct the new treesafter backup, we can use a pointer sharing mechanism inwhich a tree from the previous iteration can be part of mul-tiple new policy trees. Thus, the size of a T horizon policycan be O(QLB · T ). For n agents the total space requiredis O(n · (QLB .T + QUB)). Taking into account the addi-tional space required by the bucket elimination algorithmfor the best policy computation, the total space required is

Page 6: Constraint-Based Dynamic Programming for Decentralized ...rbr.cs.umass.edu/shlomo/papers/KZaamas09.pdfof interactions using constraint networks. These results ex-tend significantly

on the sensor network domain provided in Section 2. Our first ex-perimental result focuses on providing a head-to-head comparisonof FANS with different heuristics, against SPIDER, since SPIDERis the latest algorithm for ND-PODMPs, and also because it usesbranch-and-bound search using an MDP heuristic. The comparisonuses some of the sensor network configurations presented in [14],and an additional “7-H” configuration as shown in Figure 3. In thisexperiment, we limit the number of agents to less than 10, in orderto allow SPIDER to run.

4-Chain

5-Star 5-P

7-H 11-Helix15-3D

Figure 3: Sensor network configurations

Figure 4 shows a comparison of runtime of SPIDER vs. FANSwith different heuristics. The x-axis shows the different sensor net-work domain configurations, and the y-axis plots the runtime inlog-scale (in all our experiments, we imposed a 10000 second cut-off on runtime). Also, we chose a time horizon of 3 for policycomputation, to be consistent with the results shown in [14]. Forthe node heuristic, we chose k = 1/2; we discuss this choice of klater on. We can observe that (i) SPIDER is the slowest of all ofthe different algorithms considered and it hits the 10000 sec cut-off in the 5-P and 7-H cases; (ii) The Node and Link heuristicsare the two fastest in all cases; (iii) Unfortunately, Searcher’s extrasearch over FSMs appears have a significant runtime penalty; (iv)The Equality and Fairness heuristics are the two worst perform-ing heuristics; (v) Surprisingly, the Greedy heuristic matches theperformance of the Link heuristic for the 5-P and 7-H cases. Thekey conclusion from this graph is that FANS with either the Node,Link or Greedy heuristic is able to provide more than two orders ofmagnitude speedup over SPIDER.

Figure 4: Runtime comparison of SPIDER vs. FANS

Next, we show scale up in terms of time horizon for policy com-putation, and also provide a comparison of FANS against the LID-JESP algorithm, a locally optimal algorithm for ND-POMDPs [10].Runtime results for LID-JESP are available up to T=6, as shownbelow. Figure 5 shows a plot of the runtime as we increase the

time horizon. The x-axis shows the time horizon, and the y-axisplots the runtime in log-scale. Runtime values for different timehorizons are shown for LID-JESP and FANS with Node heuristic(k = 1/2), for the 4-chain and 5P configurations. The key obser-vations here are that FANS is able to provide orders of magnitudespeedup over LID-JESP for higher time horizons. For example, forthe 4-chain configuration with T=6, LID-JESP runs for 309 s, whileFANS terminates in 0.73 s, thus providing a 423-fold speedup overLID-JESP. Moreover, for the 5-P configuration, a speedup of morethan 1 order of magnitude is visible for T=5. We discuss solutionquality in detail later, but for both 4-chain and 5-P, the Node heuris-tic with k = 1/2 appears to match the quality of the global optimalsolution.

Figure 5: Runtime comparison of FANS with LID-JESP

Figure 6(a) shows a comparison of runtime of FANS with dif-ferent heuristics as we scale-up the number of agents. The x-axisshows the different sensor network domain configurations2 (seeFigure 3), and the y-axis plots the runtime in log-scale (again, inall our experiments, we imposed a 10000 second cutoff on run-time). Also, the time horizon for policy computation is 3 and weagain choose k = 1/2 for the Node heuristic. We can make thefollowing observations: (i) The Equality and Searcher heuristicsare not able to terminate within the cutoff time of 10000 s; (ii)The Node heuristic still remains the best for all configurations, butGreedy beats Link in the 15-3D domain. Thus FANS can scale upto as many as 15 agents, triple the number of agents SPIDER couldhandle.

The next set of experiments examine tradeoffs in solution qualityand provide analysis of these speedups. Table 1 shows a compar-ison of solution quality for SPIDER and FANS with select heuris-tics. The rows show all agent configurations considered so far whilethe columns show the algorithm used to obtain the solution quality.We can observe the following:(i) For the 4-chain and 5-star do-mains, FANS (even with its faster Greedy heuristic) is able to reachthe same quality as SPIDER, which is the global optimum. Forthe 5-P configuration, only an upper bound on the solution qualityis provided in [14], and FANS is able to obtain a solution qualitywithin 5% of the upper bound obtained by SPIDER. (ii) The Equal-ity heuristic obtains the highest solution quality in all cases; (iii)For domains of less than 10 agents, the Link heuristic matches thequality of the Equality heuristic (it loses less than 1% quality andless than 10% of quality in the 5-star and 7H domains respectively);(iv) As we scale up the Network, none of the three fastest heuristics(Link, Node, Greedy) can keep up to the solution quality providedby Equality or Searcher; (v) Even though the Equality and Searcher

2Configuration 15-mod is a modification to the 15-3D configura-tion with different target paths.

(a)

on the sensor network domain provided in Section 2. Our first ex-perimental result focuses on providing a head-to-head comparisonof FANS with different heuristics, against SPIDER, since SPIDERis the latest algorithm for ND-PODMPs, and also because it usesbranch-and-bound search using an MDP heuristic. The comparisonuses some of the sensor network configurations presented in [14],and an additional “7-H” configuration as shown in Figure 3. In thisexperiment, we limit the number of agents to less than 10, in orderto allow SPIDER to run.

4-Chain

5-Star 5-P

7-H 11-Helix15-3D

Figure 3: Sensor network configurations

Figure 4 shows a comparison of runtime of SPIDER vs. FANSwith different heuristics. The x-axis shows the different sensor net-work domain configurations, and the y-axis plots the runtime inlog-scale (in all our experiments, we imposed a 10000 second cut-off on runtime). Also, we chose a time horizon of 3 for policycomputation, to be consistent with the results shown in [14]. Forthe node heuristic, we chose k = 1/2; we discuss this choice of klater on. We can observe that (i) SPIDER is the slowest of all ofthe different algorithms considered and it hits the 10000 sec cut-off in the 5-P and 7-H cases; (ii) The Node and Link heuristicsare the two fastest in all cases; (iii) Unfortunately, Searcher’s extrasearch over FSMs appears have a significant runtime penalty; (iv)The Equality and Fairness heuristics are the two worst perform-ing heuristics; (v) Surprisingly, the Greedy heuristic matches theperformance of the Link heuristic for the 5-P and 7-H cases. Thekey conclusion from this graph is that FANS with either the Node,Link or Greedy heuristic is able to provide more than two orders ofmagnitude speedup over SPIDER.

Figure 4: Runtime comparison of SPIDER vs. FANS

Next, we show scale up in terms of time horizon for policy com-putation, and also provide a comparison of FANS against the LID-JESP algorithm, a locally optimal algorithm for ND-POMDPs [10].Runtime results for LID-JESP are available up to T=6, as shownbelow. Figure 5 shows a plot of the runtime as we increase the

time horizon. The x-axis shows the time horizon, and the y-axisplots the runtime in log-scale. Runtime values for different timehorizons are shown for LID-JESP and FANS with Node heuristic(k = 1/2), for the 4-chain and 5P configurations. The key obser-vations here are that FANS is able to provide orders of magnitudespeedup over LID-JESP for higher time horizons. For example, forthe 4-chain configuration with T=6, LID-JESP runs for 309 s, whileFANS terminates in 0.73 s, thus providing a 423-fold speedup overLID-JESP. Moreover, for the 5-P configuration, a speedup of morethan 1 order of magnitude is visible for T=5. We discuss solutionquality in detail later, but for both 4-chain and 5-P, the Node heuris-tic with k = 1/2 appears to match the quality of the global optimalsolution.

Figure 5: Runtime comparison of FANS with LID-JESP

Figure 6(a) shows a comparison of runtime of FANS with dif-ferent heuristics as we scale-up the number of agents. The x-axisshows the different sensor network domain configurations2 (seeFigure 3), and the y-axis plots the runtime in log-scale (again, inall our experiments, we imposed a 10000 second cutoff on run-time). Also, the time horizon for policy computation is 3 and weagain choose k = 1/2 for the Node heuristic. We can make thefollowing observations: (i) The Equality and Searcher heuristicsare not able to terminate within the cutoff time of 10000 s; (ii)The Node heuristic still remains the best for all configurations, butGreedy beats Link in the 15-3D domain. Thus FANS can scale upto as many as 15 agents, triple the number of agents SPIDER couldhandle.

The next set of experiments examine tradeoffs in solution qualityand provide analysis of these speedups. Table 1 shows a compar-ison of solution quality for SPIDER and FANS with select heuris-tics. The rows show all agent configurations considered so far whilethe columns show the algorithm used to obtain the solution quality.We can observe the following:(i) For the 4-chain and 5-star do-mains, FANS (even with its faster Greedy heuristic) is able to reachthe same quality as SPIDER, which is the global optimum. Forthe 5-P configuration, only an upper bound on the solution qualityis provided in [14], and FANS is able to obtain a solution qualitywithin 5% of the upper bound obtained by SPIDER. (ii) The Equal-ity heuristic obtains the highest solution quality in all cases; (iii)For domains of less than 10 agents, the Link heuristic matches thequality of the Equality heuristic (it loses less than 1% quality andless than 10% of quality in the 5-star and 7H domains respectively);(iv) As we scale up the Network, none of the three fastest heuristics(Link, Node, Greedy) can keep up to the solution quality providedby Equality or Searcher; (v) Even though the Equality and Searcher

2Configuration 15-mod is a modification to the 15-3D configura-tion with different target paths.

(b)

on the sensor network domain provided in Section 2. Our first ex-perimental result focuses on providing a head-to-head comparisonof FANS with different heuristics, against SPIDER, since SPIDERis the latest algorithm for ND-PODMPs, and also because it usesbranch-and-bound search using an MDP heuristic. The comparisonuses some of the sensor network configurations presented in [14],and an additional “7-H” configuration as shown in Figure 3. In thisexperiment, we limit the number of agents to less than 10, in orderto allow SPIDER to run.

4-Chain

5-Star 5-P

7-H 11-Helix15-3D

Figure 3: Sensor network configurations

Figure 4 shows a comparison of runtime of SPIDER vs. FANSwith different heuristics. The x-axis shows the different sensor net-work domain configurations, and the y-axis plots the runtime inlog-scale (in all our experiments, we imposed a 10000 second cut-off on runtime). Also, we chose a time horizon of 3 for policycomputation, to be consistent with the results shown in [14]. Forthe node heuristic, we chose k = 1/2; we discuss this choice of klater on. We can observe that (i) SPIDER is the slowest of all ofthe different algorithms considered and it hits the 10000 sec cut-off in the 5-P and 7-H cases; (ii) The Node and Link heuristicsare the two fastest in all cases; (iii) Unfortunately, Searcher’s extrasearch over FSMs appears have a significant runtime penalty; (iv)The Equality and Fairness heuristics are the two worst perform-ing heuristics; (v) Surprisingly, the Greedy heuristic matches theperformance of the Link heuristic for the 5-P and 7-H cases. Thekey conclusion from this graph is that FANS with either the Node,Link or Greedy heuristic is able to provide more than two orders ofmagnitude speedup over SPIDER.

Figure 4: Runtime comparison of SPIDER vs. FANS

Next, we show scale up in terms of time horizon for policy com-putation, and also provide a comparison of FANS against the LID-JESP algorithm, a locally optimal algorithm for ND-POMDPs [10].Runtime results for LID-JESP are available up to T=6, as shownbelow. Figure 5 shows a plot of the runtime as we increase the

time horizon. The x-axis shows the time horizon, and the y-axisplots the runtime in log-scale. Runtime values for different timehorizons are shown for LID-JESP and FANS with Node heuristic(k = 1/2), for the 4-chain and 5P configurations. The key obser-vations here are that FANS is able to provide orders of magnitudespeedup over LID-JESP for higher time horizons. For example, forthe 4-chain configuration with T=6, LID-JESP runs for 309 s, whileFANS terminates in 0.73 s, thus providing a 423-fold speedup overLID-JESP. Moreover, for the 5-P configuration, a speedup of morethan 1 order of magnitude is visible for T=5. We discuss solutionquality in detail later, but for both 4-chain and 5-P, the Node heuris-tic with k = 1/2 appears to match the quality of the global optimalsolution.

Figure 5: Runtime comparison of FANS with LID-JESP

Figure 6(a) shows a comparison of runtime of FANS with dif-ferent heuristics as we scale-up the number of agents. The x-axisshows the different sensor network domain configurations2 (seeFigure 3), and the y-axis plots the runtime in log-scale (again, inall our experiments, we imposed a 10000 second cutoff on run-time). Also, the time horizon for policy computation is 3 and weagain choose k = 1/2 for the Node heuristic. We can make thefollowing observations: (i) The Equality and Searcher heuristicsare not able to terminate within the cutoff time of 10000 s; (ii)The Node heuristic still remains the best for all configurations, butGreedy beats Link in the 15-3D domain. Thus FANS can scale upto as many as 15 agents, triple the number of agents SPIDER couldhandle.

The next set of experiments examine tradeoffs in solution qualityand provide analysis of these speedups. Table 1 shows a compar-ison of solution quality for SPIDER and FANS with select heuris-tics. The rows show all agent configurations considered so far whilethe columns show the algorithm used to obtain the solution quality.We can observe the following:(i) For the 4-chain and 5-star do-mains, FANS (even with its faster Greedy heuristic) is able to reachthe same quality as SPIDER, which is the global optimum. Forthe 5-P configuration, only an upper bound on the solution qualityis provided in [14], and FANS is able to obtain a solution qualitywithin 5% of the upper bound obtained by SPIDER. (ii) The Equal-ity heuristic obtains the highest solution quality in all cases; (iii)For domains of less than 10 agents, the Link heuristic matches thequality of the Equality heuristic (it loses less than 1% quality andless than 10% of quality in the 5-star and 7H domains respectively);(iv) As we scale up the Network, none of the three fastest heuristics(Link, Node, Greedy) can keep up to the solution quality providedby Equality or Searcher; (v) Even though the Equality and Searcher

2Configuration 15-mod is a modification to the 15-3D configura-tion with different target paths.

(c)

on the sensor network domain provided in Section 2. Our first ex-perimental result focuses on providing a head-to-head comparisonof FANS with different heuristics, against SPIDER, since SPIDERis the latest algorithm for ND-PODMPs, and also because it usesbranch-and-bound search using an MDP heuristic. The comparisonuses some of the sensor network configurations presented in [14],and an additional “7-H” configuration as shown in Figure 3. In thisexperiment, we limit the number of agents to less than 10, in orderto allow SPIDER to run.

4-Chain

5-Star 5-P

7-H 11-Helix15-3D

Figure 3: Sensor network configurations

Figure 4 shows a comparison of runtime of SPIDER vs. FANSwith different heuristics. The x-axis shows the different sensor net-work domain configurations, and the y-axis plots the runtime inlog-scale (in all our experiments, we imposed a 10000 second cut-off on runtime). Also, we chose a time horizon of 3 for policycomputation, to be consistent with the results shown in [14]. Forthe node heuristic, we chose k = 1/2; we discuss this choice of klater on. We can observe that (i) SPIDER is the slowest of all ofthe different algorithms considered and it hits the 10000 sec cut-off in the 5-P and 7-H cases; (ii) The Node and Link heuristicsare the two fastest in all cases; (iii) Unfortunately, Searcher’s extrasearch over FSMs appears have a significant runtime penalty; (iv)The Equality and Fairness heuristics are the two worst perform-ing heuristics; (v) Surprisingly, the Greedy heuristic matches theperformance of the Link heuristic for the 5-P and 7-H cases. Thekey conclusion from this graph is that FANS with either the Node,Link or Greedy heuristic is able to provide more than two orders ofmagnitude speedup over SPIDER.

Figure 4: Runtime comparison of SPIDER vs. FANS

Next, we show scale up in terms of time horizon for policy com-putation, and also provide a comparison of FANS against the LID-JESP algorithm, a locally optimal algorithm for ND-POMDPs [10].Runtime results for LID-JESP are available up to T=6, as shownbelow. Figure 5 shows a plot of the runtime as we increase the

time horizon. The x-axis shows the time horizon, and the y-axisplots the runtime in log-scale. Runtime values for different timehorizons are shown for LID-JESP and FANS with Node heuristic(k = 1/2), for the 4-chain and 5P configurations. The key obser-vations here are that FANS is able to provide orders of magnitudespeedup over LID-JESP for higher time horizons. For example, forthe 4-chain configuration with T=6, LID-JESP runs for 309 s, whileFANS terminates in 0.73 s, thus providing a 423-fold speedup overLID-JESP. Moreover, for the 5-P configuration, a speedup of morethan 1 order of magnitude is visible for T=5. We discuss solutionquality in detail later, but for both 4-chain and 5-P, the Node heuris-tic with k = 1/2 appears to match the quality of the global optimalsolution.

Figure 5: Runtime comparison of FANS with LID-JESP

Figure 6(a) shows a comparison of runtime of FANS with dif-ferent heuristics as we scale-up the number of agents. The x-axisshows the different sensor network domain configurations2 (seeFigure 3), and the y-axis plots the runtime in log-scale (again, inall our experiments, we imposed a 10000 second cutoff on run-time). Also, the time horizon for policy computation is 3 and weagain choose k = 1/2 for the Node heuristic. We can make thefollowing observations: (i) The Equality and Searcher heuristicsare not able to terminate within the cutoff time of 10000 s; (ii)The Node heuristic still remains the best for all configurations, butGreedy beats Link in the 15-3D domain. Thus FANS can scale upto as many as 15 agents, triple the number of agents SPIDER couldhandle.

The next set of experiments examine tradeoffs in solution qualityand provide analysis of these speedups. Table 1 shows a compar-ison of solution quality for SPIDER and FANS with select heuris-tics. The rows show all agent configurations considered so far whilethe columns show the algorithm used to obtain the solution quality.We can observe the following:(i) For the 4-chain and 5-star do-mains, FANS (even with its faster Greedy heuristic) is able to reachthe same quality as SPIDER, which is the global optimum. Forthe 5-P configuration, only an upper bound on the solution qualityis provided in [14], and FANS is able to obtain a solution qualitywithin 5% of the upper bound obtained by SPIDER. (ii) The Equal-ity heuristic obtains the highest solution quality in all cases; (iii)For domains of less than 10 agents, the Link heuristic matches thequality of the Equality heuristic (it loses less than 1% quality andless than 10% of quality in the 5-star and 7H domains respectively);(iv) As we scale up the Network, none of the three fastest heuristics(Link, Node, Greedy) can keep up to the solution quality providedby Equality or Searcher; (v) Even though the Equality and Searcher

2Configuration 15-mod is a modification to the 15-3D configura-tion with different target paths.

(d)

Figure 3: Sensor network configurations

O(n · (QLB · T + QUB) + (QUB)d) which is linear in thehorizon T .

As shown in Section 3.1.2, the complexity of calculatingthe heuristic also increases linearly with T . Once the heuris-tic is computed, the complexity depends upon the main dy-namic programing loop (steps 5-8) of Algorithm 1. If the sizeof the sampled belief set is fixed then the running time ofthe loop is independent of T . Hence, CBDP requires lineartime complexity w.r.t. the horizon length.

The next two theorems show how well the CBDP and theheuristic computation scale when we change the problemparameters–state space, action space or number of agents.

Theorem 3. For each horizon, CBDP has polynomial timeand space complexity w.r.t. the factored state space, exponen-tial in the induced width d of the interaction graph and linearin the number of agents.

Proof. The time complexity of CBDP per horizon de-pends upon the best joint policy computation. Its complex-ity is O(maxBelief · ((QUB)kL|Su|2|Si|2k|Ωj |k + n(QUB)d)).L denotes the number of links in the interaction graph, and kdenotes the maximum number of agents involved in an inter-action link and n being the number of agents. The first termdenotes the time required to setup the constraint formula-tion, second term is the time needed to solve it. Similarly,the space complexity is dominated by the bucket eliminationalgorithm, which is O(n(QUB)d), again linear in the numberof agents and exponential only in the induced width.

A direct implication of the above theorem is that CBDPcan scale up well to very large problem with low inducedwidth. For the sensor network domain, we can further showthe following property. A discussion of how to generalize itis included in the last section.

Theorem 4. The MDP heuristic requires polynomial timeand space per horizon for the sensor network domain.

Proof. This follows directly from Section 3.1.2. Thetime complexity is O(n|Su|2||Ai|d) and space complexityO(|Su|+ n|Ai|d).

150 200 250 300 350 400 450 500 550 600

3 4 5 6 7 8 9 10

Polic

y Va

lue

Horizon

CBDP Node Link

(a)

1

10

100

1000

3 4 5 6 7 8 9 10

Tim

e (s

ec)

Horizon

CBDP Node Link

(b)

Figure 4: a) compares the solution quality of CBDPand different versions of FANS (Node, Link) on thedomain 5P. b) Compares the execution time.

5. EXPERIMENTSThis section compares CBDP with FANS [10]–the most

efficient of the existing algorithms including SPIDER [18]and LID-JESP [11]. We conducted experiments on the sen-sor network configurations introduced in [10], shown in Fig-ure 3. All experiments were conducted on a Dual core ma-chine with 1GB RAM, 2.4GHz CPU. The main purpose ofthe experiments is to show the scalability of CBDP, whichhas linear time and space complexity w.r.t. the horizon. Weuse a publicly available constraint solver implementing thebucket elimination algorithm [12]. The maxbelief was setto 5 as it provided good tradeoff between speed and solutionquality across a range of tried settings.

The experiments are divided into two sets. In the firstset, we use relatively smaller domains (5-P and 7-H) becausethey are the only domains for which FANS can scale up tosignificant horizons. In the second set of experiments, we usethe remaining larger domains (11-helix, 15-3D, 15-mod). Forthese domains, FANS cannot scale well (for 15-3D it fails tosolve problems with horizons greater than 5).

Figure 4(a) shows a comparison of the solution quality ofCBDP and two different versions of FANS, Node (k = 0.5)and Link, on the 5-P domain. FANS has multiple heuristicswhich tradeoff solution quality for speed. It has been sug-gested in [10] that for smaller domains, the Link and Nodeheuristics provide the best tradeoff. Hence, we chose thesetwo for comparisons. All three algorithms achieve almostthe same solution quality for all horizons. However, the keydifference is in the runtime (Figure 4(b)). CBDP providessignificant speedup over both the Node and Link heuristics.

Page 7: Constraint-Based Dynamic Programming for Decentralized ...rbr.cs.umass.edu/shlomo/papers/KZaamas09.pdfof interactions using constraint networks. These results ex-tend significantly

250

300

350

400

450

500

550

11-helix 15-3D 15-Mod

Polic

y va

lue

CBDPGreedy

Node

(a)

1

10

100

1000

10000

11-helix 15-3D 15-Mod

Tim

e (s

ecs,

logs

cale

)

CBDPGreedy

Node

(b)

Figure 6: Solution quality and execution time for 11-helix, 15-3D and 15-mod with T = 3.

150 200 250 300 350 400 450 500 550 600 650 700

2 3 4 5 6 7 8 9 10

Polic

y Va

lue

Horizon

CBDP Link Node

(a)

0.1

1

10

100

1000

2 3 4 5 6 7 8 9 10

Tim

e (s

ec)

Horizon

CBDP Link Node

(b)

Figure 5: a) compares the solution quality of CBDPand different versions of FANS (Node, Link) on thedomain 7-H. b) Compares the execution time.

For horizon 10, CBDP is about 2 orders of magnitude fasterthan FANS with either heuristic. Another notable obser-vation is that the increase in runtime for CBDP with thehorizon is nearly linear as implied by Theorem 2.

Figure 5(a) shows a comparison of the solution quality onthe 7-H domain. We again compare with FANS using theNode (k = 0.5) and Link heuristics. On this domain, dueto increase in the number of agents, FANS can provide so-lutions up to horizon 7. With the increase in the domainsize, FANS is no longer as competitive in the solution qual-ity as it was in the 5-P domain. CBDP provides much bettersolution quality for all horizons. The runtime graph (Fig-ure 5(b)) shows again that CBDP is much faster and thatmore scalable. It is 5 times faster than the Link heuristic,

and orders of magnitude faster than the Node heuristic. Theaverage time required by CBDP per horizon is 710ms whichimplies CBDP can solve a 100 horizon instance within 1.5minutes due to its linear complexity as opposed to the Nodeheuristic which requires 685 seconds to solve a horizon 7instance. This further illustrates the scalability of CBDP.

Figure 6 shows the next set of experiments in which weuse larger domains (11-helix, 15-3D, 15-mod). To be consis-tent with the results of Marecki et al. [10], we set the horizonto 3 because for 15-3D, FANS scales up to horizon 4 only.For these larger domains, a Greedy heuristic has been pro-posed [10] and it provides best tradeoff than other heuristics.Hence, we used this heuristic along with the Node (k = 0.75)heuristic for comparisons. Figure 6(a) shows a comparison ofthe solution quality. For 11-helix, CBDP provides much bet-ter solution quality than either the Greedy or Node heuris-tic. Similar trends are observed for the 15-3D and 15-moddomains. CBDP provides better solution quality for themas well. The runtime comparison (Figure 6(b)) again showsthe stark difference between CBDP and FANS. For 15-mod,CBDP takes 0.6 secs while FANS with the Greedy heuristictakes over 2,000 secs. One can easily see the magnitude ofspeedup CBDP provides over FANS.

1500

2500

3500

4500

5500

6500

7500

8500

9500

0 10 20 30 40 50 60 70 80 90 100

210 410 610 810 1010 1210 1410 1610 1810 2010 2210 2410 2610 2810 3010

Polic

y Va

lue

Exec

utio

n Ti

me

(sec

s)

Horizon

Solution qualityTime

Figure 7: Solution Quality and Execution time for arange of horizons for 15-3D.

The results of the last set of experiments are shown inFigure 7. These experiments show the solution quality andexecution time statistics for CBDP for different horizons onthe problem 15-3D, which is the most complex problem in-stance for both CBDP and FANS. FANS could scale up toonly horizon 5 on this problem. In contrast, CBDP can eas-ily scale up to horizon 100. The execution time curve in this

Page 8: Constraint-Based Dynamic Programming for Decentralized ...rbr.cs.umass.edu/shlomo/papers/KZaamas09.pdfof interactions using constraint networks. These results ex-tend significantly

experiment clearly supports the linear time complexity ofCBDP. The solution quality too increases at a nearly con-stant rate with the horizon, which means that CBDP is ableto maintain good solution quality for large horizons.

To summarize, our experiments consistently illustrate thatCBDP provides several orders-of-magnitude of speedup overdifferent versions of FANS, while providing better solutionquality for all the instances examined. CBDP further has alinear time and space complexity w.r.t. the horizon, whichmakes it possible to apply it to problems with much largerhorizons than was previously possible.

6. CONCLUSION AND FUTURE WORKWe have introduced a point-based dynamic programing

algorithm called CBDP for ND-POMDPs–a restricted classof DEC-POMDPs characterized by a decomposable rewardfunction. CBDP takes advantage of the structure of inter-action in the domain and independence among the agentsto produce significantly better scalability over the state-of-the-art algorithms. It uses constraint networks algorithmsto improve the efficiency of key steps of the dynamic pro-gramming algorithm. Consequently, CBDP has linear timeand space complexity with respect to the horizon length andpolynomial complexity in other problem parameters. Thekey feature of CBDP which contributes to its scalability isthat the complexity per horizon step is exponential only inthe induced width of the agent interaction graph, while itis linear in the number of agents. The induced width is of-ten much smaller than the number of agents–particularly inloosely connected multi-agent systems–implying that CBDPcan scale up better and handle multi-agent systems that aresignificantly larger than the existing benchmark problems.Experimental results consistently validate these propertiesof CBDP.

Continued work on CBDP focuses on two research direc-tions. First, we are exploring additional real-world settingsthat have an underlying interaction structure that CBDPcan exploit. Sensor networks in particular provide a goodapplication area where the challenge is to coordinate the op-eration of a large team of sensors. The second research di-rection aims to make the computation of the factored MDPheuristic efficient in more complex domains. For the tar-get tracking problem, the heuristic computation is alreadyefficient. But for problems that require agents to maintainsubstantial internal states, an exponential effort is required.To avoid this blowup in computation time, we plan to usethe approach developed by Guestrin et al. [5], which haspolynomial time complexity in the state space for factoredMDPs such as the ones used in our work. Such an approachwill further increase the scalability and applicability of DEC-POMDPs to more complex real-world problems.

AcknowledgmentsWe thank Janusz Marecki for helping us perform the com-parisons with FANS. Support for this work was provided inpart by the National Science Foundation under grant IIS-0812149 and by the Air Force Office of Scientific Researchunder grant FA9550-08-1-0181.

7. REFERENCES[1] R. Becker. Exploiting Structure in Decentralized Markov

Decision Processes. PhD thesis, University ofMassachusetts Amherst, 2006.

[2] D. S. Bernstein, R. Givan, N. Immerman, andS. Zilberstein. The complexity of decentralized control ofMarkov decision processes. Mathematics of OperationsResearch, 27:819–840, 2002.

[3] A. Carlin and S. Zilberstein. Value-based observationcompression for DEC-POMDPs. In Proc. of the SeventhInternational Joint Conference on Autonomous Agents andMultiagent Systems, pages 501–508, 2008.

[4] R. Dechter. Bucket elimination: A unifying framework forreasoning. Artificial Intelligence, 113(1-2):41–85, 1999.

[5] C. Guestrin, D. Koller, and R. Parr. Multiagent planningwith factored MDPs. In Proc. of the Neural InformationProcessing Systems, pages 1523–1530, 2001.

[6] E. A. Hansen, D. S. Bernstein, and S. Zilberstein. Dynamicprogramming for partially observable stochastic games. InProc. of the Nineteenth National Conference on ArtificialIntelligence, pages 709–715, 2004.

[7] D. Koller and R. Parr. Computing factored value functionsfor policies in structured MDPs. In Proc. of the SixteenthInternational Joint Conference on Artificial Intelligence,pages 1332–1339. Morgan Kaufmann, 1999.

[8] V. Lesser, M. Tambe, and C. L. Ortiz, editors. DistributedSensor Networks: A Multiagent Perspective. KluwerAcademic Publishers, Norwell, MA, USA, 2003.

[9] W. S. Lovejoy. Computationally feasible bounds forpartially observed Markov decision processes. OperationsResearch, 39(1):162–175, 1991.

[10] J. Marecki, T. Gupta, P. Varakantham, M. Tambe, andM. Yokoo. Not all agents are equal: Scaling up distributedPOMDPs for agent networks. In Proc. of the SeventhInternational Joint Conference on Autonomous Agents andMultiagent Systems, pages 485–492, 2008.

[11] R. Nair, P. Varakantham, M. Tambe, and M. Yokoo.Networked distributed POMDPs: A synthesis of distributedconstraint optimization and POMDPs. In Proc. of theTwentieth National Conference on Artificial Intelligence,pages 133–139, 2005.

[12] A. Petcu and B. Faltings. Dpop: A scalable method formultiagent constraint optimization. In Proc. of theInternational Joint Conference on Artificial Intelligence,pages 266–271, 2005.

[13] J. Pineau, G. Gordon, and S. Thrun. Point-based valueiteration: An anytime algorithm for POMDPs. In Proc. ofthe Eighteenth International Joint Conference on ArtificialIntelligence, pages 1025–1032, 2003.

[14] S. Seuken and S. Zilberstein. Memory-bounded dynamicprogramming for DEC-POMDPs. In Proc. of the TwentiethInternational Joint Conference on Artificial Intelligences,pages 2009–2015, 2007.

[15] G. Shani, R. I. Brafman, and S. E. Shimony. Forwardsearch value iteration for POMDPs. In Proc. of theTwentieth International Joint Conference on ArtificialIntelligence, pages 2619–2624, 2007.

[16] M. Spaan and N. Vlassis. Perseus: Randomized point-basedvalue iteration for POMDPs. Journal of ArtificialIntelligence Research, 24:195–220, 2005.

[17] D. Szer and F. Charpillet. MAA*: a heuristic searchalgorithm for solving decentralized POMDPs. In Proc. ofthe Twenty-First Conference on Uncertainty in ArtificialIntelligence, pages 576–583, 2005.

[18] P. Varakantham, J. Marecki, Y. Yabu, M. Tambe, andM. Yokoo. Letting loose a SPIDER on a network ofPOMDPs: Generating quality guaranteed policies. In Proc.of the Sixth International Joint Conference on AutonomousAgents and Multiagent Systems, pages 1–8, 2007.