Open Loop Execution of Tree-Search Algorithms · 2018. 10. 23. · Open Loop Execution of...

8
an author's https://oatao.univ-toulouse.fr/19924 https://doi.org/10.24963/ijcai.2018/327 Lecarpentier, Erwan and Infantes, Guillaume and Lesire, Charles and Rachelson, Emmanuel Open Loop Execution of Tree-Search Algorithms. (2018) In: 2018 International Joint Conference on Artificial Intelligence (IJCAI 2018), 13 July 2018 - 19 July 2018 (Stockholm, Sweden).

Transcript of Open Loop Execution of Tree-Search Algorithms · 2018. 10. 23. · Open Loop Execution of...

Page 1: Open Loop Execution of Tree-Search Algorithms · 2018. 10. 23. · Open Loop Execution of Tree-Search Algorithms Erwan Lecarpentier 1;2, Guillaume Infantes 1, Charles Lesire 1, Emmanuel

an author's https://oatao.univ-toulouse.fr/19924

https://doi.org/10.24963/ijcai.2018/327

Lecarpentier, Erwan and Infantes, Guillaume and Lesire, Charles and Rachelson, Emmanuel Open Loop Execution of

Tree-Search Algorithms. (2018) In: 2018 International Joint Conference on Artificial Intelligence (IJCAI 2018), 13 July

2018 - 19 July 2018 (Stockholm, Sweden).

Page 2: Open Loop Execution of Tree-Search Algorithms · 2018. 10. 23. · Open Loop Execution of Tree-Search Algorithms Erwan Lecarpentier 1;2, Guillaume Infantes 1, Charles Lesire 1, Emmanuel

Open Loop Execution of Tree-Search Algorithms

Erwan Lecarpentier1,2, Guillaume Infantes1, Charles Lesire1, Emmanuel Rachelson2

1 ONERA - The French Aerospace Lab, Toulouse, France2 ISAE - SUPAERO, University of Toulouse, France

[email protected], [email protected],[email protected], [email protected]

AbstractIn the context of tree-search stochastic planning al-gorithms where a generative model is available, weconsider on-line planning algorithms building treesin order to recommend an action. We investigatethe question of avoiding re-planning in subsequentdecision steps by directly using sub-trees as ac-tion recommender. Firstly, we propose a methodfor open loop control via a new algorithm takingthe decision of re-planning or not at each time stepbased on an analysis of the statistics of the sub-tree.Secondly, we show that the probability of selectinga suboptimal action at any depth of the tree can beupper bounded and converges towards zero. More-over, this upper bound decays in a logarithmic waybetween subsequent depths. This leads to a dis-tinction between node-wise optimality and state-wise optimality. Finally, we empirically demon-strate that our method achieves a compromise be-tween loss of performance and computational gain.

1 IntroductionTree-search based algorithms recently encountered a real suc-cess at solving sequential, highly combinatorial problemssuch as the challenging game of Go [Enzenberger et al., 2010;Silver et al., 2016]. Such algorithms use a generative modelof the environment to simulate episodes starting from the cur-rent state of the agent [Sutton, 1991; Sutton and Barto, 1998].This allows the exploration of reachable states and actionsand results in the construction of an (unbalanced) scenariotree, that aims at identifying promising branches with a lim-ited computational budget. When the computational budgetis exhausted, the recommended action at the root node is ap-plied and a new tree is built in the resulting state. This resultsoverall in a closed loop control process.

We are interested in stochastic problems with large statespaces (e.g. continuous) with a short decision time (bud-get). In this setting, open loop planning algorithms haveproven to be successful [Bubeck and Munos, 2010] and evento outperform [Weinstein and Littman, 2012] the standard ap-proaches that consider closed loop policy trees such as UCT[Kocsis and Szepesvari, 2006]. They seek for optimal se-quences of actions (plans) rather than optimal policies de-

spite the sub-optimal nature of a plan in stochastic environ-ments. Indeed, computing the latter prevents feed-back onthe explored states but allows to break the complexity of thestate space exploration. Given a tree computed by an openloop planning algorithm, we propose to keep the sub-treereached by the application of the recommended action andto directly use it as the main tree for the subsequent time step,without re-planning. What motivates this approach is spar-ing the computational cost of tree building for subsequenttime steps, hence reducing the number of calls to the sim-ulator. The interest of this can be seen in two ways. Onone hand it is a way of reducing energy consumption for sys-tems with low computational resources [Wilson et al., 2014;2016]. On the other hand, the saved computational time canbe re-invested into other tasks. Particularly, this approachis adapted for low level control (i.e. high frequency) wheresub-sequent tree developments is cumbersome. In this frame-work, Perez et al. [2012a] and Heusner [2011] consideredkeeping the tree in deterministic environments but observeda negative impact as the sub-trees were systematically keptwithout analysis. Moreover, they lose the aforementionedcomputational gain by refining the sub-trees.

In this paper, we study the impact of using the subsequentsub-trees as main trees for the next action steps without fur-ther re-planning. We claim that in lowly-stochastic environ-ments, the reached performance is comparable to algorithmssystematically discarding the tree. Our contribution is three-fold. (1) We introduce a new algorithm called OLTA (Sec-tion 3), performing a systematic analysis of the sub-tree andtaking the decision of re-planning or not at each time step.(2) We upper bound the probability of selecting a subopti-mal action within a sub-tree, the sense of optimality beingdefined in an open loop fashion (Section 4). Additionally, weshow that this upper bound decays logarithmically with thesub-tree depth. (3) We show in our experiments the benefitof applying such a method both in terms of performance andcomputational cost saving (Section 5).

2 Background2.1 Markov Decision ProcessWe model the planning problem as a Markov Decision Pro-cess (MDP) where an agent sequentially takes actions withthe general goal of maximizing the cumulative return fed

Page 3: Open Loop Execution of Tree-Search Algorithms · 2018. 10. 23. · Open Loop Execution of Tree-Search Algorithms Erwan Lecarpentier 1;2, Guillaume Infantes 1, Charles Lesire 1, Emmanuel

back by the environment [Puterman, 2014]. We refer to thestate space as S and the action space as A. We suppose thenumber of actions to be finite with K = |A|, thus we writeA = aiKi=1. We also consider that the available actions areindependent of the state the agent lies in. The state transitionfunction is stochastic and we note P (s′|s, a) the probabilityof reaching state s′ after taking action a in state s. The re-ward model is denoted by r(s, a, s′) and refers to the scalarreward received while performing the transition (s, a, s′). Weassume that this reward function is deterministic. Finally,we suppose the horizon of the MDP is infinite and we noteγ ∈ [0, 1) the discount factor which represents the impor-tance of the subsequent collected rewards.

2.2 Tree Representation

When a generative model of the MDP is available, it becomespossible to use it within planning algorithms. Tree-search al-gorithms use this model in order to build a tree of what maypossibly occur in the current situation of the agent [Sutton,1991; Sutton and Barto, 1998; Silver et al., 2008]. In thestochastic setting with potentially infinitely many states, weuse a tree structure similar to the one used by Bubeck andMunos [2010]. The tree built at each time step consists in alook-ahead search of the possible outcomes while followingsome action plan starting from the current state of the agents0 ∈ S. Thus, the root node of the tree is labelled by theunique state s0. The edges correspond to the K available ac-tions, K being the branching factor of the tree. The tree itselfconforms to an ensemble of action sequences, or plans, orig-inating from its root node.

We emphasize the fact that this tree structure implies thatwe search for a state-independent optimal sequence of ac-tions (open loop plan) which is in general sub-optimal com-pared to a state-dependent policy search. The THTS familyof algorithms in particular [Keller and Helmert, 2013] definestrees with chance and decision nodes while our structure doesnot apply an equality operator on the sampled states. Fol-lowing Bubeck and Munos; Weinstein and Littman [2010;2012], we argue that closed-loop application of the first ac-tion in optimal open loop plans, although theoretically sub-optimal, can be competitive with these methods in practice,while being more sample-efficient.

Since the transition model is stochastic, the non-root nodesare not labelled by a unique state. Instead, every such node isassociated to a state distribution resulting from the applicationof the action plan leading to the considered node and startingfrom s0. During the exploration, we consider saving all sam-pled states at each non-root node. A comprehensive illustra-tion of such a tree can be found in Figure 1. This approach ex-tends straightforwardly to Partially Observable Markov Deci-sion Processes (POMDP) [Silver and Veness, 2010].

Given a tree-search, open loop planning algorithm, we callTd the tree at depth d ∈ N, that is the sub-tree resulting fromthe application of the d first recommended actions. HenceT0 denotes the whole tree, T1 the tree starting from the nodereached by the application of the first recommended actionand so on.

Root node

Non-root node

s0

a1 ai

s′1 s′l

aK

d = 0

d = 1∀i ∈ 1, · · · , l ,s′i ∼ P (·|s0, ai)

a1 aK

· · ·

· · · · · ·

· · ·

Figure 1: General representation of a tree, where l ∈ N is the num-ber of times the sub-tree reached by action ai has been developed.Two nodes are represented in this tree with their respective depthson the left.

2.3 Open Loop UCTFor the sake of clarity and in order to clearly separate the treebuilding properties from the open loop execution presented inthe next section, we define an open loop planning algorithmutilizing the presented tree structure that we call Open LoopUCT (OLUCT). The difference between UCT and OLUCT isthat OLUCT is not provided with an equality operator overstates. Within the THTS terminology, this means that deci-sion and chance nodes do not correspond to a single state butto the state distribution reachable by the action plan leadingto the node. Hence decision and chance nodes are associ-ated to the state distribution which makes OLUCT an openloop planning algorithm. The fundamental consequence isthat an action value within our tree is computed w.r.t. theparent node’s state distribution rather than a single state.

Apart from this, OLUCT uses the same exploration proce-dure as UCT. Within a node, we note Xi,u the estimated ex-pected return of action i after u samples of this action. Ti(t) isthe number of trials of action i up to time t of the OLUCT pro-cedure. An Upper Confidence Bound (UCB) strategy [Aueret al., 2002] is applied at each node where each action is seenas an arm of a bandit problem. The tree policy selects theaction It with the highest UCB:

It = arg maxi∈1,··· ,K

Xi,Ti(t−1) + ct−1,Ti(t−1)

,

where ct,u = 2Cp

√ln(t)u is an exploration term ensuring that

all actions will be sampled infinitely often.The Cp parame-ter drives the exploration-exploitation trade-off. The OLUCTtree building procedure is detailed in Algorithm 1.

3 OLTA (Open Loop Tree-search Algorithm)3.1 DescriptionIn order to control the execution of open loop plans, we pro-pose a new algorithm called OLTA (Algorithm 2). It relieson a generic open loop planning algorithm to generate a tree,rooting from the current state. For the next execution timestep, it decides either to use the sub-tree reached by the rec-ommended action or to trigger a re-planning by building anew tree. If no re-planning is triggered, then the recom-mended action of the sub-tree is applied without using the

Page 4: Open Loop Execution of Tree-Search Algorithms · 2018. 10. 23. · Open Loop Execution of Tree-Search Algorithms Erwan Lecarpentier 1;2, Guillaume Infantes 1, Charles Lesire 1, Emmanuel

Algorithm 1: OLUCT tree building procedureFunction createTree(state s):Parameters: budget n; default policy πdefault.Create root node νroot(s)for t ∈ 1, · · · , n do

νleaf = Select(νroot); // Select a leaf nodew.r.t. the UCT strategy and sample anew state for each encountered node.

Expand(νleaf ); // Expand the node if notterminal using the generative model.

∆ = Evaluate(νleaf , πdefault); // Simulate aroll-out using πdefault, starting fromthe last sampled state in νleaf.

Backup(νleaf ,∆); // Back-propagate thesampled return.

return T (νroot)

additional information of the new state observed after thetransition. This results in an open loop control process andspares the cost of developing a new tree starting at this state.The intuition behind OLTA is that several consecutive rec-ommended actions in an optimal branch of the tree can bereliable, despite the randomness of the environment. A majorexample of such a case is low-level control, where consecu-tive sampled states are close to each other.

In this paper, for its performance and simplicity, we choseto implement OLUCT as the open loop planning algorithmutilized by OLTA. However, any other algorithm generat-ing trees as described in Section 2.2 could be used in thesame way (e.g. OLOP [Bubeck and Munos, 2010], orHOLOP [Weinstein and Littman, 2012]).

One important feature of OLTA is the so-called “decision-Criterion”, based on which the agent decides to either usethe first sub-tree following the recommended action, or to re-build a new tree from the current state. The decision is basedon a comparison with the characteristics of the resulting sub-tree and the current state of the agent. In the next section,we discuss different decision criteria, leading to the consider-ation of a family of different algorithms.

3.2 Decision CriterionThe simplest implementation of the decision criterion is tokeep the sub-tree only if its root node is fully expanded. Thismeans that each action has been sampled at least once. Wecall the resulting algorithm Plain OLTA. It naively trusts thevalue estimates of the sub-tree, thus applies the whole plan ofrecommended actions at each depth until it reaches a partiallyexpanded node. Therefore, Plain OLTA is expected to per-form better in deterministic environments. In stochastic caseshowever, those estimates may be biased because of the differ-ent sources of uncertainty within the MDP (reward function,state transition function and action selection). For this reason,we seek more robust criteria to base the decision on.

A natural way to decide whether to keep the sub-tree ornot is to track if the recommended action is optimal w.r.t. thenew state s of the agent. Here we make an important dis-tinction between a state-wise optimal action and a node-

Algorithm 2: OLTA algorithmFunction OLTA:Parameters: initial state s0; tree building procedurecreateTree; re-planning criterion decisionCriterion.s = s0;T = createTree(s);while s is not terminal do

if decisionCriterion(s, T ) thena = recommendedAction(T ); // Get thefirst recommended action.

elseT = createTree(s); // Create a new treefrom the current state.a = recommendedAction(T ); // Get thefirst recommended action.

T = subTree(T , a); // Move the tree to thefirst sub-tree resulting from theapplication of a.

s := realWorldTransitionFunction(s, a);

wise optimal action. The first one is the action recom-mended by the optimal policy in a specific state. We noteit a∗ = arg maxa∈AQ

∗(s, a), with Q∗ : S × A → R theoptimal state-action value function. In order to define thesecond one, we introduce Sd, the state random variable atthe root node of Td. Its distribution results from the applica-tion of the d first recommended actions starting from s0, soSd ∼ P (·|s0, a0, · · · , ad−1) ≡ PSd(·). The node-wise opti-mal action maximizes the expected return given the state dis-tribution of the node. We note it a∗d = arg maxa∈AQ

∗d(a)

where Q∗d : A → R is the optimal action value functionw.r.t. the state distribution at the root node of Td, that isQ∗d(a) = Es∼PSd (Q∗(s, a)). Following Bellemare et al.[2017], a distributional Bellman equation can be expressed interms of three sources of randomness that are: R : S×A→ Rthe stochastic reward function; X : S × A → R the randomreturn; and Pπ the transition operator with PπX(s, a)

D:=

X(S′, A′), S′ ∼ P (·|s, a) and A′ ∼ π(·|S′). Mathemati-cally, we have the following distributional Bellman equations:

Qπ(s, a) = Eπ(X(s, a))

Qπd (a) = EPSd (Qπ(s, a)),

with X(s, a) ∼ R(s, a) + γPπX(s, a) and Sd ∼ PSd(·).Unfortunately, at the root node of Td for d > 0, open looptree-search algorithms do not estimate Q∗ but Q∗d. The biasintroduced by the state distribution implies that in the generalcase we have no guarantee that a∗ = a∗d. The risk is that theset ΩSd of possible realizations of Sd can include states wherea∗ is sub-optimal, in which case the resulting return evalua-tions would weight in favour of a different action than a∗. Inother words — introducing the notion of domination domainfor an action a as Da = s ∈ S|π∗(s) = a ⊂ S — if ΩSd isnot included in Da∗ , then the risk of the recommended actionto be state-wise sub-optimal is increased. Conversely, if ΩSdis included in the domination domain of a∗, then the optimal

Page 5: Open Loop Execution of Tree-Search Algorithms · 2018. 10. 23. · Open Loop Execution of Tree-Search Algorithms Erwan Lecarpentier 1;2, Guillaume Infantes 1, Charles Lesire 1, Emmanuel

action will be selected given that the budget is “big enough”w.r.t. the chosen tree-search algorithm’s performance. Conse-quently, one should base the decision criterion on the analysisof PSd and the action domination domains. To compute thesedomains, Rachelson and Lagoudakis [2010] use the proper-ties of Lipschitz-MDPs. Although the following discussion isinspired by this work, the consideration of Lipschitz-MDPs isout of the scope of this paper. We discuss below the construc-tion of decision criteria that will be illustrated in Section 5.

Current state analysis & POMDP setting. The currentstate s of the agent can be compared to the empirical statedistribution PSd at the root node of the sub-tree. If PSd(s) islarge, then the value estimators are related to the locality ofthe state space the agent lies in. If not, then the node-wiseoptimal action may not be state-wise optimal. This consider-ation supposes to identify a state-metric for which two closestates have a high chance to be in the same action domina-tion domain. Alternatively, in the case of a POMDP, a beliefdistribution on the current state is available instead of the cur-rent state itself [Kaelbling et al., 1998]. In such a case, adirect comparison between this distribution and PSd can beperformed (e.g. with a Wasserstein metric). Note that mak-ing use of the current state of the agent makes the algorithmclosed-loop, by definition. We use the terminology ”open-loop” in order to distinguish OLTA from classical closed-loop Tree Search algorithms that systematically re-plan, root-ing from the current state (e.g. OLOP [Bubeck and Munos,2010], performs closed-loop execution).

State distribution analysis. The dispersion and multi-modality of PSd could motivate not to re-use a sub-tree. Ahigh dispersion involves the possibility that ΩSd does not be-long to a single action domination domain and a re-planningshould be triggered. The same consideration applies in termsof multi-modality. Conversely, a narrow, mono-modal, statedistribution is a good hint for ΩSd to be comprised into a sin-gle action domination domain.

Return distribution analysis. A widespread or a multi-modal return distribution for the recommended action in anode may indicate a strong dependency on the region of thestate space we lie in. If ΩSd covers different action domi-nation domains, each of these domains may contribute a dif-ferent return distribution to the node’s return estimates, thusinducing a high variance on this distribution or even a multi-modality. In this case, it could be beneficial to trigger the re-planning. Alternatively, even after re-planning, widespread ormulti-modal return distributions can naturally arise as a resultof the MDP’s reward and transition models.

We do not provide a unique generic method to base the de-cision criterion on. Indeed, we believe that it is a stronglyproblem-dependent issue and that efficient heuristics can bebuilt accordingly. However, the analysis of the state and re-turn distributions constitute promising indicators and we ex-emplify their use in the experiments of the last section.

4 Theoretical AnalysisIn this section, we demonstrate that the algorithm asymptoti-cally provides node-wise optimal actions for any sub-tree Tdof depth d. We first derive an upper bound on the failure

s0

r = 1

s1

r = 0

s2

r = 0

s3

r = 0

s4

r = 1

Figure 2: 1D track environment. On top of each cell representing astate is the immediate reward of the transition to this state.

probability that converges towards zero when the initial bud-get n of the algorithm goes to infinity. Then, we character-ize the loss of performance guarantees between subsequentdepths and show a logarithmic decay of the upper bound.The demonstration unfolds as follows: first we write a lowerbound for the number of trials of the actions at the root ofTd in Lemma 1; then we write an upper bound on the failureprobability given a known budget at depth d in Lemma 2; fi-nally we derive a recursive relation between the upper boundsof subsequent trees that leads to our result in Theorem 1.Proofs are omitted due to lack of space1.

We note b(d) ∈ N the budget used to develop Td i.e. thenumber of times the d first recommended actions have beenselected by the tree policy. We note T di,t the number of timesthe ith action at the root node of Td has been selected bythe OLUCT tree policy after t expansions of Td. Similarly,Xd

i,Tdi,t≡ X

d

i,t denotes the estimate of the return of the ith

action at depth d after t expansions of the sub-tree Td. Wewrite Idt the index of the action chosen by the tree policy atdepth d after t expansions of Td. We have:

Idt = arg maxi∈1,··· ,K

Xd

i,t−1 + ct−1,Tdi,t−1

.

The recommended action at depth d given a budget b(d)

is Id = arg maxi∈1,··· ,KXd

i,b(d). Following Kocsis andSzepesvari [2006], we assume that the empirical estimatesXd

i,t converge and write Xdi,t = EXd

i,t and Xdi =

limt→∞Xdi,t. Then, we define for i ∈ 1, · · · ,K \ i∗d,

∆di = Xd

i∗d− Xd

i where we note i∗d the index of the node-wise optimal action at the root node of Td. We make the as-sumption that only one action is optimal in a given node. Theminimum return difference between a suboptimal action andthe optimal one at depth d is δd = mini∈1,··· ,K\i∗d(∆d

i ).

Lemma 1. Lower bound for the number of trials. For anysub-tree Td developed with a budget b(d) > K, there exist aconstant ρ ≥ 0 such that T di,b(d) ≥ dρ ln(b(d))e for all i ∈1, · · · ,K. Furthermore, we have the following sequenceof lower bounds for the budget with d·e the ceiling function:

b(d = 0) = n

b(d) ≥ dρ ln(b(d− 1))e .

Lemma 2. Upper bound on the failure probability at depthd given the budget b(d). For any sub-tree Td developed witha budget b(d) > K we have the following upper bound on thefailure probability, conditioned by the budget b(d):

P (Id 6= i∗d|b(d)) ≤ b(d)−ρ2 (δ

d)2 .1A longer version of the paper including the proofs can be found

at https://arxiv.org/abs/1805.01367

Page 6: Open Loop Execution of Tree-Search Algorithms · 2018. 10. 23. · Open Loop Execution of Tree-Search Algorithms Erwan Lecarpentier 1;2, Guillaume Infantes 1, Charles Lesire 1, Emmanuel

Theorem 1. Upper bound on the failure probability at depthd. For an initial budget of n and for any sub-tree Td devel-oped with a budget b(d) > K, we have the following recur-sive relation for the upper bound on the failure probability,conditioned by the initial budget n:

P (Id 6= i∗d|n) ≤ dρ ln(b(d− 1))e− ρ2 (δd)2 .Additionally, for any depth d ≥ 1 given the initial budget n:

P (Id 6= i∗d|n) ≤ fd(n)− ρ2 (δd)2

f : t 7→ dρ ln(t)e.

Where fd = f fd−1 with f1 = f and d > 0.

This result shows a logarithmic decay between the upperbounds on the failure probability of two subsequent trees.Asymptotically, at any depth, this upper bound converges to-wards zero. This result highlights the fact that the deeper thesub-tree is, the less one can rely on the recommended actionat the root node. However, we should note that these upperbounds are derived without making further hypotheses on theMDP and express a worst-case value. We show in the nextsection that equal performances to OLUCT can be reachedwith a smaller computational budget and number of calls tothe generative model.

5 Empirical AnalysisWe compared OLUCT with OLTA on a discrete 1D track en-vironment2 and a continuous Physical Travelling SalesmanProblem3 (PTSP) [Perez et al., 2012b]. We implemented fivedecision criteria, leading to five variations of OLTA.

5.1 Heuristic decision criteriaA relevant decision criterion w.r.t. the treated problem allowsOLTA to discard a sub-tree when its first recommended ac-tion may not be state-wise optimal given the current state ofthe agent. We implemented five different tests to base thisdecision on, and evaluated them independently, which led tothe following variations of OLTA.

Plain OLTA. The simplest decision criterion that discardsa sub-tree only if its root-node is not fully expanded.

State Distribution Modality (SDM-OLTA). Test whetherthe empirical state distribution is multi-modal or not. If yes,discard the tree if the current state of the agent does not be-long to a majority mode. We define a majority mode by amode comprising more than τSDM% of the sampled states.

State Distribution Variance (SDV-OLTA). Test whetherthe empirical state distribution variance is above a certainthreshold τSDV . Discard the tree if it is the case. For multi-dimensional state spaces such as in the PTSP, the Variance-Mean-Ratio (VMR) is considered for the different orders ofmagnitude to be comparable.

State Distance to State Distribution (SDSD-OLTA).Compute the Mahalanobis distance [De Maesschalck et al.,

Code available at:2https://github.com/erwanlecarpentier/1dtrack.git3https://github.com/erwanlecarpentier/flatland.git

Transition misstep probability

100

101

Loss

(tim

e st

eps

to th

e go

al)

Transition misstep probability

0%

10%

20%

30%

Mea

n lo

ss r

elat

ivel

yto

OLU

CT

loss

Transition misstep probability

100

101

Com

puta

tiona

lco

st (m

s)

0.0 0.1 0.2 0.3 0.4 0.5Transition misstep probability

103

104

Num

ber

of c

alls

OLUCTPlain OLTA

SDM-OLTASDV-OLTA

SDSD-OLTARDV-OLTA

Figure 3: Comparison between OLUCT and OLTA on the discrete1D track environment for varying values of q.

2000] of the current state from the empirical state distribution.Discard the tree if it is above a selected threshold τSDSD.

Return Distribution Variance (RDV-OLTA). Testwhether the empirical return distribution variance is above acertain threshold τRDV . Discard the tree if it is the case.

A more selective decision criterion can easily be derivedby combining the previously described decision criteria anddiscarding the tree if one of them recommends to do so.

5.2 1D Track EnvironmentThe 1D track environment (Figure 2), is a 1D discrete worldwhere an agent can either go right or left. The initial state isthe “middle” state s0 = s2. The reward is 0 everywhere ex-cept for the transition to the two terminal states s0 and s4 forwhich it is +1. The action space is A = right, left. Weintroduce a transition misstep probability q ∈ [0, 1] whichis the probability to end up in the opposite state after tak-ing an action, for i ∈ 1, 2, 3: P (si−1|si, right) = q andP (si+1|si, right) = 1 − q. The same applies for the leftaction. If q < 0.5, the optimal policy πoptimal is to go leftat s1, to act randomly at s2 and to go right at s3. The sim-ulation settings are: q ∈ 0.0, 0.05, · · · , 0.5; n = 20 (bud-get); πdefault = πoptimal; H = 10 (simulation horizon forπdefault); Cp = 0.7; γ = 0.9. The decision criteria parame-ters were tuned to: τSDM = 80; τSDV = 0.4; τSDSD = 1;τRDV = 0.9. We generated 1000 episodes for each value of qand recorded 3 performance measures: loss (number of timesteps to termination); computational cost (measured compu-tation time); and number of calls to the generative model. We

Page 7: Open Loop Execution of Tree-Search Algorithms · 2018. 10. 23. · Open Loop Execution of Tree-Search Algorithms Erwan Lecarpentier 1;2, Guillaume Infantes 1, Charles Lesire 1, Emmanuel

Trajectory derived by anOLUCT algorithm in ourPTSP setting. The startingpoint is displayed in red, thewaypoints in green and thewalls in grey.

Figure 4: PTSP illustration

display two different graphs of the loss, the second one high-lights the relative performance between OLTA and OLUCT.

The motivation behind the use of such a benchmark isto test open loop control in a highly stochastic environmentwhere feedback of the current state is highly informativeabout the optimal action. In case of misstep for the first ac-tion, OLTA has to guess that a re-planning should be triggeredwhile OLUCT does it systematically. As seen on Figure 3,the non-plain OLTA and OLUCT achieved a very comparableloss. Plain-OLTA had a weaker performance due to its sys-tematic re-use of the sub-trees. Notice that some variationsof OLTA such as SDV-OLTA achieved a better mean lossthan OLUCT for some values of q. Due to the high variance,this observation cannot lead to the conclusion that OLTA canoutperform OLUCT. However, this emphasizes the fact thatthe performance are very similar. In terms of both computa-tional cost and number of calls to the generative model, OLTAwidely outperforms OLUCT. As q increases, this computa-tional gain vanishes and catches up with OLUCT for SDM-OLTA and SDV-OLTA. This accounts for the discriminativepower of their decision criteria that discard more trees. RDV-OLTA and SDSD-OLTA kept a lower computational costwhile reasonably matching the performance of OLUCT. Ob-viously, the computational cost of Plain-OLTA stays low. Theapparent similarity between the number of calls to the genera-tive model and the computational cost proves that computingour decision criteria is less expensive than re-planning.

5.3 Physical Travelling Salesman ProblemThe PTSP is a continuous navigation problem in which anagent must reach all the waypoints within a maze (Figure 4).The state of the agent is s = (x, y, θ, v) ∈ R4 i.e. the2D position, orientation and velocity. The action space isA = +dθ, 0,−dθ which consists of the increment, decre-ment or no-change of the orientation. The reward is +1 whena waypoint is reached for the first time, −1 for a wall crashand 0 otherwise. The simulation terminates when the agentreaches all the waypoints or a time limit. The walls cannotbe crossed and the orientation is flipped when a crash occurs.We introduce a misstep probability q ∈ [0, 1] which is theprobability for another action to be undertaken instead of thecurrent one. A Gaussian noise of standard deviation σnoise isadded to each component of the resulting state from a tran-sition. The simulation settings are: s0 = (1.1, 1.1, 0, 0.1);q ∈ 0.0, 0.05, · · · , 0.5; σnoise = 0.02; n = 300 (initialtree budget); πdefault = πgo−straight that applies no orien-tation variation; H = 50 (simulation horizon for πdefault);

Transition misstep probability0

200400600800

1000

Loss

(tim

e st

eps

to r

each

all

the

way

poin

ts)

Transition misstep probability0%

50%

100%

150%

Mea

n lo

ssre

lativ

ely

toO

LUC

T lo

ss

Transition misstep probability0

10000

20000

30000

40000

50000

Com

puta

tiona

lco

st (m

s)

0.0 0.1 0.2 0.3 0.4 0.5Transition misstep probability

0.000.250.500.751.001.25

Num

ber

ofca

lls

1e7

OLUCTPlain OLTA

SDV-OLTASDSD-OLTA

RDV-OLTA

Figure 5: Comparison between OLUCT and OLTA on the continu-ous PTSP for varying values of q.

Cp = 0.7; γ = 0.99. The provided map is the one depictedin Figure 4 with three waypoints. The different decision cri-teria parameters were tuned to: τSDV = 0.02; τSDSD = 1;τRDV = 0.1. We reserve the development of SDM-OLTAin the continuous case for future work. We generated 100episodes for each transition misstep probability and recordedthe same performance measures as in the 1D track case. Theresults are presented in Figure 5. OLUCT, SDSD-OLTA andRDV-OLTA achieved a comparable loss for every q, whichshows that our method is applicable to larger scale problemsthan the 1D track environment. SDV-OLTA reached a lowerlevel of performance. Plain OLTA still realized the highestloss since it is highly sensitive to the stochasticity of the en-vironment. In terms of both computational cost and numberof calls to the generative model, the same trade-off betweenperformance and computational cost is observed. Plain OLTAand SDV-OLTA considerably lowered the number of calls atthe cost of the performance while SDSD-OLTA and RDV-OLTA realized a better compromise. The number of callsto the generative model and the computational cost are quitesimilar, meaning that — even with the higher dimensionalityof the PTSP compared to the 1D track — the cost incurredby the decision criteria computation is negligible in compari-son to the one incurred by the re-planning procedure. Noticethat SDV-OLTA achieved a good cost-performance trade-offin the 1D track environment while not in the PTSP relativelyto the other algorithms. This is explained by the decisioncriteria’s sensitivity to parameter tuning and by the problem-dependent relevance of such a criterion.

Page 8: Open Loop Execution of Tree-Search Algorithms · 2018. 10. 23. · Open Loop Execution of Tree-Search Algorithms Erwan Lecarpentier 1;2, Guillaume Infantes 1, Charles Lesire 1, Emmanuel

6 ConclusionWe introduced OLTA, a new class of tree-search algorithmsperforming open loop control by re-using subsequent sub-trees of a main tree built with the OLUCT algorithm. A de-cision criterion based on the analysis of the current sub-treeallows the agent to efficiently determine if the latter can be ex-ploited. Practically, OLTA can achieve the same level of per-formance as OLUCT given that the decision criterion is welldesigned. Furthermore, the computational cost is stronglylowered by decreasing the number of calls to the generativemodel. This saving is the main interest of the approach andcan be exploited in two ways: it decreases the energy con-sumption which is relevant for critical systems with low re-sources such as Unmanned Vehicles or Satellites; It allows asystem to re-allocate the computational effort to other tasksrather than controlling the robot. We emphasize the fact thatthis method is generic and can be combined with any othertree-search algorithm than OLUCT. Open questions includebuilding non problem-dependent decision criteria, e.g. bymaking more restrictive hypothesis on the considered classof MDPs, but also applying the method to other benchmarksand other open loop planners.

AcknowledgementsThis research was supported by the Occitanie region, France.

References[Auer et al., 2002] Peter Auer, Nicolo Cesa-Bianchi, and

Paul Fischer. Finite-time analysis of the multiarmed banditproblem. Machine learning, 47(2-3):235–256, 2002.

[Bellemare et al., 2017] Marc G. Bellemare, Will Dabney,and Remi Munos. A distributional perspective on re-inforcement learning. arXiv preprint arXiv:1707.06887,2017.

[Bubeck and Munos, 2010] Sebastien Bubeck and RemiMunos. Open loop optimistic planning. In COLT, 2010.

[De Maesschalck et al., 2000] Roy De Maesschalck, Del-phine Jouan-Rimbaud, and Desire L. Massart. The Maha-lanobis distance. Chemometrics and intelligent laboratorysystems, 50(1):1–18, 2000.

[Enzenberger et al., 2010] Markus Enzenberger, MartinMuller, Broderick Arneson, and Richard Segal. Fuego- an open-source framework for board games and Goengine based on Monte Carlo tree search. IEEE Trans-actions on Computational Intelligence and AI in Games,2(4):259–270, 2010.

[Heusner, 2011] Manuel Heusner. UCT for pac-man. Bach-elor thesis, Univ. of Basel, 2011.

[Kaelbling et al., 1998] Leslie Pack Kaelbling, Michael L.Littman, and Anthony R. Cassandra. Planning and act-ing in partially observable stochastic domains. Artificialintelligence, 101(1):99–134, 1998.

[Keller and Helmert, 2013] Thomas Keller and MalteHelmert. Trial-based heuristic tree search for finitehorizon MDPs. In ICAPS, 2013.

[Kocsis and Szepesvari, 2006] Levente Kocsis and CsabaSzepesvari. Bandit based Monte Carlo planning. InECML, volume 6, pages 282–293. Springer, 2006.

[Perez et al., 2012a] Diego Perez, Philipp Rohlfshagen, andSimon M. Lucas. Monte Carlo tree search: Long-termversus short-term planning. In Computational Intelligenceand Games (CIG), 2012 IEEE Conference on, pages 219–226. IEEE, 2012.

[Perez et al., 2012b] Diego Perez, Philipp Rohlfshagen, andSimon M. Lucas. The physical travelling salesman prob-lem: WCCI 2012 competition. In Evolutionary Compu-tation (CEC), 2012 IEEE Congress on, pages 1–8. IEEE,2012.

[Puterman, 2014] Martin L. Puterman. Markov decision pro-cesses: discrete stochastic dynamic programming. JohnWiley & Sons, 2014.

[Rachelson and Lagoudakis, 2010] Emmanuel Rachelsonand Michail G. Lagoudakis. On the locality of actiondomination in sequential decision making. In ISAIM,2010.

[Silver and Veness, 2010] David Silver and Joel Veness.Monte Carlo planning in large POMDPs. In Advances inneural information processing systems, pages 2164–2172,2010.

[Silver et al., 2008] David Silver, Richard S. Sutton, andMartin Muller. Sample-based learning and search withpermanent and transient memories. In Proceedings of the25th international conference on Machine Learning, pages968–975. ACM, 2008.

[Silver et al., 2016] David Silver, Aja Huang, Chris J.Maddison, Arthur Guez, Laurent Sifre, George VanDen Driessche, Julian Schrittwieser, Ioannis Antonoglou,Veda Panneershelvam, Marc Lanctot, et al. Mastering thegame of Go with deep neural networks and tree search.Nature, 529(7587):484–489, 2016.

[Sutton and Barto, 1998] Richard S. Sutton and Andrew G.Barto. Reinforcement learning: An introduction. MITpress Cambridge, 1998.

[Sutton, 1991] Richard S. Sutton. Dyna, an integrated archi-tecture for learning, planning, and reacting. ACM SIGARTBulletin, 2(4):160–163, 1991.

[Weinstein and Littman, 2012] Ari Weinstein andMichael L. Littman. Bandit-based planning and learningin continuous-action markov decision processes. InICAPS, 2012.

[Wilson et al., 2014] Mark A. Wilson, James McMahon, andDavid W. Aha. Bounded expectations for discrepancy de-tection in goal-driven autonomy. In AI and Robotics: Pa-pers from the AAAI Workshop, 2014.

[Wilson et al., 2016] Mark A. Wilson, James McMahon, Ar-tur Wolek, David W. Aha, and Brian H. Houston. Towardgoal reasoning for autonomous underwater vehicles: Re-sponding to unexpected agents. In Goal Reasoning: Pa-pers from the IJCAI Workshop, 2016.