I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is...

76
Introduction to Articial Intelligence Lecture 7: Reasoning over time Prof. Gilles Louppe [email protected] 1 / 75

Transcript of I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is...

Page 1: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Introduction to Arti�cialIntelligence

Lecture 7: Reasoning over time

Prof. Gilles Louppe [email protected]

1 / 75

Page 2: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Markov models

Markov processes

Inference tasks

Hidden Markov models

Filters

Kalman �lter

Particle �lter

TodayMaintain a belief state about the world, and update it as time passes andevidence is collected.

Do not overlook this lecture!

―――Image credits: CS188, UC Berkeley. 2 / 75

Page 3: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Pacman revenge: How to make good use of the sonar readings?

0:00 / 1:00

―――Image credits: CS188, UC Berkeley. 3 / 75

Page 4: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Markov models

4 / 75

Page 5: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Reasoning over timeOften, we want to reason about a sequence of observations:

Robot localization

Speech recognition

Medical monitoringn

Machine translation

Part-of-speech tagging

Handwriting recognition

...

For this reason, we need to introduce time (or space) in our model.

5 / 75

Page 6: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Modelling the passage of time

We will consider the world as a discrete series of time slices, each of whichcontains a set of random variables:

denotes the set of unobservable state variables at time .

denotes the set of observable evidence variables at time .

X t t

E t t

6 / 75

Page 7: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

We specify:

a prior that de�nes our inital belief state over hidden state variables.

a transition model (for ) that de�nes the probability

distribution over the latest state variables, given the previous (unobserved)values.

a sensor model (for ) that de�nes the probability

distribution over the latest evidence variables, given all previous (observedand unobserved) values.

P(X )0

P(X ∣X )t 0:t−1 t > 0

P(E ∣X ,E )t 0:t 0:t−1 t > 0

7 / 75

Page 8: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Markov processesMarkov assumption

The current state of the world depends only on its immediate previous

state(s), i.e., depends on only a bounded subset of .

Random processes that satisfy this assumption are called Markov processes.

X t X 0:t−1

8 / 75

Page 9: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

First-order Markov processes

Markov processes such that

i.e., and are conditionally independent given .

P(X ∣X ) = P(X ∣X ).t 0:t−1 t t−1

X t X 0:t−2 X t−1

9 / 75

Page 10: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Second-order Markov processes

Markov processes such that

i.e., and are conditionally independent given and .

P(X ∣X ) = P(X ∣X ,X ).t 0:t−1 t t−2 t−1

X t X 0:t−3 X t−1 X t−2

10 / 75

Page 11: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Sensor Markov assumption

Additionally, we make a (�rst-order) sensor Markov assumption:

Stationarity assumption

The transition and the sensor models are the same for all (i.e., the laws of

physics do not change with time).

P(E ∣X ,E ) = P(E ∣X )t 0:t 0:t−1 t t

t

11 / 75

Page 12: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Joint distribution

A Markov process can be described as a growable Bayesian network, unrolledin�nitely through time, with a speci�ed restricted structure between time steps.

Therefore, the joint distribution of all variables up to in a (�rst-order) Markov

process is

t

P(X ,E ) = P(X ) P(X ∣X )P(E ∣X ).0:t 1:t 0i=1

∏t

i i−1 i i

12 / 75

Page 13: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

?

?

?

Example: Will you take your umbrella today?

P(Umbrella ∣Rain )t t

P(Rain ∣Umbrella )t 0:t−1

P(Rain ∣Rain )t+2 t

―――Image credits: CS188, UC Berkeley. 13 / 75

Page 14: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

The transition model can equivalently be represented by a

state transition diagram.

P(Rain ∣Rain )t t−1

14 / 75

Page 15: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Inference tasksFiltering:

Filtering is what a rational agent does to keep track of the current hidden state , its beliefstate, so that rational decisions can be made.

Prediction: for

Computing the posterior distribution over future states.

Used for evaluation of possible action sequences.

Smoothing: for

Computing the posterior distribution over past states.

Used for building better estimates, since it incorporates more evidence.

Essential for learning.

Most likely explanation:

Decoding with a noisy channel, speech recognition, etc.

P(X ∣e )t 1:t

X t

P(X ∣e )t+k 1:t k > 0

P(X ∣e )k 1:t 0 ≤ k < t

arg max P (x ∣e )x 1:t 1:t 1:t

15 / 75

Page 16: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

(Predict) Push forward

through the transition model.(Update) Update with the

evidence , given the sensor model.

Base cases

P(X )2 = P(X ,x )x 1

∑ 2 1

= P (x )P(X ∣x )x 1

∑ 1 2 1

P(X )1

P(X ∣e )1 1 =

P (e )1

P(X , e )1 1

∝ P(X , e )1 1

= P(X )P(e ∣X )1 1 1

P(X )1

e 1

―――Image credits: CS188, UC Berkeley. 16 / 75

Page 17: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Prediction

To predict the future :

Push the prior belief state through the transition model:

Repeat up to , using to compute .

P(X ∣e )t+k 1:t

P(X ∣e )t 1:t

P(X ∣e ) = P(X ∣x )P (x ∣e )t+1 1:tx t

∑ t+1 t t 1:t

t + k P(X ∣e )t+k−1 1:t P(X ∣e )t+k 1:t

―――Image credits: CS188, UC Berkeley. 17 / 75

Page 18: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Random dynamics (Ghostbusters)

0:00 / 1:20

―――Image credits: CS188, UC Berkeley. 18 / 75

Page 19: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Circular dynamics (Ghostbusters)

0:00 / 0:32

―――Image credits: CS188, UC Berkeley. 19 / 75

Page 20: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Whirlpool dynamics (Ghostbusters)

0:00 / 0:41

―――Image credits: CS188, UC Berkeley. 20 / 75

Page 21: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

As time passes, uncertainty "accumulates" if we do not accumulate new evidence.

―――Image credits: CS188, UC Berkeley. 21 / 75

Page 22: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Stationary distributionsWhat if ?

For most chains, the in�uence of the initial distribution gets lesser and lesserover time.

Eventually, the distribution converges to a �xed point, called the stationarydistribution.

This distribution is such that

t → ∞

P(X ) = P(X ) = P(X ∣x )P(x )∞ ∞+1x ∞

∑ ∞+1 ∞ ∞

22 / 75

Page 23: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

0.9

0.1

0.3

0.7

Example

Therefore, .

Which implies that and .

Xt−1 Xt P

sun sun

sun rain

rain sun

rain rain

P (X = sun) =∞

=

=

P (X = sun)∞+1

P (X = sun∣X = sun)P (X = sun)∞+1 ∞ ∞

+ P (X = sun∣X = rain)P (X = rain)∞+1 ∞ ∞

0.9P (X = sun) + 0.3P (X = rain)∞ ∞

P (X = sun) = 3P (X = rain)∞ ∞

P (X = sun) = ∞ 43 P (X = rain) = ∞ 4

1

23 / 75

Page 24: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Filtering

What if we collect new observations? Beliefs get reweighted, and uncertainty"decreases":

P(X ∣e ) = αP(e ∣X )P(X ∣e )t+1 1:t+1 t+1 t+1 t+1 1:t

24 / 75

Page 25: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Bayes �lter

An agent maintains a belief state estimate and updates it as new

evidences are collected.

Recursive Bayesian estimation:

(Predict step): Project the current belief state forward from to

through the transition model.

(Update step): Update this new state using the evidence .

P(X ∣e )t 1:t

e t+1

P(X ∣e ) = f(e ,P(X ∣e ))t+1 1:t+1 t+1 t 1:t

t t + 1

e t+1

25 / 75

Page 26: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

where

the normalization constant

is used to make probabilities sum to 1;

in the last expression, the �rst and second terms are given by the model whilethe third is obtained recursively.

P(X ∣e )t+1 1:t+1 = P(X ∣e , e )t+1 1:t t+1

= αP(e ∣X , e )P(X ∣e )t+1 t+1 1:t t+1 1:t

= αP(e ∣X )P(X ∣e )t+1 t+1 t+1 1:t

= αP(e ∣X ) P(X ∣x , e )P (x ∣e )t+1 t+1xt

∑ t+1 t 1:t t 1:t

= αP(e ∣X ) P(X ∣x )P (x ∣e )t+1 t+1xt

∑ t+1 t t 1:t

α = = 1/ P (e ∣x )P (x ∣e )P (e ∣e )t+1 1:t

1

x t+1

∑ t+1 t+1 t+1 1:t

26 / 75

Page 27: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

We can think of as a message that is propagated forward along

the sequence, modi�ed by each transition and updated by each new observation.

Thus, the process can be implemented as .

The complexity of a forward update is constant (in time and space) with .

P(X ∣e )t 1:t f 1:t

f = α forward(f , e )1:t+1 1:t t+1

t

27 / 75

Page 28: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Example

R t−1 P (R )t

true 0.7

false 0.3

R t P (U )t

true 0.9

false 0.2

28 / 75

Page 29: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Ghostbusters with a Bayes �lter

0:00 / 0:36

―――Image credits: CS188, UC Berkeley. 29 / 75

Page 30: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

SmoothingWe want to compute for .

Divide evidence into and . Then,

P(X ∣e )k 1:t 0 ≤ k < t

e 1:t e 1:k e k+1:t

P(X ∣e )k 1:t = P(X ∣e , e )k 1:k k+1:t

= αP(X ∣e )P (e ∣X , e )k 1:k k+1:t k 1:k

= αP(X ∣e )P (e ∣X ).k 1:k k+1:t k

30 / 75

Page 31: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Let the backward message correspond to . Then,

where is a pointwise multiplication of vectors.

This backward message can be computed using backwards recursion:

The �rst and last factors are given by the model. The second factor is obtainedrecursively. Therefore,

b k+1:t P(e ∣X )k+1:t k

P(X ∣e ) = α f × b ,k 1:t 1:k k+1:t

×

P(e ∣X )k+1:t k = P(e ∣X ,x )P(x ∣X )x k+1

∑ k+1:t k k+1 k+1 k

= P (e ∣x )P(x ∣X )x k+1

∑ k+1:t k+1 k+1 k

= P (e ∣x )P (e ∣x )P(x ∣X )x k+1

∑ k+1 k+1 k+2:t k+1 k+1 k

b = backward(b , e ).k+1:t k+2:t k+1

31 / 75

Page 32: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Forward-backward algorithm

Complexity:

Smoothing for a particular time step takes:

Smoothing a whole sequence (because of caching):

k O(t)

O(t)

32 / 75

Page 33: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Example

33 / 75

Page 34: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Most likelyexplanationSuppose that is the umbrella sequence.

What is the weather sequence that is the most likely to explain this?

Does the absence of umbrella at day 3 means it wasn't raining?

Or did the director forget to bring it?

If it didn't rain on day 3, perhaps it didn't rain on day 4 either, but the directorbrought the umbrella just in case?

Among all sequences, is there an (ef�cient) way to �nd the most likely one?

[true, true, false, true, true]

25

―――Image credits: CS188, UC Berkeley. 34 / 75

Page 35: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

The most likely sequence is not the sequence of the most likely states!

The most likely path to each , is the most likely path to some plus one

more step. Therefore,

Identical to �ltering, except that the forward message is

replaced with

where gives the probability of the most likely path to state .

The update has its sum replaced by max, resulting in the Viterbi algorithm:

x t+1 x t

P(x ,X ∣e )x 1:t

max 1:t t+1 1:t+1

= αP(e ∣X ) (P(X ∣x ) P(x ,x ∣e ))t+1 t+1xt

max t+1 tx 1:t−1max 1:t−1 t 1:t

f = P(X ∣e )1:t t 1:t

m = P(x ,X ∣e ),1:tx 1:t−1max 1:t−1 t 1:t

m (i)1:t i

m = αP(e ∣X ) P(X ∣x )m 1:t+1 t+1 t+1x t

max t+1 t 1:t

35 / 75

Page 36: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Example

36 / 75

Page 37: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Hidden Markov modelsSo far, we described Markov processes over arbitrary sets of state variables

and evidence variables .

A hidden Markov model (HMM) is a Markov process in which the state

and the evidence are both single discrete random variables.

, with domain

, with domain

This restricted structure allows for a reformulation of the forward-backwardalgorithm in terms of matrix-vector operations.

X t

E t

X t

E t

X = X t t D = {1, ..., S}X t

E = E t t D = {1, ..., R}E t

37 / 75

Page 38: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Note on terminology

Some authors instead divide Markov models into two classes, depending on theobservability of the system state:

Observable system state: Markov chains

Partially-observable system state: Hidden Markov models.

We follow here instead the terminology of the textbook.

38 / 75

Page 39: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Simpli�ed matrix algorithms

The prior becomes a (normalized) column vector .

The transition model becomes an transition matrix ,

such that

The sensor model is de�ned as an sensor matrix , such

that

P(X )0 f ∈ R 0 +S

P(X ∣X )t t−1 S × S T

T = P (X = j∣X = i).ij t t−1

P(E ∣X )t t S × R B

B = P (E = j∣X = i).ij t t

39 / 75

Page 40: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Let the observation matrix be a diagonal matrix whose elements

corresponds to the column of the sensor matrix .

If we use column vectors to represent forward and backward messages, thenwe have

where is an all-one vector of size .

Therefore the forward-backward algorithm needs time and space

.

O t

e t B

f = αO T f 1:t+1 t+1T

1:t

b = TO b ,k+1:t k+1 k+2:t

b t+1:t S

O(S t)2

O(St)

40 / 75

Page 41: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Example

Suppose that is the umbrella sequence.

See code/lecture7-forward-backward.ipynb for the execution.

[true, true, false, true, true]

f 0

T

B

O = O = O = O 1 2 4 5

O 2

= (0.50.5

)

= (0.70.3

0.30.7

)

= (0.90.2

0.10.8

)

= (0.90.0

0.00.2

)

= (0.10.0

0.00.8

)

41 / 75

Page 42: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Stationary distribution

The stationary distribution of a HMM is a distribution such that

Therefore, the stationary distribution corresponds to the (normalized)eigenvector of the transposed transition matrix with an eigenvalue of .

f

f = T f .T

1

42 / 75

Page 43: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Filters

43 / 75

Page 44: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Suppose we want to track the position and velocity of a robot from noisyobservations collected over time.

Formally, we want to estimate continuous state variables such as

the position of the robot at time ,

the velocity of the robot at time .

We assume discrete time steps.

X t t

Xt t

―――Image credits: CS188, UC Berkeley. 44 / 75

Page 45: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Continuous variablesLet be a random variable.

When is uncountably in�nite (e.g., ), is called a continuous

random variable.

If is absolutely continuous, its probability distribution is described by a

density function that assigns a probability to any interval

such that

where is non-negative piecewise continuous and such that

X : Ω → D X

D X D = RX X

X

p [a, b] ⊆ D X

P (a < X ≤ b) = p(x)dx,∫a

b

p

p(x)dx = 1.∫D X

45 / 75

Page 46: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Uniform

The uniform distribution is described by the density function

where and are the bounds of its support.

U (a, b)

p(x) = { b−a

1

0if x ∈ [a, b]otherwise

a ∈ R b ∈ R

46 / 75

Page 47: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Normal

The normal (or Gaussian) distribution is described by the density

function

where and are its mean and standard deviation parameters.

N (μ, σ)

p(x) = exp −

2πσ2

1(

2σ2

(x − μ)2

)

μ ∈ R σ ∈ R+

47 / 75

Page 48: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Multivariate normal

The multivariate normal distribution generalizes to random variables. Its (joint)

density function is de�ned as

where and is positive semi-de�nite.

N

p(x = x , ..., x ) = exp − (x− μ) Σ (x− μ)1 n (2π) ∣Σ∣n

1(

21 T −1 )

μ ∈ Rn Σ ∈ Rn×n

48 / 75

Page 49: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

The (multivariate) Normal density is the only density for real random variablesthat is closed under marginalization and multiplication. Also

a linear (or af�ne) function of a Normal random variable is Normal;

a sum of Normal variables is Normal.

For these reasons, most algorithms discussed in this course are tractable only fordiscrete random variables or Normal random variables.

49 / 75

Page 50: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Continuous Bayes �lterThe Bayes �lter similarly applies to continuous state and evidence variables

and , in which case summations are replaced with integrals and probability

mass functions with probability densities:

where the normalization constant is

X t

E t

p(x ∣e )t+1 1:t+1 = α p(e ∣x ) p(x ∣x )p(x ∣e )dx t+1 t+1 ∫ t+1 t t 1:t t

α = 1 / p(e ∣x )p(x ∣e )dx .∫ t+1 t+1 t+1 1:t t+1

50 / 75

Page 51: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Kalman �lterThe Kalman �lter is a special case of the Bayes �lter, which assumes:

Gaussian prior

Linear Gaussian transition model

Linear Gaussian sensor model

51 / 75

Page 52: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Transition model

Sensor model

Linear Gaussian models

p(x ∣x ) = N (x ∣Ax + b,Σ )t+1 t t+1 t x

p(e ∣x ) = N (e ∣Cx + d,Σ )t t t t e

52 / 75

Page 53: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Cheat sheet for Gaussian models (Bishop, 2006)

Given a marginal Gaussian distribution for and a linear Gaussian distribution

for given in the form

the marginal distribution of and the conditional distribution of given are

given by

where

xy x

p(x)p(y∣x)

= N (x∣μ,Λ )−1

= N (y∣Ax+ b,L )−1

y x y

p(y)

p(x∣y)

= N (y∣Aμ+ b,L +AΛ A )−1 −1 T

= N (x∣Σ A L(y − b) +Λμ ,Σ)( T )

Σ = (Λ+A LA) .T −1

53 / 75

Page 54: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Filtering Gaussian distributions

Prediction step:

If the distribution is Gaussian and the transition model

is linear Gaussian, then the one-step predicted distribution given

by

is also a Gaussian distribution.

Update step:

If the prediction is Gaussian and the sensor model

is linear Gaussian, then after conditioning on new evidence,

the updated distribution

is also a Gaussian distribution.

p(x ∣e )t 1:t

p(x ∣x )t+1 t

p(x ∣e ) = p(x ∣x )p(x ∣e )dx t+1 1:t ∫ t+1 t t 1:t t

p(x ∣e )t+1 1:t

p(e ∣x )t+1 t+1

p(x ∣e ) ∝ p(e ∣x )p(x ∣e )t+1 1:t+1 t+1 t+1 t+1 1:t

54 / 75

Page 55: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Therefore, for the Kalman �lter, is a multivariate Gaussian distribution

for all .

Filtering reduces to the computation of the parameters and .

By contrast, for general (nonlinear, non-Gaussian) processes, the descriptionof the posterior grows unboundedly as .

p(x ∣e )t 1:t

N (x ∣μ ,Σ )t t t t

μ t Σ t

t → ∞

55 / 75

Page 56: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

1D example

Gaussian random walk:

Gaussian prior:

The transition model adds random perturbations of constant variance:

The sensor model yields measurements with Gaussian noise of constantvariance:

p(x ) = N (x ∣μ , σ )0 0 0 02

p(x ∣x ) = N (x ∣x , σ )t+1 t t+1 t x2

p(e ∣x ) = N (e ∣x , σ )t t t t e2

56 / 75

Page 57: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

The one-step predicted distribution is given by

Note that the same result can be obtained by using instead the Gaussian modelsidentities.

p(x )1 = p(x ∣x )p(x )dx ∫ 1 0 0 0

= α exp − exp − dx ∫ (21

σ x2

(x − x )1 02

) (21

σ 02

(x − μ )0 02

) 0

= α exp − dx ∫ (21

σ σ 02

x2

σ (x − x ) + σ (x − μ )02

1 02

x2

0 02

) 0

... (simplify by completing the square)

= α exp − (21

σ + σ 02

x2

(x − μ )1 02

)

= N (x ∣μ , σ + σ )1 0 02

x2

57 / 75

Page 58: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

For the update step, we need to condition on the observation at the �rst timestep:

p(x ∣e )1 1 = αp(e ∣x )p(x )1 1 1

= α exp − exp − (21

σ e2

(e − x )1 12

) (21

σ + σ 02

x2

(x − μ )1 02

)

= α exp −

⎝⎜⎛

21

σ +σ +σ 02

x2

e2

(σ +σ )σ 02

x2

e2

x − ( 1 σ +σ +σ 02

x2

e2

(σ +σ )e +σ μ 02

x2

1 e2

0 )2

⎠⎟⎞

= N x , ( 1∣∣∣∣

σ + σ + σ 02

x2

e2

(σ + σ )e + σ μ 02

x2

1 e2

0

σ + σ + σ 02

x2

e2

(σ + σ )σ 02

x2

e2

)

58 / 75

Page 59: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

In summary, the update equations given a new evidence are:e t+1

μ t+1

σ t+12

=

σ + σ + σ t2

x2

e2

(σ + σ )e + σ μ t2

x2

t+1 e2

t

=

σ + σ + σ t2

x2

e2

(σ + σ )σ t2

x2

e2

59 / 75

Page 60: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

General Kalman update

The same derivations generalize to multivariate normal distributions.

Assuming the transition and sensor models

we arrive at the following general update equations:

where is the Kalman gain matrix.

p(x ∣x )t+1 t

p(e ∣x )t t

= N (x ∣Fx ,Σ )t+1 t x

= N (e ∣Hx ,Σ ),t t e

μt+1

Σt+1

Kt+1

= Fμ +K (e −HFμ )t t+1 t+1 t

= (I−K H)(FΣ F +Σ )t+1 tT

x

= (FΣ F +Σ )H (H(FΣ F +Σ )H +Σ )tT

xT

tT

xT

e−1

K t+1

60 / 75

Page 61: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

2D tracking: �ltering

61 / 75

Page 62: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

2D tracking: smoothing

62 / 75

Page 63: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Apollo guidance computer

The Kalman �lter put man on the Moon, literally!

The onboard guidance software of Saturn-V used a Kalman �lter to mergenew data with past position measurements to produce an optimal positionestimate of the spacecraft.

―――Credits: Apollo-11 source code 63 / 75

Page 64: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

The Kalman Filter (with music)The Kalman Filter (with music)The Kalman Filter (with music)Watch laterWatch laterWatch later ShareShareShare

64 / 75

Page 65: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Dynamic Bayesian networks

Dynamics Bayesian networks (DBNs) can be used for tracking multiple variablesover time, using multiple sources of evidence. Idea:

Repeat a �xed Bayes net structure at each time .

Variables from time condition on those from .

DBNs are a generalization of HMMs and of the Kalman �lter.

t

t t− 1

―――Image credits: CS188, UC Berkeley. 65 / 75

Page 66: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Exact inference

Unroll the network through time and run any exact inference algorithm (e.g.,variable elimination)

Problem: inference cost for each update grows with .

Rollup �ltering: add slice , sum out slice using variable elimination.

Largest factor is and the total update cost per step is .

Better than HMMs, which is , but still infeasible for large numbers of variables.

t

t+ 1 t

O(d )n+k O(nd )n+k

O(d )2n

66 / 75

Page 67: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Approximate inference

If exact inference in DBNs intractable, then let's use approximate inferenceinstead.

Likelihood weighting? Generated samples pay no attention to the evidence!

The fraction of samples that remain close to the actual series of events dropsexponentially with .

We need a better solution!

t

67 / 75

Page 68: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Particle �lterBasic idea:

Maintain a �nite population of samples, called particles.

The representation of our beliefs is a list of particles.

Ensure the particles track the high-likelihood regions of the state space.

Throw away samples that have very low weight, according to the evidence.

Replicate those that have high weight.

This scales to high dimensions!

N

―――Image credits: CS188, UC Berkeley. 68 / 75

Page 69: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Update cycle

―――Image credits: CS188, UC Berkeley. 69 / 75

Page 70: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

70 / 75

Page 71: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

Robot localization

(See demo)

71 / 75

Page 72: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

The RAGI robot makes use of a particle �lter to locate itself within Monte�ore. (See RTBF, mars 2019.)

72 / 75

Page 73: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

―――Credits: Tom Ewbank, RAGI. 73 / 75

Page 74: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

SummaryTemporal models use state and sensor variables replicated over time.

Their purpose is to maintain a belief state as time passes and as more evidence is collected.

The Markov and stationarity assumptions imply that we only need to specify

a transition model ,

a sensor model .

Inference tasks include �ltering, prediction, smoothing and �nding the mostlikely sequence.

Filter algorithms are all based on the core of idea of

projecting the current belief state through the transition model,

updating the prediction according to the new evidence.

P (X ∣X )t+1 t

P (E ∣X )t t

74 / 75

Page 75: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

The end.

74 / 75

Page 76: I n telligen ce I n tro d u ctio n to Articia l...I n f eren ce ta sk s F ilt e rin g : Filtering is what a rational agent does to keep track of the current hidden state , its belief

ReferencesKalman, Rudolph Emil. "A new approach to linear �ltering and predictionproblems." Journal of basic Engineering 82.1 (1960): 35-45.

Bishop, Christopher "Pattern Recognition and Machine Learning" (2006).

75 / 75