11. Markov Chains (MCs) 2 Courtesy of J. Bard, L. Page, and J. Heyl.

Post on 18-Jan-2018

213 views 0 download

description

Transition prob. matrix n-step transition prob. from state i to j is n-step transition matrix (for all states) is then For instance, two step transition matrix is 3

Transcript of 11. Markov Chains (MCs) 2 Courtesy of J. Bard, L. Page, and J. Heyl.

11. Markov Chains (MCs) 2

Courtesy of J. Bard, L. Page, and J. Heyl

11.2.1 n-step transition probabilities (review)

2

Transition prob. matrix

n-step transition prob. from state i to j is

n-step transition matrix (for all states) is then

For instance, two step transition matrix is

3

.Ej,i ,n]iX|jX[P)n(p kknij 0

)n(p)n(P ij

2112 P)(P)(P)(P

Chapman-Kolmogorov equations

Prob. of going from state i at t=0, passing though state k at t=m, and ending at state j at t=m+n is

In matrix notation,

4

.Ej,i all,m,n allfor

)n(p)m(p)nm(pEk

kjikij

0

mnP)m(P)n(P)mn(P nP)n(P

11.2.2 state probabilities

5

State probability (pmf of an RV!)Let p(n) = {pj(n)} be the row vector of state probabilities at time n (i.e. state prob. vector)

Thus, p(n) is given byFrom the initial state

In matrix notation

6

)n(pp

]iX[P]iX|jX[P)n(p

ii

ij

innnj

1

11

P)n()n( 1pp

)(p)n(p]iX[P]iX|jX[P)n(p ii

iji

nj 000

nP)()n(P)()n( 00 ppp

How an MC changes (Ex 11.10, 11.11)

A two-state system

7

Silence(state 0)

Speech(state 1)

0.10.90.8

0.2

80201090

..

..P

Suppose p(0)=(0,1)Then p(1) = p(0)P= (0,1)P = (0.2, 0.8)p(2)= (0.2,0.8)P= (0,1)P2 = (0.34,

0.66) p(4)= (0,1)P4 = (0.507, 0.493)p(8)= (0,1)P8 = (0.629, 0.371)p(16)= (0,1)P16 = (0.665, 0.335)

p(32)= (0,1)P32 = (0.667, 0.333)p(64)= (0,1)P64 = (0.667, 0.333)

Suppose p(0)=(1,0)p(1) = p(0)P = (0.9, 0.1)p(2) = (1,0)P2 = (0.83,

0.17)p(4)= (1,0)P4 = (0.747, 0.253)p(8)= (1,0)P8 = (0.686, 0.314)p(16)= (1,0)P16 = (0.668, 0.332)

p(32)= (1,0)P32 = (0.667, 0.333)p(64)= (1,0)P64 = (0.667, 0.333)

Independence of initial condition

8

The lesson to take away

No matter what assumptions you make about the initial probability distribution, after a large number of steps,the state probability distribution is approximately (2/3, 1/3)

9

See p.666, 667

11.2.3 steady state probabilities

10

State probabilities (pmf) convergeAs n , then transition prob. matrix Pn approaches a matrix whose rows are equal to the same pmf.

In matrix notation,

where 1 is a column vector of all 1’s, and =(1, 1, … )

The convergence of Pn implies the convergence of the state pmf’s

11

iallfor)n(p jij

jii

jii

ijj )(p)(p)n(p)n(p 00

1nP

Steady state probabilitySystem reaches “equilibrium” or “steady state”, i.e., n , pj(n) j, pi(n-1) i

In matrix notation,

here is stationary state pmf of the Markov chainTo solve this,

12

ii

ijj p

P

1i

i

Speech activity system

From the steady state probabilities

13

= P

(1, 2) = (1, 2) 0.9 0.10.2 0.8

1 = 0.91 + 0.12

2 = 0.21 + 0.82

1 + 2 = 1

1 = 2/3 = 0.6672 = 1/3 = 0.333

14

Question 11-1: Alice, Bob and Carol are playing Frisbee. Alice always throws to Carol. Bob always throws to Alice.Carol throws to Bob 2/3 of the time and to Alice 1/3 of the time.

In the long run, what percentage of the time do each of the players have the Frisbee?

11.3.1 classes of states11.3.2 recurrence properties

15

Why classification?

The methods we have learned with steady state probabilities (in Section 11.2.3) work only with regular Markov chains (MCs).

A regular MC has Pn whose entries are all non-zero for some integer n. (then as n ?)

There are non-regular MCs; how can we check?

16

Classification of StatesAccessible: Possible to go from state i to state j (path exists from i to

j).Let ai and di be the events of arriving and departing of a customer to a

system in state i2 3 4 …10

d4d1 d2 d3

a0a1 a2 a3

2 3 4 …a a1 a0 a0 1 2 3

Two states communicate if both are accessible from each other. A system is irreducible if all states communicate.

State i is recurrent if the system will return to it after leaving some time in the future.

If a state is not recurrent, it is transient.

Classification of States (cont’)A state is periodic if it can only return to itself after a fixed number of transitions greater than 1 (or multiple of a fixed number).

A state that is not periodic is aperiodic.

2

0

1

(1) (1)

(1)

Each state is visited every 3 iterations

2 3 4 …10

d4d1 d2 d3

a0a1 a2 a3

How about this?

Classification of States (cont’)

(1)

2

0

1

(1)(0.5)

(1)

4(0.5)

Each state is visited in multiples of 3 iterations

The period of a state i is the smallest k > 1 such that all paths leading back to i have lengths that are a multiple of k;

i.e., pii(n) = 0 unless n = k, 2k, 3k, . . .

If the smallest k, gcd of all path lengths, is 1, then state i is aperiodic.Periodicity is a class property; all the states in a class have the same period.

Classification of States (cont’)An absorbing/trapping state is one that locks in the system once it enters.

2 3 4a a

1a

0

d d d1 2 3

1 2 3

This diagram might represent the wealth of a gambler who begins with $2, makes a series of wagers for $1 each, and stops when his money becomes $4 or $0.

Let ai be the event of winning in state i and di the event of losing in state i.

There are two absorbing states: 0 and 4.

An absorbing state is a state j with pjj = 1.

Classification of States (cont’)Class: set of states that communicate with each other.

A chain is irreducible if there is only one class

A class is either all recurrent or all transient and may be all periodic or aperiodic.

States in a transient class communicate only with each other so no arcs enter any of the corresponding nodes from outside the class. Arcs may leave, though, passing from a node in the class to one outside.

2

0

1 5

3

4

6

Illustration of Concepts

3 1

0

2

0 0 X 0 X

1 X 0 0 0

2 X 0 0 0

3 0 0 X X

0 1 2 3

State

Example 1

Every pair of states that communicates forms a single recurrent class; however, the states are not periodic.

Thus the stochastic process is aperiodic and irreducible.

* X is a prob. and 0<X1

Illustration of Concepts

Example 2

4

0

0 X X 0 0 X

1 X X 0 0 0

2 0 0 X X 0

3 0 0 0 X 0

0 1 2 3 4

State 4 0 0 0 0 0

23

1

States 0 and 1 communicate and for a recurrent class.

States 3 and 4 form separate transient classes.

State 2 is an absorbing state and forms a recurrent class.

* X is a prob. and 0<X1 * especially, p2,2 is 1

Illustration of ConceptsExample 3

3 1

0

2

0 0 0 0 X

1 X 0 0 0

2 X 0 0 0

3 0 X X 0

0 1 2 3

State

Every state communicates with every other state, so we have irreducible stochastic process.

Periodic?

Yes, so this MC is irreducible and periodic.

* X is a prob. and 0<X1

Classification of States

2.08.00001.04.05.000

07.03.0000005.05.00006.04.0

54321

P

.5.4

.6

.5.3 .5

.4

.8

.7

.1

1

5

24

.2

Example 4* Sometimes states start with 1, not 0

A state j is accessible from state i if pij(n) > 0 for some n > 0.

In Example 4, state 2 is accessible from state 1 states 3 and 4 are accessible from state 5but states 3 is not accessible from state 2.

States i and j communicate if i is accessible from j and j is accessible from i.

States 1 & 2 communicate; states 3, 4 & 5 communicate;states 1,2 and states 3,4,5 do not communicate.

States 1 & 2 form one communicating class.States 3, 4 & 5 form another communicating class.

Example 4 Review

If all states in an MC communicate, (i.e., all states are in the same class) then the chain is irreducible.

Example 4 is not an irreducible MC. Gambler’s example has 3 classes: {0}, {1, 2, 3} and {4}.Example 3?

Recurrence Properties (solve Ex 11.19)Let fi = prob. that the process will return to state i (eventually) given that it starts in state i.

If fi = 1 then state i is called recurrent.

If fi < 1 then state i is called transient.

.)n(pn

ii

1

.)n(pn

ii

1

Recurrence properties

11.3.3 limiting probabilities

28

Transient vs. recurrentIf a Markov chain has transient classes and recurrent classes, the system will eventually enter and remain in one of recurrent classes (Fig. 11.6(a)). Thus, we can assume only irreducible Markov chains

Suppose a system starts with a recurrent state i at time 0.

Let Ti(k) be the time that elapses between the (k-1)-th and k-th returns.

29

Fig 11.8: recurrence times Ti(k)The process will return to state i at Ti(1), Ti(1)+Ti(2), Ti(1)+Ti(2)+Ti(3),…

30

Steady state prob. vs. recurrence timeProportion of time in state i after k returns to i is

31

Ti’s form an iid sequence since each recurrence time is independent of previous ones.

As the state is recurrent, the process returns to state i an infinite number of times. Then the law of large numbers implies that

Proportion of time in state i

Recurrent timesIf E[Ti] < , then state i is positive recurrent, which

implies that i > 0

32

If E[Ti] = , then state i is null recurrent, which implies that i = 0

But how can we obtain E[Ti] ? )n(nf]T[En

iii

1

Solve Ex 11.26

}iX,iX,,iX|iX{P)n(f nnii 011

n

iii )n(ff If fi is 1, recurrent; if fi <1, transient

Question 11-2Consider an MC with infinite states, where p0,1 = 1,

and for all state i1, we have

33

Check whether state 0 is recurrent or transient.If recurrent, check whether positive or null.

Existence of Steady-State ProbabilitiesA state is ergodic if it is aperiodic and positive recurrent

Once an MC enters an ergodic state, then the process will remain in the state’s class forever.

Also, the process will visit all states in the class sufficiently frequently, so that the long term proportion of time in a state will be non zero.

We mostly deal with MCs, each of which has a single class that consists of ergodic states only.

Thus we can apply = P

Regular vs. ergodic MC

Regular MCs have Pn whose entries are all non-zero for some integer n.

Ergodic MCs have aperiodic, positive recurrent states.

They are almost same, practically.

35

Not regular, but ergodic MC

36

0110

P