G12: Management Science Markov Chains. Outline Classification of stochastic processes Markov...

Post on 18-Dec-2015

231 views 0 download

Tags:

transcript

G12: Management Science

Markov Chains

Outline

• Classification of stochastic processes• Markov processes and Markov chains

• Transition probabilities

• Transition networks and classes of states

• First passage time probabilities and expected first passage time

• Long-term behaviour and steady state distribution

Analysing Uncertainty• Computer Models of Uncertainty:

– Building blocks: Random number generators– Simulation Models

• Static (product launch example)• Dynamic (inventory example and queuing models)

• Mathematical Models of Uncertainty: – Building blocks: Random Variables– Mathematical Models

• Static: Functions of Random Variables • Dynamic: Stochastic (Random) Processes

Stochastic Processes• Collection of random variables Xt, t in T

– Xt’s are typically statistically dependent

• State space: set of possible values of Xt’s– State space is the same for all Xt’s – Discrete space: Xt’s are discrete RVs– Continuous space: Xt’s are continuous RVs

• Time domain: – Discrete time: T={0,1,2,3,…}– Continuous time: T is an interval (possibly unbounded)

Examples from Queuing Theory• Discrete time, discrete space

– Ln: queue length upon arrival of nth customer

• Discrete time, continuous space– Wn: waiting time of nth customer

• Continuous time, discrete space– Lt: queue length at time t

• Continuous time, continuous space– Wt: waiting time for a customer arriving at time t

A gambling example

• Game: Flip a coin. You win £ 10 if coin shows head and loose £ 10 otherwise

• You start with £ 10 and you keep playing until you are broke

• Typical questions– What is the expected amount of money after t

flips?– What is the expected length of the game?

£10

£ 0

£20

10

£30

0.5

0.5

0.5

0.5

0.5

0.5

0.5

0.5

A Branching Process

….

….

….

Discrete Time - Discrete State Stochastic Processes

• Xt: Amount of money you own after t flips• Stochastic Process: X1,X2,X3,…• Each Xt has its own probability distribution• The RVs are dependent: the probability of

having £ k after t flips depends on what you had after t’ (<t) flips – Knowing Xt’ changes the probability distribution

of Xt (conditional probability)

Outline• Classification of stochastic processes

• Markov processes and Markov chains• Transition probabilities

• Transition networks and classes of states

• First passage time probabilities and expected first passage time

• Long-term behaviour and steady state distribution

Markovian Property• Waiting time at time t depends on waiting time at times t’<t

– Knowing waiting time at some time t’<t changes the probability distribution of waiting time at time t (Conditional probability)

– Knowledge of history generally improves probability distribution (smaller variance)

• Generally: The distribution of states at time t depends on the whole history of the process – Knowing states of the system at times t1,…tn<t changes the distribution of states at

time t

• Markov property: The distribution of states at time t, given the states at times t1<…<tn<t is the same as the distribution of states at time t, given only knowledge of the state at time tn. – The distribution depends only on the last observed state– Knowledge about earlier states does not improve probability distribution

Discrete time, discrete space

• P(Xt+1= j | X0=i0,…,Xt=it) = P(Xt+1= j | Xt=it)

• In words: The probabilities that govern a transition from state i at time t to state j at time t+1 only depend on the state i at time t and not on the states the process was in before time t

Transition Probabilities• The transition probabilites are

P(Xt+1= j | Xt=i)

• Transition probabilities are called stationary if

P(Xt+1= j | Xt=i) = P(X1= j | X0=i)

• If there are only finitely many possible states of the RVs Xt then the stationary transition probabilities are conveniently stored in a transition matrix

Pij= P(X1= j | X0=i)

• Find the transition matrix for our first example if the game ends if the gambler is either broke or has earned £ 30

Markov Chains• Stochastic process with a finite number, say n,

possible states that has the Markov property• Transitions between states in discrete time steps• MC is completely characterised by transition

probabilities Pij from state i to state j are stored in an n x n transition matrix P– Rows of transition matrix sum up to 1. Such a matrix

is called a stochastic matrix• Initial distribution of states is given by an initial

probability vector p(0)=(p1(0),…,pn

(0))• We are interested in the change of the probability

distribution of the states over time

Markov Chains as Modelling Templates

• Lawn mower example:– Weekly demand D for lawn mowers has distribution

P(D=0)=1/3, P(D=1)=1/2, P(D=2)=1/6

– Mowers can be ordered at the end of each week and are delivered right at the beginning of the next week

– Inventory policy: Order two new mowers if stock is empty at the end of the week

– Currently (beginning of week 0) there are two lawn mowers in stock

• Determine the transition matrix

Market Shares• Two software packages, B and C, enter a market

that has so far been dominated by software A• C is more powerful than B which is more

powerful than A• C is a big departure from A, while B has some

elements in common with both A and C• Market research shows that about 65% of A-users

are satisfied with the product and won’t change over the next three months

• 30% of A-users are willing to move to B, 5% are willing to move to C….

Transition Matrix• All transition probabilities over the next three months

can be found in the following transition matrix

• What are the approximate market shares going to be?

ToFrom

A B C

A 65% 30% 5%

B 10% 75% 15%

C 0% 10% 90%

Machine Replacement• Many identical machines are used in a

manufacturing environment

• They deteriorate over time with the following monthly transition probabilities:

ToFrom

New OK Worn Fail

New 0 0.9 0.1 0

OK 0 0.6 0.3 0.1

Worn 0 0 0.6 0.4

Fail 1 0 0 0

Outline• Classification of stochastic processes

• Markov processes and Markov chains

• Transition probabilities• Transition networks and classes of states

• First passage time probabilities and expected first passage time

• Long-term behaviour and steady state distribution

2-step transition probability (graphically)

i

0

1

2

j

Pi0

Pi1

Pi2

P2j

P1j

P0j

2-step transition probabilities (formally)

kikkj

k

k

k

ij

PP

iXkXPkXjXP

iXkXPkXjXP

iXkXPkXiXjXP

iXjXPP

)|()|(

)|()|(

)|(),|(

)|(

0101

0112

01102

02)2(

Chapman-Kolmogorov Equations

• Similarly, one shows that n-step transition probabilities Pij

(n)=P(Xn=j | X0=i) obey the following law (for arbitrary m<n:)

• The n-step transition probability matrix P(n) is the n-th power of the 1-step TPM P:

P(n) =Pn=P…P (n times)

k

mik

mnkj

nij PPP )()(

Example

iesprobabilitn transitiostep-3 theFind

9.01.0

7.03.0

matrix n transitioGiven the

P

see spreadsheet Markov.xls

Distribution of Xn

• Given – Markov chain with m states (1,…,m) and transition matrix P

– Probability vector for initial state (t=0): p(0)=(p1(0),…, pm

(0))

• What is the probability that the process is in state i after n transitions?

• Bayes’ formula:P(Xn=i)=P(Xn=i¦X0=1)p1

(0)+…+P(Xn=i¦X0=m)pm(0)

• Probability vector for Xn: p(n)= p(0)Pn

• Iteratively: p(n+1)= p(n)P • Open spreadsheet Markov.xls for lawn mower, market

share, and machine replacement examples

Outline• Classification of stochastic processes

• Markov processes and Markov chains

• Transition probabilities

• Transition networks and classes of states• First passage time probabilities and expected first passage

time

• Long-term behaviour and steady state distribution

An Alternative Representation of the Machine Replacement Example

New

Fail

OK

Worn

0.9

0.1

0.6

0.30.1

0.60.4

1

The transition network

• The nodes of the network correspond to the states

• There is an arc from node i to node j if Pij > 0 and this arc has an associated value Pij

• State i is accessible from state j if there is a path in the network from node i to node j

• A stochastic matrix is said to be irreducible if each state is accessible from each other state

Classes of States • State i and j communicate if i is accessible

from j and j is accessible from i• Communicating states form classes • A class is called absorbing if it is not possible

to escape from it• A class A is said to be accessible from a class

B if each state in A is accessible from each state in B– Equivalently: …if some state in A is accessible from

some state in B

Find all classes in this example and indicate their accessibility from other classes

1

4

7

2

6

3

5

1/3

2/3

1

1/3

1/2

1/6

1

1/2

1/2

1

1/2

2/3

Return to Gambling Example

• Draw the transition network

• Find all classes

• Is the Markov chain irreducible?

• Indicate the accessibility of the classes

• Is there an absorbing class?

Outline• Classification of stochastic processes

• Markov processes and Markov chains

• Transition probabilities

• Transition networks and classes of states

• First passage time probabilities and expected first passage time

• Long-term behaviour and steady state distribution

First passage times

• The first passage time from state i to state j is the number of transitions until the process hits state j if it starts at state i

• First passage time is a random variable

• Define fij(k) = probability that the first passage from state i to state j occurs after k transitions

Calculating fij(k)

• Use Bayes’ formula

P(A)=P(A|B1)P(B1)+…+ P(A|Bn)P(Bn)

• Event A: starting from sate i the process is in state j after n transitions (P(A)=Pij

(n))

• Event Bk: first passage from i to j happens after k transitions

Calculating fij(k) (cont.)

)1()1()(

)1(

1111

)(

)1(...)1()(

)1(

)()1(...)1(

)()|()()|(...)()|(

)(

jjijn

jjijn

ijij

ijij

ijijjjijn

jj

nnnn

nij

PnfPfPnf

Pf

nfnfPfP

BPBAPBPBAPBPBAP

APP

Bayes’ formula gives:

This results in the recursion formula:

Alternative: Simulation

• Do a number of simulations, starting from state i and stopping when you have reached state j

• Estimate fij(k) = Percentage of runs of length k

• BUT: This may take a long time if you want to do this for all state combinations (i,j) and many k’s

Expected first passage time

• If Xij = time of first passage from i to j then

E(Xij)=fij(1)+2fij(2)+3fij(3)+….

• Use conditional expectation formula

E(Xij)=E(Xij|B1)P(B1)+…+ E(Xij|Bn)P(Bn)

• Event Bk: first transition goes from i to k

• Notice– E(Xij |Bj)=1 and E(Xij|Bk)=1+E(Xkj)

Hence

jkikkj

k jkikkjik

jkikkjij

ikkk

ijij

PXE

PXEP

PXEP

PBXEXE

)(1

)(

))(1(

)|()(

unknowns in equations Fix mmj

Example

4/34/1

3/23/1P

4)(,3/11)(

)(4/31)(

)(3/21)(

2111

2121

2111

XEXE

XEXE

XEXE

Outline• Classification of stochastic processes

• Markov processes and Markov chains

• Transition probabilities

• Transition networks and classes of states

• First passage time probabilities and expected first passage time

• Long-term behaviour and steady state distribution

Long term behaviour• We are interested in distribution of Xn as n tends to

infinity: lim p(n)=lim p(0)P(n)= p(0) lim P(n)

• If lim P(n) exists then P is called

• The limit may not exist, though:

• See Markov.xls• Problem: Process has periodic behaviour

– Process can only recur to state i after t,2t,3t,… steps– There exists t: if n Not in {t,2t,3t} then Pii

(n) = 0

• Period of a state i: maximal such t

01

10P

Find the periods of the states

1

4

7

2

6

3

5

1/3

2/3

1

1/3

1/2

1/6

1

1/2

1/2

1

1/2

2/3

Aperiodicity

• A state with period 1 is called aperiodic

• State i is aperiodic if and only if there exists N such that Pii

(N) > 0 and Pii(N+1) > 0

• The Chapman-Kolmogorov Equations therefore imply that Pii

(n)>0 for every n>=N

• Aperiodicity is a class property, i.e. if one state in a class is aperiodic, then so are all others

Regular matrices

• A stochastic matrix P is called regular if there exists a number n such that all entries of Pn are positive

• A Markov chain with a regular transition matrix is aperiodic (i.e. all states are aperiodic) and irreducible (i.e. all states communicate)

Back to long-term behaviour

• Mathematical Fact: If a Markov chain is irreducible and aperiodic then it is ergodic, i.e., all limits

exist

)(lim nij

nij PP

Finding the long term probabilities

• Mathematical Result: If a Markov chain is irreducible and aperiodic then all rows of its long term transition probability matrix are identical to the unique solution =(1,…, m) of the equations

11

m

ii

P

P

However,...

• …the latter system is of the form P=, 1+…+m=1 and has m+1 equations and m unknowns– It has a solution because P is a stochastic matrix and

therefore has 1 as an eigenvalue (with eigenvector x=(1,…,1)). Hence is just a left eigenvector of P to the eigenvalue 1 and the additional equation normalizes the eigenvector

• Calculation: solve the system without the first equation - then check first equation

Example

• Find the steady state probabilities for

• Solution: (1,2)=(0.6,0.4)

2/12/1

3/13/2P

Steady state probabilities

• The probability vector with P= and 1+..+m=1 is called the steady state (or stationary) probability distribution of the Markov chain

• A Markov chain does not necessarily have a steady state distribution

• Mathematical result: an irreducible Markov chain has a steady state distribution

Tending towards steady state

• If we start with the steady state distribution then the probability distribution of the states does not change over time

• More importantly: If the Markov chain is irreducible and aperiodic then, independently of the initial distribution, the distribution of states gets closer and closer to the steady state distribution

• Illustration: see spreadsheet Markov.xls

More on steady state distributions

• j can be interpreted as the long-run proportion of time the process is in state j

• Alternatively: j=1/E(Xjj) where Xjj is the time of the first recurrence to j– E.g. if the expected recurrence time to state j

is 2 transitions then, on the long run, the process will be in state j after every 1 out of two transitions,i.e. 1/2 of the time

Average Payoff Per Unit Time

• Setting: If process hits state i, a payoff of g(i) is realized (costs = negative payoffs)

• Average payoff per period after n transitions

Yn=(g(X1)+…+g(Xn))/n

• Long-run expected average payoff per time period: lim E(Yn) as n tends to infinity

Calculating long-run average pay-offs

Mathematical Fact: If a Markov chain is irreducible and aperiodic then

)()(lim1

jgYEm

jjn

n

Example

2/12/1

3/13/2P

A transition takes place every week. A weekly cost of £ 1 has to be payed if the process is in state 1, while a weekly profit of £ 1 is obtained if the process is in state 1. Find the average payoff per week. (Solution: £ -0.2 per week)

Key Learning Points• Markov chains are a template for the analysis of

systems with finitely many states where random transitions between states happen at discrete points in time

• We have seen how to calculate n-step transition probabilities, first passage time probabilities and expected first passage times

• We have discussed steady state behaviour of a Markov chain and how to calculate steady state distributions