0% found this document useful (0 votes)
140 views

Continuous Time Markov Chains

The document defines and discusses continuous-time Markov chains (CTMC). It begins by defining a CTMC as a stochastic process with discrete states and time-dependent transition probabilities between states. It provides an example of a Poisson process and discusses the exponential distribution of waiting times in each state. The document then covers the Markov chain of state transitions, birth-death processes, and the transition probability function Pij(t). It concludes by defining the instantaneous transition rates qij in terms of the state transition rates ν(i) and transition probabilities Pij.

Uploaded by

pouty567
Copyright
© Attribution Non-Commercial (BY-NC)
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
140 views

Continuous Time Markov Chains

The document defines and discusses continuous-time Markov chains (CTMC). It begins by defining a CTMC as a stochastic process with discrete states and time-dependent transition probabilities between states. It provides an example of a Poisson process and discusses the exponential distribution of waiting times in each state. The document then covers the Markov chain of state transitions, birth-death processes, and the transition probability function Pij(t). It concludes by defining the instantaneous transition rates qij in terms of the state transition rates ν(i) and transition probabilities Pij.

Uploaded by

pouty567
Copyright
© Attribution Non-Commercial (BY-NC)
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 20

Continuous time Markov chains

01KRR - STOCHASTIC PROCESSES


October 2010
Continuous time Markov chains (CTMC): denition

Let X(t), t 0 be a continuous-time stochastic process taking on


values i , j , . . . in a discrete (nite or countable) state space.

We assume step-wise constant and right-continuous trajectories.

We say that X is a continuous-time Markov chain if for all times


s < t + s, states i , j and trajectory x(u), 0 u < s, we have
P{X(t + s) = j |X(s) = i , X(u) = x(u), 0 u < s }
= P{X(t + s) = j |X(s) = i } = P
ij
(t)

As in the case of discrete time, the rst equality is the Markov


property, the second is the stationarity of the transition
probabilities.

Example: Poisson process P


i ,j
(t) = e
t
(t)
j i
(j i )!
i j
Waiting times
Let X(t), t 0, be a CTMC.
Theorem

Given X(s) = i , let R = T s be the residual time in i .

Therefore,
P{R > t + u|R > t, X(s) = i } = P{X(s+t+u) = i |X(s+t) = i , X(s) = i }
= P{X(s + t + u) = i |X(s + t) = i } = P{X(s + u) = i |X(s) = i }
= P{R > u|X(s) = i }

The conditional distribution is exponential:


P{R > t|N(s) = i } = e

i
t
.
In the case of the Poisson process the intensity is constant
i
= .
MC of the transitions
Let T
n
denote the time of the n-th change of state and T
0
= 0.
Theorem
The process X
n
= X
T
n
, n = 0, 1, . . . is a Markov chain, with transition
matrix P such that P
ii
= 0.
Poisson In the case of the Poisson process, transition is always i i + 1
with probability 1.
Exercise Let N(t), t > 0, be a Poisson process. Dene X(t) = N(t) mod 2.
Therefore,
P{X(t) = 0} =

k=1
e
t
(t)
2k
(2k)!
= e
t
cosh(t) =
1
2
_
1 + e
2t
_
P{X(t) = 1} =

k=0
e
t
(t)
2k+1
(2k + 1)!
= e
t
sinh(t) =
1
2
_
1 e
2t
_
Continuous-time MC: second denition
Theorem
Given exponential independent times with state-dependent intensity
(rate) (i ), and a transition matrix P, with zero diagonal, P
ii
= 0, the
process obtained by waiting in each state for the given exponential time,
them moving to a new state according to the transition, is a
continuous-time MC.
Simulation of a CTMC
The following algorithm generates a continuous-time MC. Let s = 0, then
Step 1 Generate a value x from the distribution of X
s
Step 2 Generate a value t from Exp(1); the trajectory is x(u) = x
for 0 u <
t
(x)
; s =
t
(x)
Step 3 Generate a new value for x from the x-row of P.
Ross Example 6.5
Consider a shoe-shine shop consisting of two chairs (1 and 2).
Assumptions are:
1. Potential customers arrive according a Poisson process with
intensity .
2. Service starts in chair 1 if both chairs are free.
3. After an Exp(
1
) service time, the customer moves to chair 2.
4. The service time in chair 2 is Exp(
2
).
Then, the state space is 0, 1, 2, where 0 means that no chair is occupied,
1 o 2 means that that chair is occupied. Transitions are deterministic.
The exit times from 1 and 2 are exponential with rate
1
,
2
.
_
_
0 1 2
0 0 1 0
1 0 0 1
2 1 0 0
_
_
(1) =
1
(2) =
2
Discussion

The exit rate from 0 requires a discussion to show it is equal to .

Let T
2
an exit time from state 2. After that, the system goes to
state 0 and can accept a new client. Note that T
2
is independent
from any inter-arrival time of the Poisson process posterior to the
last client accepted.

If S is the inter-arrival time running at time T


2
, then the condition
is satised and S and T
2
are independent.

The probability for the exit time from 0 to be greater then t, given
that the system is in state i is equal to
P{S > T
2
+ t|S > T
2
} = P{S > t} = e
t
therefore (0) =
A complement on the Poisson process
A continuous time MC can be described either by is continuous time
transitions P
ij
(t) = P{X(t) = j |X(0) = i }, or by its transition rates (i ),
together with the discrete time transitions P
ij
. The connection between
the two descriptions rests on the following property of the Poisson
process.
Let N(t), t 0, be a Poisson process with rate . Then
1. P{N(h) = 1} = h + o(h) as h 0.
2. P{N(h) 2} = o(h) as h 0.
This follows directly from the explicit formul
p{N(h) = 1} = e
h
h
1!
P{N(h) 2} = e
h

k2
(h)
k
k!
by Calculus.
Therefore
P
ij
(h) = P
ij

i
h + o(h) h 0 j = i
Birth and death process

A birth and death process is a continuous time MC X(t), t 0,


with state space Z
+
= 0, 1, . . . , which is described through two
independent Poisson processes describing the process of birth
(arrival) and the process of death (departures).

Given the process is in state n > 0 at time t, {X(t) = n}, the


process waits for the rst of two independent exponential RV B and
D of respective rate
i
and
i
. If B comes rst, then the system
moves to state n + 1. If D comes rst, then the system moves to
state n 1. If the initial value is n = 0, only a birth can occur.

min(B, D) is Exp(
i
+
i
); P{B < D} =
i
/(
i
+
i
).

State transition rates are


0
=
0
,
i
=
i
+
i
, i > 0.

Transition probabilities are P


01
= 1, P
i ,i +1
=

i

i
+
i
,
P
i ,i 1
=

i

i
+
i
, i > 0
Examples of B&D process
Yule
n
= n,
n
= 0.
Linear growth with immigration
n
= n + , = n. is the
immigration rate. Let M(t) = E[X(t)|X(0) = i ] the expected population
at time t, given X(0) = i . By computing the transition probability with
small h, we can nd the dierential equation
M

(t) = ( )M(t) +
whose solution for = is
M(t) =

mu
_
e
()t
1
_
+ ie
(mu)t
Note the behavior when t .
M/M/1 Customers arrive as a Poisson process with rate . They enter
the queue and are served with the priority rule 1st arrived 1st served. The
process of services is a Poisson process with rate . The length of the
queue is a B&D process with constant rates.
M/M/s If there are s servers, then the birth rate is constant, while the
death rare is
n
= n for 1 n s,
n
= s, n > s.
Expected transition times in B&D process
A possible question related to a birth and death is the time T
i
needed
to move from state i to state i + 1 (not necessarily equal to 1), or its
expected value.
Obviously, E[T
0
] =
1

0
, while for others it can be shown that
E[T
i
] =
1

i
+

i

E[T
i 1
]
so that E[T
i
] may be calculated recursively.
In the particular case
i
= and
i
= , = , we get
E[T
0
] =
1

0
E[T
n
] =
1 (/)
n+1
1 (/)
, n 1.
In the case = then E[T
n
] =
n+1

.
Transition probability function P
ij
(t)
Let X(t), t 0 be a continuous-time MC.
Denition
P
ij
(t) The transition probability function is dened to be
P
ij
(t) = P{X(s + t) = j |X(s) = i }
For the Yule process (pure birth process) with distinct birth rates
i
, let
X
i
be the time spent in state i before moving to state i + 1. The time
spent before entering state j is

j 1
k=1
X
k
. Then the events
{X(t) < j , X(0) = i }, {X
i
+ + X
j 1
> t, X(0) = i }
are equal. Then, for i < j ,
P{X(t) < j |X(0) = i } = P{X
i
+ + X
j 1
> t|X(0) = i }
=
j 1

k=i
e

k
t
j 1

r =k,r =i

r

k
Proof is a (long) exercise.
Instantaneous transition rates
Let X(t), t 0 be a continuous-time MC with state transition rates
(i ) and state transition probabilities P
ij
.
Denition: For any couple of states i , j we dene the instantaneous
transition rates
q
ij
= (i )P
ij
Viceversa:

j
q
ij
= (i ) and
q
ij

j
q
ij
=
q
ij
(i )
= P
ij
.
Q and P(t): 1 P
ii
(h) is the probability of no transition or at least two
transitions. P
ij
(t) is the probability of one transition from
i to j or at least two transitions. Then, for h 0,
1 P
ii
(h) = (i )h + o(h) lim
h0
1 P
ii
(h)
h
= (i )
P
ij
(h) = (i )P
ij
h + o(h) lim
h0
P
ij
(h)
h
= q
ij
Champan-Kolmogorov
Exactly as in the discrete-time case, we derive the CK Equations:
P
ij
(t + s) = P{X(t + s) = j |X(0) = i }
=

k
P{X(t + s) = j , X(t) = k|X(0) = i }
=

k
P{X(t + s) = j |X(t) = k, X(0) = i }P{X(t) = k|X(0) = i }
=

k
P{X(t + s) = j |X(t) = k}P{X(t) = k|X(0) = i }
=

k
P{X(s) = j |X(0) = k}P{X(t) = k|X(0) = i }
=

k
P
ik
(t)P
kj
(s)
Kolmogorovs backward equations
Using the previous denitions and results,
P
ij
(t + h) P
ij
(t) =

k
P
ik
(h)P
kj
(t) P
ij
(t)
=

k=i
P
ik
(h)P
kj
(t) [1 P
ii
(h)]P
ij
(t)
=

k=i
P
ik
(h)P
kj
(t) [1 P
ii
(h)]P
ij
(t)
=

k=i
q
ik
hP
kj
(t) (i )hP
ij
(t) + o(h)
then, we have the Kolmogorovs Backward Equations
P

ij
(t) =

k=i
q
ik
P
kj
(t) (i )P
ij
(t)
It is a system of linear dierential equations. Note the matrix product QP
Kolmogorovs forward equations
By taking the product in the CKEs dierently, we get
P
ij
(t + h) P
ij
(t) =

k
P
ik
(t)P
kj
(h) P
ij
(t)
=

k=j
P
ik
(t)P
kj
(h) [1 P
jj
(h)]P
ij
(t)
=

k=j
P
ik
q
kj
(t)h (j )hP
ij
(t) + o(h)
then, we obtain the Kolmogorovs Forward Equations:
P

ij
(t) =

k=j
P
ik
(t)q
kj
(j )P
ij
(t)
Note that here the product is PQ. The (in-formal) derivation above fails
in some cases of countable state space, while the KBE are always
satised.
Example: For birth and death processes, Kolmogorovs forward
equations are:
P

i 0
(t) =
1
P
i 1
(t)
0
P
i 0
(t),
P

ij
(t) =
j 1
P
i ,j 1
(t)
j
P
ij
(t), j = 0.
Balance equations
Remember: for MCs, under suitable assumptions, P
j
= lim
n
P
n
ij
is the probability of being in state j for n . Similarly, it can be
proved that:
P
j
= lim
t
P
ij
(t) j S.
Letting t in the Kolmogorov forward equations:
lim
t
P

ij
(t) = lim
t
_

k=j
q
kj
P
ik
(t)
j
P
ij
(t)

0 =
_

k=j
q
kj
P
k

j
P
j

Thus, probabilities P
j
may be found by means of

j S
P
j
= 1 and

k=j
q
kj
P
k
=
j
P
j
j S.
Example: for a M/M/1 queue, with rates and , if < then
P
n
=
(

)
n
1 +

m
(

)
m
= (

)
n
(1

), n 0.
The are no limiting values of P
n
for since in this case the
states are all transient states or null recurrent states.
Erlangs Loss Formula
Consider an M/M/k queue where a client that nd the server occupied
moves out (loss system). Let the arrival rate, and the service rate.
The balance equations are:
P
0
= P
1
........
P
k1
= kP
k
.
Adding the condition

j S
P
j
= 1, one gets:
P
i
=
(

)
i
/i !

k
j =0
(

)
j
/j !
, i = 0, 1, . . . , k.
This formula can be generalized to a M/G/k queueing loss model (see
Ross, Theorem 5.7.4)

You might also like