Dokumen - Pub - Probability Amp Random Processes For Engineers Solution Manual 1nbsped 9389976413 9789389976410
Dokumen - Pub - Probability Amp Random Processes For Engineers Solution Manual 1nbsped 9389976413 9789389976410
978-93-89976-41-0
9 789389 976410
TM
PROBABILITY AND
RANDOM
PROCESSES FOR
ENGINEERS
Dr. J. Ravichandran
Professor
Department of Mathematics
Amrita Vishwa Vidyapeetham
Coimbatore, India
©Copyright 2020 I.K. International Pvt. Ltd., New Delhi-110002.
This book may not be duplicated in any way without the express written consent of the publisher,
except in the form of brief excerpts or quotations for the purposes of review. The information
contained herein is for the personal use of the reader and may not be incorporated in any commercial
programs, other books, databases, or any kind of software without written consent of the publisher.
Making copies of this book or any portion for any purpose other than your own is a violation of
copyright laws.
Limits of Liability/disclaimer of Warranty: The author and publisher have used their best efforts in
preparing this book. The author make no representation or warranties with respect to the accuracy or
completeness of the contents of this book, and specifically disclaim any implied warranties of
merchantability or fitness of any particular purpose. There are no warranties which extend beyond the
descriptions contained in this paragraph. No warranty may be created or extended by sales
representatives or written sales materials. The accuracy and completeness of the information provided
herein and the opinions stated herein are not guaranteed or warranted to produce any particulars
results, and the advice and strategies contained herein may not be suitable for every individual.
Neither Dreamtech Press nor author shall be liable for any loss of profit or any other commercial
damages, including but not limited to special, incidental, consequential, or other damages.
Trademarks: All brand names and product names used in this book are trademarks, registered
trademarks, or trade names of their respective holders. Dreamtech Press is not associated with any
product or vendor mentioned in this book.
ISBN: 978-93-89976-41-0
EISBN: 978-93-90078-54-7
Preface
Probability and Random Processes is one of the most important courses being
offered in engineering colleges. Particularly, more attention is paid by those who
work with signals, random walks, and Markov chains. For aspiring engineering
students, whether at graduate level or at postgraduate level, when they take up
a project work or research related to signals, image processing, etc., knowledge
of random processes is very useful. Looking at the existing texts on Probability
and Random Processes, there are no texts that are well-structured ones and as a
result it is difficult to understand the basics not to mention about the higher level
concepts.
Therefore, in order to help teachers for the best of their teaching and students
for the best of their understanding on the subject Probability and Random Pro-
cesses, first of all I have made attempts to put the contents in well-structured chap-
ters. It is well known that for understanding the concepts of random processes
from the point of view of teachers and students lies with the understanding of the
concepts of probability and statistics since the concepts of random processes are
built upon the concepts of probability and statistics. Hence, one full chapter is fully
dedicated for the topics on probability and statistics.
Probability and Random Processes for Engineers caters to the needs of the
engineering students both at graduate and postgraduate levels. The text contains
nine chapters that are well organized and presented in an order as the contents
progress from one topic in one chapter to another topic in the proceeding chapter.
In addition, there are appendixes that help in knowing some of the derivations
for the results used in the text. Clearly, the book is user-friendly, as it explains
the concepts with suitable examples and graphical representations before solving
problems. I am of the opinion that this book will be of much value to the faculty in
developing or fine-tuning a good syllabus on Probability and Random Processes
and also in teaching the subject.
This book will have a tinge of the author’s expertise. The author has been
teaching this subject for many years at both undergraduate and postgraduate levels
and, therefore, this book has been written taking into account the needs of teaching
faculty and students. Where appropriate, examples with graphical representations
that are engineering in nature are given to illustrate the concepts. A number of
problems have been solved and exercise problems are given with answers. Putting
it in simple terms, this book is written in such a way that it will stimulate the
interest of students in learning of this subject and also in preparing for their exam-
inations.
As an author I always expect from both students and faculty their critical eval-
uations and suggestions by which the book can be further improved with more
v
vi • Preface
Dr. J. Ravichandran
Contents
Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . v
List of Acronyms . . . . . . . . . . . . . . . . . . . . . . . . . . xii
1. An Overview of Random Variables and Probability Distributions 1
1.0 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1 Probability . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1.1 Classical Approach . . . . . . . . . . . . . . . . . . 1
1.1.2 Statistical Approach . . . . . . . . . . . . . . . . . 2
1.1.3 Axiomatic Approach . . . . . . . . . . . . . . . . . 2
1.1.4 Some Important Results . . . . . . . . . . . . . . . 2
1.1.5 Bayes Theorem . . . . . . . . . . . . . . . . . . . . 3
1.2 One-Dimensional Random Variable . . . . . . . . . . . . . 3
1.2.1 Discrete Random Variable . . . . . . . . . . . . . . 4
1.2.1.1 Discrete Probability Distribution: Probability
Mass Function (PMF) . . . . . . . . . . . . . . . . . . . 5
1.2.2 Continuous Random Variable . . . . . . . . . . . . . 6
1.2.2.1 Continuous Probability Distribution:
Probability Density Function (PDF) . . . . . . . . . . . . 6
1.2.3 Cumulative Distribution Function (CDF) . . . . . . 8
1.3 Expectation (Average or Mean) . . . . . . . . . . . . . . . . 8
1.3.1 Definition and Important Properties of Expectation . 8
1.3.2 Moments and Moment Generating Function . . . . . 10
1.3.3 Characteristic Function . . . . . . . . . . . . . . . . 11
1.4 Special Distribution Functions . . . . . . . . . . . . . . . . 11
1.4.1 Binomial Random Variable and Its Distribution . . . 11
1.4.1.1 Derivation of Mean and Variance using Moment
Generating Function . . . . . . . . . . . . . . . . . . . . 12
1.4.2 Poisson Random Variable and Its Distribution . . . . 13
1.4.2.1 Derivation of Mean and Variance using Moment
Generating Function . . . . . . . . . . . . . . . . . . . . 14
1.4.3 Uniform Random Variable and Its Distribution . . . 15
vii
viii • Contents
xiii
C HAPTER 1
AN OVERVIEW OF RANDOM VARIABLES
AND PROBABILITY DISTRIBUTIONS
1.0 INTRODUCTION
Probability, random variables and probability distributions play a major role in
modeling of random processes in which the outputs observed over a period of time
are quite random in nature. As a matter of fact, one can easily understand the con-
cept of random process, if the basics of probability, random variables and distribu-
tions are properly understood. As probability goes in tune with random variables
and probability distributions, in this chapter we concentrate more on defining the
concepts such as random variables and different types of distributions. In fact, in
this chapter we present only the essentials that are required for understanding the
concept of random process. Similar to that of random variables, in random process
also, each outcome is associated with a probability of its happening. Such prob-
abilities may be according to some probability distribution as well. For example,
an outcome of a random process may take a form according to the outcomes of
tossing a coin, throwing a dice, etc. Or the outcomes of a random process may be
according to a uniform distribution, normal distribution, etc. Also, in this chap-
ter, the most required concepts such as expectation, covariance, correlation and
multivariable distribution function are considered.
1.1 PROBABILITY
Probability is defined as a measure of degree of uncertainty, that is, a measure of
happening or not happening of an event in a trial of a random experiment. Proba-
bility can be determined using three different approaches:
Suppose an event A can occur in ‘a’ ways and cannot occur in ‘b’ ways, then the
probability that event ‘A’ can occur is given as
m a
P= = ,
n a+b
and the probability that it cannot occur is
n−m b
Q= = = 1 − P.
n a+b
Obviously, P and Q are non-negative and cannot exceed unity, that is P + Q = 1
and 0 ≤ P, Q ≤ 1.
If the events A and B are mutually exclusive, then the additive law becomes
(ii) Conditional probability and multiplicative law: For two events A and B,
we have
P (A ∩ B) = P (A) P (B/A) , P (A) > 0
(1.5)
= P (B) P (A/B), P (B) > 0
P (Ei ) P (A/Ei )
P (Ei /A) = , i = 1, · · · · · · , n (1.8)
P (A)
where
n
P (A) = ∑ P (Ei ) P (A/Ei ). (1.9)
i=1
Example 1.1
For example, if two coins are tossed once (or one coin is tossed twice) and if
X is a random variable representing number of heads turning up, then we have the
possible outcomes and the related random variable as
Outcome S ξ1 = HH ξ2 = HT ξ3 = T H ξ4 = T T
Example 1.2
From these examples, one can understand that in the first case (Example 1.1)
the number of heads obtained can be either 0 or 1 or 2, whereas in the second
case (Example 1.2) the temperature at a point of time may be any value within
18 ± 2◦ C.
X =x x1 = 0 x2 = 1 x3 = 2
Here X = x exhausts all possible values 0, 1, 2 and hence the probabilities add
to 1. The probabilities shown are, in fact, the weights assigned to each and every
value assigned by the random variable. Hence, we have
P (X = x1 = 0) = P (X = 0) = 1/4
P (X = x2 = 1) = P (X = 1) = 1/2
P (X = x3 = 2) = P (X = 2) = 1/4
P(x)
1.00
0.75
0.50
0.25
0
0 1 2 x
Definition
The function P (X = x) of the numerical values of the discrete random variable X is
said to be probability mass function (PMF) if it satisfies the following properties:
(i) 0 ≤ P (X = x) ≤ 1
∞
(ii) ∑ P (X = x) = 1
x=−∞
n n
(iii) P ( ∪ Ei ) = ∪ P (Ei ), if E1 , E2 , · · · · · · , En are mutually exclusive (disjoint)
i=1 i=1
events.
One may be of the opinion that since a particular thermometer measures the
temperature as 18◦ C, 19◦ C, 20◦ C, 21◦ C and 22◦ C, the random variable may be
a discrete one. However, there exist measuring equipment (thermometers) that
can measure all possible values such as 19.0001◦ C, 19.0002◦ C, and so on. This
means that the random variable can assign all possible values in the given interval
(16 to 20)◦ C.
That is, given two values a and b such that (a < b), the probability can be
computed as the probability that the random variable X lies between a and b and
is denoted as P (a < X < b) or P (a ≤ X ≤ b). In this regard, we need a function of
the numerical values of the random variable X which could be integrated over the
range (a, b) to get the required probability. Such a function is notationally given as
f (x) and called probability density function (PDF).
Consider the following probability for a more intuitive interpretation of the
density function
Z ε /2
c+
P (c − ε /2 ≤ X ≤ c + ε /2) = = ε f (x)
f (x) dx ∼
c−ε /2
where ε is small. This probability is depicted by the shaded area in Figure 1.2.
f (x)
f (x)
c − ε/2 c c + ε/2 x
Definition
The function f (x), also denoted by fX (x), of the numerical values of the continu-
ous random variable X is said to be probability density function (PDF) if it satisfies
the following properties:
(i) f (x) ≥ 0 for all x ∈ R
Z∞
(ii) f (x)dx = 1
−∞
Zb
(iii) P (a ≤ X ≤ b) = f (x)dx
a
8 • Probability and Random Processes for Engineers
Z∞
E(X) = µx = x f (x) dx if X is continuous random variable
−∞
An Overview of Random Variables and Probability Distributions • 9
Properties
It may be noted that E(b) = b implies that the expected value of a constant
is the same constant only.
(ii) If X is a random variable and h (X) is a function of X, then
∞
E [h(X)] = ∑ h (x) P(X = x) if X is a discrete random variable
x=−∞
Z∞ (1.11)
E [h(X)] = h (x) f (x) dx if X is a continuous random variable
−∞
Therefore,
V (X) = σx2 = E(X 2 ) − {E(X)}2 (1.12)
It may be noted that, the variance V (X) is nothing but the average (mean
or expectation) of the squared differences of each observation from its own
mean value and is always greater than or equal to zero, that is, V (X) ≥ 0.
If X is a random variable (whether discrete or continuous) and if a sample
of n observations is drawn whose mean is E(X), then the variance of X can
be defined as
1 n
V (X) = ∑ [xi − E (X)]2
n i=1
10 • Probability and Random Processes for Engineers
SD (X) = σx = V (X)
p
(1.16)
1.3.2 Moments and Moment Generating Function
Raw moments
It may be noted that E(X r ), r = 1, 2, 3, 4, · · · · · · are known as raw moments (or
moments about origin) of order r. For example, the mean E(X) is the first order
moment and is obtained with r = 1 and E(X 2 ) is the second order is obtained with
r = 2 and so on. The r th order raw moments E(X r ), r = 1, 2, 3, 4, · · · · · · can be
obtained as
d r MX (t)
µr′ = = E (X r )
dt r t=0
Central moments
µr = E[(X − E(X)]r , r = 1, 2, 3, 4, · · · · · · are known as the r th order central
moments (or moments about mean) as the deviations are taken from the mean.
Clearly, the first order moment is given as µ1 = E[(X − E(X)] = 0 and the second
order moment is given as µ2 = E[(X − E(X)]2 , which is the variance of X.
d r ΦX (t)
µr′ = (−i)r = E (X r )
dt r t=0
1 if i th trial is success
where Xi = (1.20)
0 if i th trial is failure
P(x)
0.3
0.2
0.1
0
0 1 2 3 4 5 6 7 8 9 10 x
= p et + q n
d MX (t)
Now, Mean = µ1′ = E (X) =
dt t=0
d MX (t) d n
⇒ = p et + q
dt dt
An Overview of Random Variables and Probability Distributions • 13
n−1 t
= n p et + q pe
d MX (t)
⇒ = n (p + q)n − 1 p
dt t=0
= np (∵ p + q = 1)
2
We know that V (X) = E(X 2 ) − {E(X)}2 = µ2′ − µ1′
d 2 MX (t) d h t n−1 i
Consider µ2′ = = npe p et + q
dt 2 t=0 dt t=0
n−2 t n−1 t
= np [et (n − 1) p et + q p e + p et + q e] t=0
= np [(n − 1) (p + q)n−2 p + 1]
= np [(n − 1) p + 1]
e−λ λ x
P(X = x) = , x = 0, 1, 2, · · · · · · , (1.21)
x!
14 • Probability and Random Processes for Engineers
1 n
where e = lim 1 + , is a commonly used constant in mathematics and is
n→∞ n
approximately equal to 2.7183.
A random variable X that follows Poisson distribution with parameter λ is
usually denoted by: X ∼ P(λ ). An example for Poisson distribution with λ = 1.5
is shown in Figure 1.4.
P(x)
0.3
0.2
0.1
0
0 1 2 3 4 5 x
∞
e−λ λ x
= ∑ etx x!
x=0
∞
(λ et )x
= e−λ ∑ x!
x=0
" #
(λ et )1 (λ et )2
= e−λ 1+ + +......
1! 2!
t t
= e−λ eλ e = eλ (e −1)
d MX (t)
∴ Mean = µ1′ = E (X) =
dt t=0
d MX (t)
⇒ =λ
dt t=0
An Overview of Random Variables and Probability Distributions • 15
2
V (X) = E(X 2 ) − {E(X)}2 = µ2′ − µ1′
Consider
d 2 MX (t)
µ2′ =
dt 2 t=0
d h λ (et −1) +t i
= λe
dt t=0
t
= λ eλ (e −1) +t λ et + 1
h i
t=0
= [λ (λ + 1)]
= λ2 +λ
∴ V (X) = λ 2 + λ − λ 2 = λ .
It is important to note that in case of Poisson distribution, the mean and variance
are same.
1
f (x) = , a<x<b (1.22)
b−a
As shown in Figure 1.5, the random variable X is uniformly distributed over (a, b),
meaning that it puts all its mass on that interval and any point in this interval is
equally likely to occur. By virtue of its appearance as in Figure 1.5, the uniform
distribution is also called a “rectangular distribution”.
f(x)
1
b−a
a b x
Zx
1 x−a
P (X ≤ x) = dx =
b−a b−a
a
Zb
1
µr′ r
= E (x ) = xr dx
b−a
a
r+1 b
1 x br+1 − ar+1
= =
b − a r + 1 a (b − a) (r + 1)
b2 − a2 a+b
Mean = µ1′ = =
2 (b − a) 2
(b − a) b2 + ab + a2 b2 + ab + a2
b3 − a3
µ2′ = = =
3 (b − a) 3 (b − a) 3
b2 + ab + a2 a + b 2 (b − a)2
= − =
3 2 12
reason that many phenomena that occur in nature, industry, and research appear to
have the features of normality. Physical measurements in areas such as tempera-
ture studies, meteorological experiments, rainfall studies and product dimensions
in manufacturing industries are often conveniently explained with a normal dis-
tribution. This is particularly true because, when the size of the sample increases,
almost all distributions can be approximated to a normal distribution. Of course,
the data analyses with normality assumption are always handy.
A random variable X is said to be normally distributed with parameters µ and
σ if its probability density function is given by
2 !
x−µ
1 1
f (x) = √ exp − , −∞ < x < ∞ (1.23)
2πσ 2 σ
f(x)
-∞ ∞
µ-3σ µ-2σ µ-σ µ µ+σ µ+2σ µ+3σ
X −µ
Z=
σ
is normal with mean 0 and variance 1. Such a random variable Z is said to have a
standard normal distribution.
A random variable Z that follows standard normal distribution with mean 0
and variance 1 is usually denoted by: Z ∼ N(0, 1). Accordingly, the probability
density function of the standard normal variate is given as
1
f (z) = √ exp − z 2 /2 , −∞ < z < ∞, (1.24)
2π
X −µ
The result that Z = has a standard normal distribution when X is normal
σ
with mean µ and variance σ 2 is quite useful because it allows us to evaluate all
probabilities concerning X in terms of ϕ . For example, the distribution function of
X can be expressed as
X −µ x−µ
F(x) = P (X ≤ x) = P ≤
σ σ
= P (Z ≤ z) = ϕ (z)
of the values of ϕ (z) (Refer Table given in Appendix C). For example, consider
the following:
f(z)
1- α
α
–∞ ∞
0 zα
MX (t) = E etX
Z∞
= etx f (x) dx
−∞
Z∞
x−µ 2
1 −1 σ
= etx √ e 2 dx
2π σ
−∞
Z∞
1 1 2
MX (t) = √ et (zσ +µ ) e− 2 z σ dz
2π σ
−∞
Z∞
eµ t
− 12 z2 −2t σ z
=√ e dz
2π
−∞
20 • Probability and Random Processes for Engineers
µ t+ σ 2 t 2 /2 Z∞
e 1 2
MX (t) = √ e− 2 (z−σ t) dz
2π
−∞
µ t+ σ 2 t 2 /2 Z∞
e 1 2
MX (t) = √ e− 2 u du
2π
−∞
Z∞
µ t+ σ 2 t 2 /2
e 1 2
= √ 2 e− 2 u du (by even function property)
2π
0
1 dy
Let y = u2 then dy = udu ⇒ du = √ , and now we have
2 2y
µ t+ σ 2 t 2 /2 Z∞
e 1
MX (t) = √ e−y y 2 −1 dy
π
0
Z∞
µ t+ σ 2 t 2 /2 1 1 √
=e ∴ e−y y 2 −1 dy = Γ = π
2
0
d MX (t)
∴ Mean = µ1′ = E (X) =
dt t=0
d MX (t) d µ t+ σ 2t 2 /2
⇒ = MX′ (t) = e
dt dt t=0
µ t + σ 2 t 2 /2
= e µ + σ t t=0 = µ
2
2
We know that V (X) = E(X 2 ) − {E(X)}2 = µ2′ − µ1′
Consider,
d2
µ ′2 = MX (t)
dt 2 t=0
d ′ d µ t+ σ 2 t 2 /2
= MX (t) = e µ +σ t2
dt dt t=0
µ t + σ 2 t 2 /2 µ t + σ 2 t 2 /2
= e σ + µ +σ t e
2 2
µ +σ t
2
t=0
An Overview of Random Variables and Probability Distributions • 21
= σ 2 + µ2
∴ V (X) = σ 2 + µ 2 − µ 2 = σ 2
σ2
P (|X − µ | ≥ k) ≤ (Upper bound for probability)
k2
σ2
or [P (|X − µ |) ≤ k] ≥ 1 − (Lower bound for probability) (1.25)
k2
1
[P (|X − µ |) ≥ kσ ] ≤
k2
1
or [P (|X − µ |) ≤ kσ ] ≥ 1 −
k2
X −µ
Z= √ ∼ N(0, 1) as n → ∞ (1.26)
σ/ n
n
∑ Xi σ2
where X = i=1
. It may be noted that E(X) = µ and V (X) = .
n n
σ12
or simply ρ (X1 , X2 ) = ρ12 =
σ1 σ2
Note:
Two random variables X1 and X2 are said to be independent if P(X1 = x1 , X2 = x2 )
= P(X1 = x1 )P(X2 = x2 ) for discrete case and fX1 X2 (x1 , x2 ) = fX1 (x1 ) fX2 (x2 ) for
continuous case. Also two random variables X1 and X2 are said to be uncorrelated
if C(X1 X2 ) = E(X1 X2 ) − E(X1 )E(X2 ) = 0.
An Overview of Random Variables and Probability Distributions • 23
P (Y = y) = P [w (y)]
Similarly, let X and Y be two discrete random variables with joint probability mass
function P (X = x, Y = y). Let U = g (X, Y ) and V = h (X, Y ) define a one-to-one
transformation between points (x, y) and (u, v) so that the equations u = g (x, y)
and v = h (x, y) may be solved uniquely for x and y in terms of u and v, say
x = w1 (u, v) and y = w2 (u, v), then the joint probability mass function of U and
V is given as
P (U = u, V = v) = P [w1 (u, v), w2 (u, v)] (1.29)
∂x ∂x
∂u ∂v
J=
∂y ∂y
∂u ∂v
24 • Probability and Random Processes for Engineers
or
fUV (u, v) = f [w1 (u, v), w2 (u, v)]/ | J | (1.32)
∂u ∂u
∂x ∂y
J=
∂v ∂v
∂x ∂y
µ1
X1 E(X1 )
X2 E(X2 ) µ2
.. .. ..
. . .
X = Xi , µ = E(Xi ) = µi
.. .. ..
. . .
Xd E(Xd ) µd
2
σ σ12 · · · σ1 j · · · σ1d
1
σ21 σ 2 · · · σ2i · · · σ2d
2
.. .. .. .. .. ..
. . . . . .
and Σ =
2
σi1 σi2 · · · σi · · · σid
. .. .. .. .. ..
.
. . . . . .
2
σd1 σd2 · · · σd j · · · σd
j = 1, 2, · · · · · · , d, i 6= j.
An Overview of Random Variables and Probability Distributions • 25
Cov (Xi , X j ) σi j
Since the correlation coefficient ρi j = p = , we can write
V (Xi ) V (X j ) σi σ j
p
σi j = ρi j σi σ j , i = 1, 2, · · · · · · , d; j = 1, 2, · · · · · · , d, i 6= j. Now, the joint prob-
ability density function can be obtained as
1 (X−µ )T (X−µ )
e− 2
f (x1 , x2 , · · · , xi , · · · , xd ) = , −∞ < xi < ∞, ∀ i (1.33)
(2π )d/2 |Σ|1/2
x1 − µ1
x2 − µ2
..
.
X −µ =
where xi − µi
.
..
.
xd − µd
d d
|Σ| = ∏ σi |Σ|1/2 = ∏ σi
2
⇒ and
i=1 i=1
Also,
2
1/σ1 0 ··· 0 ··· 0
2
1/σ2 · · ·
0 0 ··· 0
.. .. .. .. .. ..
. . . . . .
Σ =
−1
2
0 0 · · · 1/σi · · · 0
.. .. .. .. .. ..
. . . . . .
2
0 0 ··· 0 · · · 1/σd
26 • Probability and Random Processes for Engineers
⇒ (X − µ )T Σ−1 (X − µ ) = (x1 − µ1 x2 − µ2 , . . . , xi − µi , . . . , xd − µd )
2
1/σ1 0 ··· 0 ··· 0 x1 − µ1
0 1/σ 2 · · · 0 · · · 0 x2 − µ2
2
.. .. .. .. .. .. ..
. . . . . . .
xi − µi
2
0 0 · · · 1/σi · · · 0
. .. .. .. .. .. ..
. .
. . . . . .
2 xd − µd
0 0 · · · 0 · · · 1/σd
1 1 1 1
= (x1 − µ1 )2 +(x2 − µ2 )2 2 +· · · · · ·+(xi − µi )2 2 +· · · · · ·+(xd − µd )2 2
σ12 σ2 σi σd
xi − µi
d 2
= ∑ σi
i=1
d x −µ 2
− 21 ∑ i i
σi
e i=1
∴ f (x1 , x2 , · · · , xi , · · · , xd ) = , −∞ < xi < ∞, ∀ i (1.34)
d
(2π )d/2 ∏ σi
i=1
d x − µ 2
− 21 ∑ i i
σ
e i=1
∴ f (x1 , x2 , · · · , xi , · · · , xd ) = , −∞ < xi < ∞, ∀ i (1.35)
(2π )d/2 σ d
2
!
µ1 σ1 σ12
X1
X= , µ= and Σ=
X2 µ2 2
σ21 σ2
An Overview of Random Variables and Probability Distributions • 27
2
!
σ1 ρ12 σ1 σ2 2 σ2σ2,
Clearly, Σ = , |Σ| = 1 − ρ12
2 1 2
ρ12 σ1 σ2 σ2
q
|Σ|1/2 = 1 − ρ12
2 σ σ
1 2 and
2
1 σ2 −ρ12 σ1 σ2
Σ−1 =
1 − ρ12 σ12 σ22
2
2
−ρ12 σ1 σ2 σ1
e− 2 (X−µ )
1 T Σ−1 (X− µ )
f (x1 , x2 ) =
(2π ) |Σ|1/2
( ( 2
!) )
σ2 −ρ12 σ1 σ2 x 1 − µ1
1 1
x 1 − µ1 x 2 − µ2
exp −
1 − ρ12
2 σ 2σ 2
−ρ12 σ1 σ2 σ1
2
x 2 − µ2
2 1 2
= q
2π 1 − ρ12
2 σ σ
1 2
( " #)
x 1 − µ1 x 1 − µ1 x 2 − µ2 x 2 − µ2 2
2
−1
exp − 2ρ12 +
2 1 − ρ12 σ1 σ1 σ2 σ2
2
= q (1.36)
2π σ1 σ2 1 − ρ12
2
If X1 and X2 are independent and hence uncorrelated then ρ12 = 0 and hence we
have
( " 2 #)
x1 − µ1 x2 − µ2
2
1
exp − +
2 σ1 σ2
f (x1 , x2 ) =
2π σ1 σ2
( x −µ 2 ) ( x − µ 2 )
1 −1 1 1
σ1 1 −1 2 2
σ2
f (x1 , x2 ) = √ e 2 √ e 2 ,
2π σ1 2π σ2
− ∞ < x1 , x2 < ∞ (1.37)
which is the product of the density functions of two independent normal random
variables. The graphical representation of the bivariate (two-dimensional) normal
probability density function is shown in Figure 1.7.
28 • Probability and Random Processes for Engineers
f (x1, x2)
x2
x1
∞
−∞
µ1
µ2
−∞ ∞
SOLVED PROBLEMS
Problem 1. A and B are two events such that P (A ∪ B) = 3/4, P (A ∩ B) = 1/4
and P (Ac ) = 2/3. Find P (Ac /B).
S OLUTION :
It is given that
3 1 2
P (A ∪ B) = , P (A ∩ B) = and P (Ac ) =
4 4 3
2 1
⇒ P (A) = 1 − P (Ac ) ⇒ 1− =
3 3
3 1 1 2
∴ P (B) = P (A ∪ B) − P (A) + P (A ∩ B) = − + =
4 3 4 3
We know that
Problem 2. Machine A was put into use 15 years ago and the probability that it
may work for the next 10 years is 0.2. Machine B was put into use eight years ago
and that it may work for the next 10 years is 0.9. The machines being independent,
what is the probability that these two machines can work for the next 10 years?
An Overview of Random Variables and Probability Distributions • 29
S OLUTION :
Probability that M/C A will work for next 10 years: P (A) = 0.2
Probability that M/C B will work for next 10 years: P (B) = 0.9
Since the machines are independent, the probability that M/C A and M/C B
will work for the next 10 years can be obtained as
Problem 3. A, B, and C in order hit a target. The first one to hit the target wins. If
A starts, find their respective chances of winning.
S OLUTION :
It is known that
1 1
P (A wins) = P (A) = and P (A loses) = P (A) =
2 2
1 1
P (B wins) = P (B) = and P (B loses) = P (B) =
2 2
1 1
P (C wins) = P (C) = and P (C loses) = P (C) =
2 2
Now we can have the sequence
A, if A wins
AB, if A loses and B wins
A BC, if A and B lose and C wins
A BCA, if A, B, C lose and A wins and so on
1 −1 4
1
= 1− =
2 8 7
2 " 3 6 #
1 1 1
= 1+ + +······
2 2 2
2
1 −1 2
1
= 1− =
2 8 7
Problem 4. A system with three components with the probabilities that they work
is given below. Calculate the probability that the system will work.
0.8
0.9
0.8
S OLUTION :
The system S has two independent subsystems S1 and S2 and S will work if
both S1 and S2 work. That is,
Problem 5. A signal which can be green with probability 4/5 or red with proba-
bility 1/5 is received by Station A and then transmitted to Station B. The probabil-
ity of each station receiving the signals correctly is 3/4. If the signal received at
Station B is green, then find (i) the probability that the original signal was green
and (ii) the probability that the original signal was red. Also if the signal received
at Station B is red, then find (iii) the probability that the original signal was green
and (iv) the probability that the original signal was red.
S OLUTION :
Let us present the problem situation using a flow diagram as shown below:
P(G)=4/5 P(R)=1/5
Signal
Station A
Station B
Let E be the event that a signal at Station B is received, then E can be either green,
say EG or red, say ER . Therefore, we have
(i) Now, if the signal received at Station B is green, then the probability that the
original signal was green can be obtained using Bayes formula as follows:
20 3
P (R/EG ) = 1 − P (G/EG ) = 1 − =
23 23
(iii) Similarly,
= P (G)P (GA )P (RB )+P (G)P (RA )P (RB )+P (R)P (GA )P (RB )
12 5
(iv) ∴ P (R/ER ) = 1 − P(G/ER ) = 1 − =
17 17
Problem 6. An assembly consists of three mechanical components. Suppose that
the probabilities that the first, second, and third components meet specifications
are 0.95, 0.98, and 0.99. Assume that the components are independent. Determine
the probability mass function of the number of components in the assembly that
meet specifications.
S OLUTION :
Probability that first component meets specification = 0.95
Probability that second component meets specification = 0.98
Probability that third component meets specification = 0.99
Out of three components, let X be the number of components that meets specifica-
tions. Therefore, we have X = 0, 1, 2, 3.
Now,
= 10−5 = 0.00001
P (X = 1) = P (One component meets the specification)
= (0.95)(1 − 0.98)(1 − 0.99) + (1 − 0.95)(0.98)
(1 − 0.99) + (1 − 0.95)(1 − 0.98)(0.99)
= 0.00167
P (X = 2) = P (Two components meet the specification)
= (0.95)(0.98)(1 − 0.99) + (0.95)(1 − 0.98)(0.99)
+ (1 − 0.95)(0.98)(0.99)
= 0.07663
P (X = 3) = P (Three components meet the specification)
= (0.95)(0.98)(0.99)
= 0.92169
34 • Probability and Random Processes for Engineers
X =x 0 1 2 3
P(X = x) 0.00001 0.00167 0.07663 0.92469
Problem 7. A car agency sells a certain brand of foreign car either equipped with
power steering or not equipped with power steering. The probability distribution
of number of cars with power steering sold among the next 4 cars is given as
P (X = x) = 4Cx /16, x = 0, 1, 2, 3, 4
Find the cumulative distribution function of the random variable X. Using cumu-
lative probability approach verify that P(X = 2) = 3/8.
S OLUTION : nC
x
It is given that P (X = x) = , x = 0, 1, 2, 3, 4
16
Probability distribution is
4C 1 4C 4 4C 6
0 1 2
P (X = 0) = = , P (X = 1) = = , P (X = 2) = = ,
16 16 16 16 16 16
4C 4 4C 1
3 4
P (X = 3) = = , P (X = 4) = =
16 16 16 16
Cumulative distribution function:
x F(x) = P(X≤ x)
x<0 0
0≤x<1
1 1
(x = 0) 0+ =
16 16
1≤x<2
1 4 5
(x = 1) + =
16 16 16
2≤x<3
5 6 11
(x = 2) + =
16 16 16
3≤x<4
11 4 15
(x = 3) + =
16 16 16
x≥4
15 1 16
(x = 4) + = =1
16 16 16
An Overview of Random Variables and Probability Distributions • 35
Now, P (X = 2) = P (X ≤ 2) − P (X ≤ 1)
11 5 3
= − = (Verified)
16 16 8
Problem 8. Let X denote the diameter of a hole drilled in a sheet metal compo-
nent. The target diameter is 12.5 mm. Most random disturbances to the process
result in larger diameters. Historical data show that the distribution of X can be
modeled by a probability density function given below:
(
20 e−20(x−12.5) , x ≥ 12.5
f (x) =
0, Otherwise
If a part with a diameter larger than 12.6 mm is scrapped, (i) what proportion of
parts is scrapped? and (ii) what proportion of parts is not scrapped?
S OLUTION :
(i) A part is scrapped if X ≥ 12.6 then
Z∞
P (X ≥ 12.6) = f (x)dx
12.6
Z∞ ∞
− e−20(x−12.5)
⇒ 20 e−20(x−12.5) dx = 20 = 0.135
−20
12.6 12.6
It may be noted that we can obtain this probability value using the
relationship
P (X ≥ 12.6) = 1 − P (12.5 ≤ X ≤ 12.6)
Problem
( 9. If X is a random variable, then find k so that the function f (x) =
0, for x ≤ 0
2 can serve as the probability density function of X.
kxe−4x for x > 0
S OLUTION :
It is given that
2
f (x) = kxe−4x , 0≤x≤∞
36 • Probability and Random Processes for Engineers
We know that
Z∞ Z∞
2
f (x)dx = 1 ⇒ kxe−4x dx = 1
−∞ 0
1
Let 4x2 = y ⇒ 8x dx = dy ⇒ xdx = dy
8
Z∞ ∞
k −y k e−y
⇒ e dy = 1 ⇒ =1
8 8 −1 0
0
k
⇒ (1) = 1 ⇒ k=8
8
Problem 10. With equal probability, the observations 5, 10, 8, 2 and 7, show the
number of defective units found during five inspections in a laboratory. Find (a) the
first four central moments and (b) the moments about origin (raw moments).
S OLUTION :
(a) Central moments: In order to find the first four central moments, we first
obtain the computations for µr = E [X − E(X)]r , r = 1, 2, 3, 4 as given
in the following table with
1 5 32
E (X) = ∑
5 i=1
xi =
5
= 6.4
µr = E [X − E(X)]r , r = 1, 2, 3, 4,
37.2
that is, µ1 = E [X − E(X)]1 = 0, µ2 = E [X − E(X)]2 = = 7.44
5
An Overview of Random Variables and Probability Distributions • 37
−36.96
µ3 = E [X − E(X)]3 = = −7.392,
5
553.296
µ4 = E [X − E(X)]4 = = 110.66
5
(b) Raw moments: In order to obtain the raw moments we consider the
following table:
X X2 X3 X4
5 25 125 625
10 100 1000 10000
8 64 512 4096
2 4 8 16
7 49 343 2401
Total 32 242 1988 17138
1 5 1
µ1′ = E(X) = ∑ xi = 5 (32) = 6.4
5 i=1
1 5 2 1
µ2′ = E(X 2 ) = ∑ xi = 5 (242) = 48.4
5 i=1
1 5 3 1
µ3′ = E(X 3 ) = ∑ xi = 5 (1988) = 397.6
5 i=1
1 5 4 1
µ4′ = E(X 4 ) = ∑ xi = 5 (17138) = 3427.6
5 i=1
Problem 11. A man draws 3 balls from an urn containing 5 white and 7 black
balls. He gets Rs. 10 for each white ball and Rs. 5 for black ball. Find his
expectation.
S OLUTION :
Out of three balls drawn, the following combinations are possible:
(i) 3 white balls, (3W )
(ii) 2 white balls and 1 black ball, (2W , 1B )
(iii) 1 white ball and 2 black balls (1W , 2B ) and
(iv) 3 black balls (3B )
38 • Probability and Random Processes for Engineers
3W 3 × 10 = 30
2W , 1B 2 × 10 + 1 × 5 = 25
1W , 2B 1 × 10 + 2 × 5 = 20
3B 3 × 5 = 15
X =x 15 20 25 30
7 21 14 2
P(X = x)
44 44 44 44
30
∴ E(X) = ∑ xP (X = x) = ∑ xP (X = x)
x x=15
7 21 14 2
= (15) + (20) + (25) + (30)
44 44 44 44
= Rs. 21.25
An Overview of Random Variables and Probability Distributions • 39
S OLUTION :
(i) Since X is a continuous random variable, by definition, we know that
Z∞
E(X) = x f (x) dx
−∞
Z2 2 2
x x4 16 1 15
= x dx = = − =
3 12 −1 12 12 12
−1
Z2 2
x2 x5
2 32 −1 33
= x dx = = − =
3 15 −1 15 15 15
−1
Therefore,
2
33 15 33 225
V (X) = − = − = 0.6375
15 12 15 144
(iii) For finding standard deviation, by definition, we know that
p √
SD(X) = V (X) = 0.6375 = 0.7984
Now we find the mean and variance of g (X) = 4X + 3 as follows:
E [g(X)] = E(4X + 3) = 4E(X) + 3 = 4(15/12) + 3 = 8
V [g(X)] = V (4X + 3) = 16V (X) = 16(0.6375) = 10.2
40 • Probability and Random Processes for Engineers
Problem 13. The fraction X of male runners and the fraction Y of female runners
who compete in marathon races is described by the joint density function
(
8xy, 0 < x < 1, 0 < y < x
f (x, y) =
0, otherwise
S OLUTION :
In order to find the covariance, first we have to find the marginal probability
density functions for X and Y as follows. By definition we know that
Zx x
y2
f (x) = 8xydy = 8x = 4x3 , 0<x<1
2 0
0
Similarly,
Z1 1
x2
f (y) = 8xydx = 8y = 4y(1 − y2 ), 0<y<1
2 y
y
Now from the marginal density functions given above, we can compute the
expected values of X and Y as follows:
Z1 Z1 1
x5
4 4
E(X) = x f (x)dx = 4x dx = 4 =
5 0 5
0 0
Z1 Z1 1
y3 y5
2 2 8
E(Y ) = y f (y)dy = 4y (1 − y )dy = 4 − =
3 5 0 15
0 0
Also using joint density function of X and Y , we can find E(XY ) as follows:
Z1 Zx
E(XY ) = xy f (x, y)dydx
0 0
Z1 Zx Z1 Zx
= xy(8xy)dydx = 8x2 y2 dy dx
0 0 0 0
Z1 x Z1
y3
2 8
= 8x dx = x5 dx
3 0 3
0 0
An Overview of Random Variables and Probability Distributions • 41
1
8 x6 4
= =
3 6 0 9
Problem 14. The variables X and Y have the joint probability function
1 (x + y) , 0 < x < 1, 0<y<2
f (x, y) = 3
0, otherwise
S OLUTION :
The marginal probability density functions of X and Y are given by
Z2 Z2
1 2
f (x) = f (x, y)dy = (x + y)dy = (1 + x), 0<x<1
3 3
0 0
Z1 Z1
1 1 1
f (y) = f (x, y)dx = (x + y)dx = +y , 0<y<2
3 3 2
0 0
E(XY ) − E(X)E(Y )
ρXY = p p
E(X 2 ) − [E(X)]2 E(Y 2 ) − [E(Y )]2
Consider
Z1 Z1
2 5
E(X) = x f (x)dx = x (1 + x) dx =
3 9
0 0
Z1 Z1
2 2 2 2 7
E(X ) = x f (x)dx = x (1 + x) dx =
3 18
0 0
Z2 Z2
1 1 11
E(Y ) = y f (y)dy = y +y dy =
3 2 9
0 0
42 • Probability and Random Processes for Engineers
Z2 Z2
2 2 2 1 1 16
E(Y ) = y f (y)dy = y +y dy =
3 2 9
0 0
Z1 Z2 Z1 Z2
1 2
E(XY ) = xy f (x, y)dydx = xy (x + y) dy dx =
3 3
0 0 0 0
2 5 11 1
− − r
3 9 9 81 2
∴ ρXY =s 2 s 2 = r 13 r 23 = − 299
7 5 16 11
− − 162 81
18 9 9 9
Problem 15. Suppose that 10 cones are selected for weight test. From the past
records 2 out of the 10 cones on the lot are expected to be below standards for
weight, what is the probability that at least 2 cones will be found not meeting
weight standards?
S OLUTION :
It is given that the probability that the cones are below standards, say
2
p= = 0.2
10
q = 1 − p = 0.8
Let X be the number of cones not meeting standards, then the probability that out
of 10, at least two cones will not meet weight standards
10 10
P (X ≥ 2) = ∑ 10
Cx px q10−x = ∑ 10Cx (0.2)x (0.8)10−x
x=2 x=2
1
= 1 − ∑ 10Cx (0.2)x (0.8)10−x
x=0
h i
10
= 1− C0 (0.2)0 (0.8)10 + 10C1 (0.2)1 (0.8)10−1
= 1 − [0.10737 + 0.26844]
= 0.6242
S OLUTION :
It is given that the probability that a bulb is defective, say p = 0.05
Number of bulbs in one box, say n = 100
Since n is large and p ≤ 0.05, we can use Poisson approximation with
λ = np = (100)(0.05) = 5
Therefore, a box will fail to meet the guaranteed quality if the number of defective
bulbs, say X, exceeds 10. Then the required probability is
P (X ≥ 11) = 1 − P (X ≤ 10)
∞ −5 x
10
e −λ λ x e 5
= 1− ∑ = 1− ∑
x=0 x! x=0 x!
51 52 53 54 55 56 57 58 59 510
−5
= 1− e 1+ + + + + + + + + +
1! 2! 3! 4! 5! 6! 7! 8! 9! 10!
( !)
1 + 5 + 12.5 + 20.83 + 26.04 + 26.04 + 21.70
= 1 − e −5
+15.5 + 9.68 + 5.38 + 2.69
S OLUTION :
Let the probability of an accident on any given day be p = 0.005
It is given that the number of days is n = 400
Since n is large and p is small, we can approximate this to a Poisson distribu-
tion as mean λ = np = (400)(0.005) = 2
Now, if we let X as the random variable that represents number of accidents,
then X follows a Poisson distribution with mean λ = 2. Therefore,
e−2 2x
P(X = x) = , x = 0, 1, 2, . . . . . .
x!
44 • Probability and Random Processes for Engineers
(i) Now the probability that there is one accident on a day is given by
e−2 21
P(X = 1) = = 0.271
1!
(ii) The probability that there are at most three days with an accident is
given by
3
e−2 2x
P (X ≤ 3) = ∑
x=0 x !
S OLUTION :
It is given that mean E(X) = µ ⇒ E(x) = µ
And standard deviation σ = 0.01 mm
2 (0.01)2
Then variance V (X) = σ 2 ⇒ V (x) = σn =
n
σ 0.01
⇒√ = √
n n
8
We need to find the sample size n such that P X − µ ≤ 0.005 ≥
9
By Chebyshev’s theorem, we know that
σ
1
P X −µ ≤ k
√ ≥ 1− 2
n k
0.01 1
P X −µ ≤ k √
⇒ ≥ 1− 2
n k
1 8
Let 1 − 2
= ⇒ ε =3
k 9
0.01 8
P X −µ ≤ 3 √
∴ ≥
n 9
0.01
Letting 3 √ = 0.005 ⇒ n = 36
n
An Overview of Random Variables and Probability Distributions • 45
Therefore, 36 diameters must be measured that will ensure that the average
of the measured diameters is within 0.005 mm of the process mean diameter with
probability at least 8/9.
S OLUTION :
Let X be the thickness of the photo resist. Then X is uniform in the interval
(0.205, 0.215)
1 1
f (x) = = , 0.205 < x < 0.215
b − a 0.215 − 0.205
1
= , 0.205 < x < 0.215
0.01
Let a be the thickness such that 10% of the wafers exceed this thickness
That is,
P(X ≥ a) = 0.10
0.215 0.215
1
Z Z
⇒ f (x)dx = dx = 0.10
0.01
a a
0.215 − a
⇒ = 0.10
0.01
⇒ a = 0.214
Problem 20. Let X be a normal variable with mean µ and standard deviation σ .
If Z is the standard normal variable such that Z = −0.8 when X = 26 and Z = 2
when X = 40, then find µ and σ . Also find P(X > 45) and P (| X − 30 |) > 5).
S OLUTION :
Given a normal random variable X with mean µ and standard deviation σ the
X −µ
standard normal variate is given by Z =
σ
26 − µ
It is given that Z = = − 0.8 ⇒ −0.8σ + µ = 26
σ
46 • Probability and Random Processes for Engineers
40 − µ
and Z= = 2 ⇒ 2σ + µ = 40
σ
Solving the above two equations, we get µ = 30 and σ = 5
Now, consider
X − 30 45 − 30
P (X > 45) = P >
5 5
= P (Z > 3)
Similarly,
= 1 − P (−5 ≤ X − 30 ≤ 5)
= 1 − P (25 ≤ X ≤ 35)
25 − 30 X − 30 35 − 30
= 1−P ≤ ≤
5 5 5
= 1 − P (−1 ≤ Z ≤ 1)
= 1 − [P (Z ≤ 1) − P (Z ≤ −1)]
= 1 − [ϕ (1) − ϕ (−1)]
S OLUTION :
Since X is uniform in the interval (0, 1), we have
(
1, 0<x<1
f (x) =
0, otherwise
√
Now, y = g (x) = x ⇒ x = w (y) = y 2
dx
J = w ′ (y) = = 2y
dy
An Overview of Random Variables and Probability Distributions • 47
We know that the probability density function of Y say h (y) can be given as
Problem 22. The joint probability density function of random variables X and Y is
given by
(
e−(x+y) , x > 0, y > 0
f (x, y) =
0, otherwise
Then find whether X and Y are independent. Also find the probability density
function of the random variable U = (X +Y )/2.
S OLUTION : (
e−(x+y) , x > 0, y > 0
Given f (x, y) =
0, otherwise
Z∞
Now, f (x) = e−(x+y) dy = e−x , x>0
0
Z∞
f (y) = e−(x+y) dx = e−y , y>0
0
∂x ∂x
∂v 2 −1
⇒ J = du = =2
∂y ∂y 0 1
∂u ∂v
Z2u
⇒ f (u) = 2 e−2u du = 4ue−2u , u>0
0
(
4u e−2u , u>0
∴ f (u) =
0, otherwise
Problem 23. If X = (X1 , X2 ) is a bivariate normal random vector with mean vector
′ 5 4
(0, 0) and covariance matrix then obtain the means and variances of
4 5
X1 and X2 and also the correlation between X1 and X2 . Hence, obtain the joint
probability density function of X1 and X2 .
S OLUTION :
It is given that µ = (µ1 , µ2 )′ = (0, 0)′
µ1
! ! !
E(X1 ) 0
⇒µ= = =
µ2 E(X2 ) 0
σ12 σ12
! !
5 4
Also it is given that Σ = =
σ21 σ22 4 5
( " 2 #)
x2 − 0 2
−1 x1 − 0 x1 − 0 x2 −0
exp √ − 2(0.8) √ √ + √
2 1 − 0.82 5 5 5 5
= √ √ q
2π 5 5 1 − 0.82
EXERCISE PROBLEMS
1. Two newspapers A and B are published in a city and a survey of readers
indicates that 20% read A, 16% read B, and 8% read both A and B. For
a person chosen at random, find the probability that he reads none of the
papers.
2. The following circuit operates if and only if there is a path of functional
devices from left to right. The probability that each functions is as shown
in figure. Assume that the probability that a device is functional does not
depend on whether or not the other devices are functional. What is the prob-
ability that the circuit operates?
X =x 1 2 3 4 5 6 7 8
P (X = x) a 3a 5a 7a 11a 13a 15a 17a
50 • Probability and Random Processes for Engineers
(a) Find the value of a, (b) Find P (X < 3) and (c) find the cumulative
probability distribution of X.
5. If the cumulative distribution function of a random variable X is given by
0, x<0
2
F(X) = x /2, 0 ≤ x < 4 then find P (X > 1/X < 3).
1,
x≥4
where a and k are constants, then find the value of k and find the probability
that a given batch will be accepted.
7. A box is to be constructed so that its height is 5 inches and its base is X
inches by X inches, where X is a random variable described by the proba-
bility density function
6x(1 − x), 0 ≤ x ≤ 1
f (x) =
0, otherwise
10. In 1 out of 6 cases, material for bulletproof vests fails to meet puncture
standards. If 405 specimens are tested, what does Chebyshev’s theorem tell
us about the probability of getting at most 30 or more than 105 cases that do
not meet puncture standards?
11. The lifetime of a certain brand of electric bulb may be considered as a ran-
dom variable with mean 1200 hours and standard deviation 250 hours. Find
the probability, using central limit theorem, that the average lifetime of 60
bulbs exceeds 1250 hours.
12. Let X and Y be two random variables with the joint probability density func-
tion given as
x (1 + 3y2 )/4, 0 < x < 2, 0 < y < 1
f (x, y) =
0, otherwise
2.0 INTRODUCTION
In this chapter, the concept of random process is explained in such a way that it
is easy to understand. The concepts of random variable and random function are
discussed. Many examples are presented to illustrate the nature of various random
processes. Since a random process depends on both time and state space, the ran-
dom process is properly interpreted and classified into different categories based
on the combination of time index and state space. Further, the statistical averages
of random processes are presented since the outcomes of a random process are
probabilistic in nature. Several problems are worked out and exercise problems
are given.
Outcome(e) e1 = T e2 = H
X(e) 0 1
0
0
1
0
0
1
1
and so on
0
0
1
1
0
1
1
0 t
t1 t2 t3 ⋯ ti ⋯ tm ⋯ ⋯
Figure 2.1. Possible combinations of outcomes of tossing a coin at different time points
54 • Probability and Random Processes for Engineers
Time
Outcome
t1 t2 t3 t4 t5 t6 t7 t8 t9 t10
ξ1 1 0 0 1 0 1 0 0 1 0
ξ2 0 1 1 0 1 0 0 1 0 1
.. .. .. .. .. .. .. .. .. .. ..
. . . . . . . . . . .
ξi 0 0 1 0 0 1 1 0 1 1
.. .. .. .. .. .. .. .. .. .. ..
. . . . . . . . . . .
ξn 1 1 0 0 0 0 1 0 0 0
For illustration purpose, let us assume that the values shown in the rows of
Table 2.2 are the n possible outcomes, say ξ = ξ1 , ξ2 , · · · , ξi , · · · , ξn , when the
coin is tossed ten times, that is at time points t1 , t2 , · · · , ti , · · · , t10 in (0, t). For
example, one of the possible outcomes (there are 210 outcomes of different com-
binations in total) from ten tosses (first row of Table 2.2) of the coin at time points
t1 ,t2 , · · · , ti , · · · , t10 could be, say ξ1 = 1, 0, 0, 1, 0, 1, 0 , 0, 1, 0 respectively.
Naturally, the events of ξ1 ∈ (e1 , e2 ) assume the values from (0, 1). It may be noted
that in every time point the possible outcomes are either e1 = 0 or e2 = 1, whereas
the outcomes for 10 tosses may happen in combination of ‘0’s and ‘1’s. Similarly,
we can obtain ξi , i = 2, 3, · · · · · · , n. In general, the events of ξ ∈ (e1 , e2 ) assume
values from (0, 1).
Let us assume that for every toss you will get rupees ten multiplied by the
time t at which it is tossed and for every toss you will lose rupees five multiplied
by the time t at which it is tossed. That is, if the time point is t1 then the multi-
plying factor is 1, if the time point is t2 , then the multiplying factor is 2 and so
on. Under this condition, if we define the random function X(t, ξ ) as the amount
gained at time point t, then we will have the values of random functions X(t, ξi ),
i = 1, 2, 3, · · · · · · , n as given in Table 2.3.
Table 2.3. Possible outcomes for gain in rupees when a coin is tossed
Time
Outcome
t1 t2 t3 t4 t5 t6 t7 t8 t9 t10
ξ1 10 −10 −15 40 −25 60 −35 −40 90 −50
ξ2 −5 20 30 −20 50 −30 −35 80 −45 100
.. .. .. .. .. .. .. .. .. ..
. . . . . . . . . .
ξi −5 −10 30 −20 50 60 70 80 90 100
.. .. .. .. .. .. .. .. .. ..
. . . . . . . . . .
ξn 10 20 −15 −20 −25 −30 70 −40 −45 −50
Introduction to Random Processes • 55
For example, from the first row of Table 2.3, we have X(t1 , ξ1 ) = 10,
X(t2 , ξ1 ) = −10, X(t3 , ξ1 ) = −15 and so on X(t10 , ξ1 ) = −50. Therefore, in this
example, based on the outcomes ξ at the time points t1 = 1, · · · · · · , t = t10 = 10 in
(0, t), we can represent the random function X(t, ξ ) as:
If we plot this function connecting the points with a smooth curve (at this
stage never mind the time points are discrete) then we have the graph as shown
in Figure 2.2. Note that each curve is occurring in a random fashion. Also, in this
case, it may be noted that both the outcome ξ and the time parameter t change
randomly simultaneously. In addition, the changes in the outcomes depend on the
changes in the time points. However, it may be noted that at a particular time point
ti for some i = 1, 2, · · · · · · , m, · · · · · · we have the random variable X(ti , ξ ) with
different outcomes. For example, In Figure 2.2, at time t4 the intersecting points of
the vertical line and the curves show the values (outcomes) of the random variable
X(t4 , ξ ).
X (t,ξ )
Outcome (gain in rupees)
100 X (t4,ξ )
X (t, ξi)
50
X (t, ξ2)
0
X (t, ξ1)
−50
X (t, ξn)
−100 t
0 1 2 3 4 5 6 7 8 9 10
t4
Time
Here, it may be noted that the experimental outcomes are fixed. That is, either ξ1
or ξ2 happens based on which the function changes over a given period of time.
Hence, we get the random function as X(t, ξ1 ) = − sin(1 + t) when tail turns up
(e1 = T ) or the function X(t, ξ2 ) = sin(1 + t) when head turns up (e2 = H). If
56 • Probability and Random Processes for Engineers
these functions are plotted against the time points, we obtain the smooth curves as
shown in Figure 2.3.
In Figure 2.3, at time t1 the intersecting points of the vertical line and the
curves show the values of the random variable X(t1 , ξ ).
X (t , ξ )1.5 X (t 1, ξ )
1
0.5 X (t , ξ1 )
Outcome
0
−0.5
X (t , ξ 2 )
−1
−1.5 t
0 1 2 3 4 5 6 7 8 9 10
Time
t1
Time
Outcome
t1 t2 t3 t4 t5 t6 t7 t8 t9 t10
ξ1 6 3 6 6 2 4 2 2 5 1
ξ2 2 5 1 3 3 6 2 1 2 4
.. .. .. .. .. .. .. .. .. .. ..
. . . . . . . . . . .
ξi 1 5 3 4 4 1 5 1 4 3
.. .. .. .. .. .. .. .. .. .. ..
. . . . . . . . . . .
ξn 4 2 1 5 2 3 2 6 1 2
Introduction to Random Processes • 57
Let us assume that you win an amount equivalent to the face of the dice
that turned up multiplied by the time t at which it is thrown. If we let the ran-
dom function X(t, ξ ) as the amount won at time point t, the values of the n
random functions X(t, ξi ), i = 1, 2, 3, · · · · · · , n based on the possible outcomes
ξ = ξ1 , ξ2 , · · · , ξi , · · · , ξn are presented in Table 2.5.
Table 2.5. Possible outcomes for gain in rupees when a dice is thrown
Time
Outcome
t1 t2 t3 t4 t5 t6 t7 t8 t9 t10
ξ1 6 6 18 24 10 24 14 16 45 10
ξ2 2 10 3 12 15 36 14 8 18 40
.. .. .. .. .. .. .. .. .. ..
. . . . . . . . . .
ξi 1 10 9 16 20 6 35 8 36 30
.. .. .. .. .. .. .. .. .. ..
. . . . . . . . . .
ξn 4 4 3 20 10 18 14 48 9 20
Looking at Table 2.5, in this example of throwing a dice ten times, for time
points t1 = 1, · · · · · · , t = t10 = 10 in (0,t), we can represent the random function
X(t, ξ ) as a function of t and ξ as follows:
X(t, ξ ) = at
where a = 1, 2, 3, 4, 5 or 6. If we plot this function connecting the points (though
discrete) with a smooth curve then we have the graph as shown in Figure 2.4. Note
that each curve is occurring in a random fashion.
Also, in this case, it may be noted that both the outcome ξ and the time param-
eter t change randomly simultaneously. In addition, the changes in the outcomes
depend on the changes in the time points.
X (t , ξ ) 50
X (t6 , ξ)
Outcome (gain in rupees)
X (t, ξ2 )
40
30 X (t, ξi )
20 X (t, ξn )
10 X (t, ξ1 )
0 t
0 1 2 3 4 5 6 7 8 9 10
t6
Time
In Figure 2.4, at time t6 the intersecting points of the vertical line and the
curves show the values of the random variable X(t6 , ξ ).
58 • Probability and Random Processes for Engineers
X (t , ξ ) 2 X(ti ,ξ )
X (t , ξ n )
X (t , ξ i )
X (t , ξ 2 )
X (t , ξ1 )
t
ti
Figure 2.5. Graphical representation of the random function X(t, ξ ) in Example 2.4
Here, if we look at the values of the function at a particular time point of the
time parameter, say ti , then the function X(t, ξ ) becomes a random variable as
X(ti , ξi ) at time point ti with outcomes ξ1 = 0.05, ξ2 = 0.4, ξi = 0.75, ξn = 0.95,
that is, the points of the four curves as they pass through the time point ti . Also,
it could be seen that at time point ti , the outcome ξ is uniformly distributed in the
interval (0, 1). In Figure 2.5, at time ti the intersecting points of the vertical line
and the curves show the values of the random variable X(ti , ξ ).
Introduction to Random Processes • 59
X (t , ξ 2 )
20.5
X (t , ξ i )
19.5
X (t , ξ n )
18.5 X (t , ξ1 )
17.5 t
0 1 ti 2 3 4
Time
2.2.1 Definition
A random process is defined as the collection of random functions, say {X(t, ξ )},
ξ = ξ1 , ξ2 , · · · · · · , ξi , · · · · · · , ξn · · · , together with a probability rule.
The probability rule implies that each random function X(t, ξi ), i = 1,
2, · · · · · · , n, · · · · · · is associated with a probability of its happening over a time
period.
That is, the random process denoted by {X(t, ξ )} is the collection of (the
uncountably infinite if the state space ξ is continuous or countably infinite if the
state space ξ is discrete) random functions X(t, ξ1 ), X(t, ξ2 ), . . . , X(t, ξi ), . . . ,
X(t, ξn ), . . . . with state space ξ = ξ1 , ξ2 , · · · · · · , ξi , · · · · · · , ξn , . . . . . . It could be
60 • Probability and Random Processes for Engineers
seen that each random function of the random process {X(t, ξ )} is indexed by the
time parameter t and the state space ξ . Here, the state space represents the state
(outcome) of the random function at time t. The collection of the random functions
is also called an ensemble.
Further, it may be noted that at each time point ti , i = 1, 2, · · · · · · , m, · · · · · · in
the interval (0, t), we have a random variable denoted by X(ti , ξ ) whose realiza-
tions are X(ti , ξ1 ), X(ti , ξ2 ), . . . , X(ti , ξi ), . . . , X(ti , ξn ) · · · · · · . Therefore, a ran-
dom process can also be defined as the collection of random variables {X(t, ξ )}
at time points t = t1 , t2 , · · · , ti , · · · , tm , · · · in (0, t) together with a probability rule
indexed by time parameter t and the state space ξ .
Such collection of random variables is uncountably infinite if the time param-
eter t is continuous or countably infinite if the time parameter t is discrete.
In an ensemble, since the happening of each member function X(t, ξi ),
i = 1, 2, · · · · · · , n, · · · · · · , depends on the happening of the corresponding exper-
imental outcome ξi according to the known probability rule, the random process
is usually denoted by {X(t)} or simply X(t). Note that in case of random variable
X(e), we denote the same as simply X, omitting e. Apparently, when we denote a
random process by X(t), we mean that it is a random process observed in the time
interval (0, t).
For example, recall the Illustrative Example 2.2 in which the member functions
X(t, ξ1 ) and X(t, ξ2 ) of the ensemble are given as follows:
Since we know that for a fair coin, probabilities are equal for the happening of a
tail or a head, according to the probability law we have
1 1
P [X(t, ξ1 ) = − sin(1 + t)] = and P [X(t, ξ2 ) = sin(1 + t)] =
2 2
1 1
P [X(t) = − sin(1 + t)] = and P [X(t) = sin(1 + t)] =
2 2
Table 2.6. Probability distribution of the random process given in Example 2.2
x − sin(1 + t) sin(1 + t)
P[X(t) = x] 1/2 1/2
State Space (ξ )
Fixed Variable
Fixed A value X(ti , ξi ) Random variable X(ti , ξ )
Time Parameter (t)
Variable Single Random function X(t, ξi ) An ensemble {X(t, ξ )}
Interpretation 1:
If the state space ξ and time parameter t are fixed at (ti , ξi ) for some
i = 1, 2, · · · · · · , m, · · · · · · , then the process stops by assigning a value say X(ti ξi ).
Consider the Illustrative Example 2.1, in tossing the coin at the fixed time point
t = 4, for a fixed outcome of a head we have X(t4 , ξ1 ) = 40 and for a fixed outcome
of a tail we have X(t4 , ξ2 ) = −20 (Refer to Table 2.3).
Interpretation 2:
If the state space ξ is fixed and time parameter t is allowed to vary such as
t1 , t2 , · · · , ti , · · · , tm , · · · in (0, t), then we have a single random function X(t, ξi )
for some i = 1, 2, · · · · · · , m, · · · · · · . Consider the Illustrative Example 2.2, in which
the coin is tossed repeatedly at the time points t1 , t2 , · · · , ti , · · · ,t10 in (0, t)
we get the function X(t, ξ1 ) = − sin(1 + t) if the outcome of the toss is a tail
and the function X(t, ξ2 ) = sin(1 + t) if the outcome results in a head. When
these functions are plotted with smooth curves for t = t1 , t2 , · · · , ti , · · · , t10 in
(0, t) then we get the graphical representation of these two functions as shown in
Figure 2.3.
Interpretation 3:
If the state space ξ is allowed to vary such as ξ = ξ1 , ξ2 , · · · · · · ,
ξi , · · · · · · , ξn , · · · · · · and time parameter t is fixed at ti , then we have a random
variable X(ti , ξ ) for some i = 1, 2, · · · · · · , m, · · · · · · in (0, t). Consider the Illus-
trative Example 2.3, in which if the dice is thrown at a time point t6 for some i =
1, 2, · · · · · · , m, · · · · · · in (0, t), we get the random variable X(t6 , ξi ) that assumes
the values X(t6 , ξ1 ) = 24, X(t6 , ξ2 ) = 36, X(t6 , ξi ) = 6 and X(t6 , ξn ) = 18
62 • Probability and Random Processes for Engineers
(Refer Table 2.5 and Figure 2.4). When these functions are plotted with smooth
curves for t1 , t2 , · · · , ti , · · · , t10 in (0, t) then we get the graphical representation
of these functions as shown in Figure 2.4.
Interpretation 4:
If the state space ξ is allowed to vary such as ξ = ξ1 , ξ2 , · · · · · · , ξi , · · · · · · ,
ξn , · · · · · · and time parameter t is allowed to vary such as t1 , t2 , · · · ,
ti , · · · , tm , · · · in (0, t), then we have an ensemble of random functions {X(t, ξ )}
whose member functions are X(t, ξi ) for i = 1, 2, · · · · · · , n, · · · · · · in (0, t).
Consider the Illustrative Example 2.4, in which the state space ξ is continuous
uniform random variable in (0, 1) and the random functions are observed over a
period of time in (0, t) continuously. Similarly, in the Illustrative Example 2.5,
the temperature values are continuous between 18◦ C and 22◦ C, and is observed
over a continuous time interval (0, t). In the given time interval (0, t) the random
functions X(t, ξ1 ), X(t, ξ2 ), X(t, ξi ) and X(t, ξn ) are observed and the graphical
representation of these functions are shown in Figures 2.5 and 2.6.
State Space (ξ )
Discrete Continuous
Discrete Discrete random Continuous random
sequence sequence
Time Parameter (t)
Continuous Discrete random Continuous random
process process
Discrete random sequence: If time parameter t is discrete and state space ξ is also
discrete then each member function of the ensemble {X(t, ξ )} is called a discrete
random sequence. Refer to Illustrative Example 2.1 of tossing a coin in which the
outcomes ξ are discrete (0 for tail and 1 for head) and Illustrative Example 2.3 of
throwing a dice in which the outcomes are discrete (1, 2, 3, 4, 5, 6). In these cases
the time parameter t is also discrete as the experiments are conducted at specific
time points t1 , t2 , · · · , ti , · · · , tm , · · · .
Introduction to Random Processes • 63
Here, P {X(t) = x} and f (x,t) are respectively called the first order PMF and first
order PDF of the random process {X(t)}. The second order PMF (joint PMF)
and second order PDF (joint PDF) of the random process {X(t)} are respectively
denoted as P {X(t1 ) = x1 , X(t2 ) = x2 } and fXX (x1 , x2 ; t1 , t2 ) or f (x1 , x2 ; t1 , t2 ) or
fX(t1 )X(t2 ) (x1 , x2 ). Now, the second order CDFs for discrete and continuous cases
denoted by FXX (x1 , x2 ; t1 , t2 ) or F(x1 , x2 ; t1 , t2 ) or FX(t1 )X(t2 ) (x1 , x2 ) can be
obtained as
Zx1 Zx2
= f (x1 , x2 ; t1 , t2 ) dx2 dx1
−∞ −∞
Similarly, the m th order PMF and m th order PDF the random process {X(t)}
are respectively given as P {X(t1 ) = x1 , X(t2 ) = x2 , · · · · · · , X(tm ) = xm } and
f (x1 , x2 , · · · · · · , xm ; t1 ,t2 , · · · · · · , tm ) and hence the m th order CDFs for discrete
and continuous cases can be obtained as
Note:
If {X1 (t)} and {X2 (t)} are two random processes observed over a period
of time (0, t), and X1 (t1 ) is a random variable of the process {X1 (t)} at the
time point t1 , and X2 (t2 ) is a random variable of the process {X2 (t)} at the time
point t2 , then their joint PMF and joint PDF are respectively denoted by
Introduction to Random Processes • 65
P {X1 (t1 ) = x1 , X2 (t2 ) = x2 } and fX1 X2 (x1 , x2 ; t1 , t2 ). Similarly, the CDF for two
variable case is denoted as FX1 (t)X2 (t) (x1 , x2 ).
+∞
µx (t)=E {X(t)}=∑ xP {X(t) = x} if the outcome of the process {X(t)} is discrete
x=−∞
Z∞
= x f (x,t) dx if the outcome of the process {X(t)} is continuous
−∞
(2.4)
Autocorrelation:
If {X(t)} is a random process and X(t1 ) and X(t2 ) are the two random variables
of the process at two time points t1 and t2 , then the autocorrelation of the process
{X(t)} denoted by Rxx (t1 , t2 ) is obtained as the expected value of the product of
X(t1 ) and X(t2 ). That is,
Correlation:
If {X(t)} is a random process and X(t1 ) and X(t2 ) are the two random variables
of the process at two time points t1 and t2 , then the correlation between X(t1 ) and
X(t2 ), denoted by ρxx (t1 , t2 ), is given by
Cx1 x2 (t1 , t2 ) = E {X1 (t1 )X2 (t2 )} − E {X1 (t1 )} E {X2 (t2 )}
In this case, the correlation between the random variables X1 (t1 ) and X2 (t2 ),
denoted as ρx1 x2 (t1 , t2 ), is given by
It may be noted that without loss of generality, and of course for clarity, it is
assumed that t = t1 in {X1 (t)} and t = t2 in {X2 (t)}.
Note:
In discrete case, we have random sequence denoted by {Xn } instead of {X(t)},
since the time t is discrete the same is represented in terms of steps, say
n = 0, 1, 2, · · · · · · .
Introduction to Random Processes • 67
It may be noted that unless or otherwise it is provided, the mean of the white
noise process is always assumed as zero. If X(t1 ) and X(t2 ) are uncorrelated and
independent, then the process {X(t)} is called strictly white noise process.
Given two random variables X(t1 ) and X(t2 ) of a white noise process {X(t)},
the auto-covariance is usually of the form
SOLVED PROBLEMS
Problem 1. A random process {X(t)} has the sample functions of the form X(t) =
Y sin ω t where ω is a constant and Y is a random variable that is uniformly dis-
tributed in (0, 1). Sketch three sample functions for Y = 0.25, 0.5, 1 by fixing
ω = 2. Assume 0 ≤ t ≤ 10.
S OLUTION :
Since Y is a random variable that is uniformly distributed in (0, 1) we consider
three arbitrary values Y = 0.25, 0.5, 1. Now, for ω = 2, we have three sample
functions of {X(t)} as
X (t )
X (t ) = (1) sin 2t
X (t ) = (0.5) sin 2t
X (t ) = (0.25) sin 2t
−
Problem 2. If {X(t)}is a random process then obtain the probability mass func-
tion of the sample function X(t)at time point t = 5 in cases of (i) repeated tossing
of a coin (ii) repeated rolling of a die.
S OLUTION :
(i) Let X (5) be the outcome of the coin tossed at time point 5, and let
(
1, if head turns at time point 5
X(5) =
0, if tail turns at time point 5
(ii) Now, let X (5) be the outcome of the die rolled at time point 5, then we have
X(5) = 1 or 2 or 3 or 4 or 5 or 6
S OLUTION :
It is given that A is a random variable uniformly distributed over some interval
(0, a0 ), therefore we have the probability density function of A as
1
, 0 < a < a0
f (a) = a0
0, otherwise
Za0
E {X(t)} = cos(2π at) f (a)da
0
n o Za0
Consider 2
E X (t) = cos2 (2π at) f (a)da
0
Za0
1
= cos2 (2π at)da
a0
0
Za0
1 + cos(4π at)
1
= da
a0 2
0
a
Z0 Z a
1 0
= da + cos(4π at)da
2a0 0
0
sin 4π at a0
1
= a0 +
2a0 4π t 0
sin 4π a0 t
1 1
= a0 + = (1 + sin c(4a0t))
2a0 4π t 2
70 • Probability and Random Processes for Engineers
n o
∴ V {X(t)} = E X 2 (t) − {E[X(t)]}2
1
= {1 + sin c (4a0t)} − (sin c (2a0t))2
2
S OLUTION :
It is given that
− (t1 µx + µy )(t2 µx + µy )
Note:
(i) Variance of {Z(t)} may also be obtained by letting t1 = t2 = t in Czz (t1 ,t2 )
given in (iv).
(ii) If X and Y are independent random variables, then we have E(XY ) =
Cov (X, Y )
E(X) E(Y ) and hence ρxy = = 0, therefore we have the results
σx σx
Rzz (t1 ,t2 ) = t1t2 (σx2 + µx2 ) + (t1 + t2 )µx µy + (σy2 + µy2 )
Problem 5. Suppose that {X(t)} is a random process with µ (t) = 3 and C(t1 ,t2 ) =
4e−0.2|t1 −t2 | . Find (i) P [X(5) ≤ 2] and (ii) P [|X(8) − X(5)| ≤ 1] using central limit
theorem.
S OLUTION :
It is given that E {X(t)} = µ (t) = 3 and V {X(t)} = C(t,t) = 4e−0.2|0|!= 4
X(5) − E {X(5)} 2 − E {X(5)}
(i) Consider P [X(5) ≤ 2] = P p < p
V {X(5)} V {X(5)}
!
2 − µ (5)
2−3
=P Z< p =P Z< √
C (5, 5) 4
(ii) Consider
P {|X(8) − X(5)| ≤ 1}
!
X(8) − X(5) − E {X(8) − X(5)} 1 − E {X(8) − X(5)}
=P p < p
V {X(8) − X(5)} V {X(8) − X(5)}
= P (|Z| ≤ 0.526)
= P (−0.526 ≤ Z ≤ 0.526)
S OLUTION :
It is given that E {X(t)} = µ (t) = 8 and R(t1 ,t2 ) = 64 + 10e−2|t1 −t2 |
(i) Mean
(ii) Variance
(iii) Covariance
S OLUTION :
Given X(t) = Y cos ω t
Since Y is a uniform random variable in the interval (0, 1) the probability density
function (PDF) of Y is given as
1, 0<y<1
f (y) =
0, otherwise
The following figure depicts three samples of X(t) = Y cos(2t) for Y = 0, Y = 0.5
and Y = 1.0 for a fixed value of ω = 2.
1.5
X (t ) X(t) = (1)cos(2t)
1
0.5
X(t) = (0.5)cos(2t)
0
−0.5
X(t) = (0)cos(2t)
−1
t
−1.5
−7 −6 −5 −4 −3 −2 −1 0 1 2 3 4 5 6 7
74 • Probability and Random Processes for Engineers
We know that if the PDF of the random variable Y is known and the random vari-
able X = g (Y ) ⇒ y = w (x) then the PDF of the random variable X can be obtained
using the transformation
dy
where |J| = w ′ (x) = .
dx
Unlike Equation 1.30, note that in this case the new variable is X and the old
variable is Y .
When, t = 0 we have X(0) = Y cos 0 = Y
Therefore, the probability density function f (x) of {X(t)} at t = 0 becomes
(
dy 1, 0<x<1
f (x) = f [w (x)] =
dx 0, otherwise
1
(i) When, t = π /4ω , we have X(π /4ω ) = Y cos(ωπ /4ω ) = Y cos π /4 = √ Y
2
Therefore, the probability density function f (x) of {X(t)} at t = π /4ω
becomes
(√ √
dy 2, 0 < x < 1/ 2
f (x) = f [w (x)] =
dx 0, otherwise
(i) Find E {X(t)} at t = 1/4. and (ii) Find the probability distribution function
F(x, t) at t = 1/4.
S OLUTION :
We know that in the experiment of tossing a fair coin,
1
P(head) = P(tail) =
2
(i) Now,
1 1
P(head) = ⇒ P {X(t) = sin π t} =
2 2
1 1
P(tail) = ⇒ P {X(t) = 2t} =
2 2
E {X(t)} = ∑ xP {X(t) = x)}
x
S OLUTION :
Consider the cumulative probability distribution function of {X(t)}
= P {Y ≤ x/ |cos(2π f t)|}
Z π f t)|
x/|cos(2
= fY (y)dy
0
Z π f t)|
x/|cos(2
1 −y/10
= e dy
10
0
= 1 − e−x/10|cos 2π f t|
(
0, x<0
∴ FX(t) (x) =
1 − e−x/10|cos 2π f t| , x ≥ 0
We know that the probability density function of the process {X(t)} can be
given as
1
e−x/10|cos 2π f t| , x ≥ 0
′
fX(t) (x) = FX(t) (x) = 10 |cos 2π f t|
0, x<0
S OLUTION :
Given X(t) = A cos(ω t + θ )
Since θ is a uniform random variable distribution in the interval (−π , π ), we have
the probability density function of θ as
1
f (θ ) = , −π ≤ θ ≤ π
2π
We know that mean of {X(t)} is given by
E {X(t)} = E(A)E [cos(ω t + θ )] (since A and θ are independent)
Zπ
1
E [cos(ω t + θ )] = cos(ω t + θ ) dθ
2π
−π
Zπ
1
= cos(ω t + θ )d θ
2π
−π
1
= [sin(ω t + θ )]π−π
2π
1
= [sin(ω t + π ) − sin(ω t − π )] = 0
2π
∴ E {X(t)} = E(A)(0) = 0
pendent)
1 + cos 2(ω t + θ )
Consider E [cos (ω t + θ )] = E
2
2
1 1
= + E[cos 2(ω t + θ )]
2 2
Zπ
1 1 1
= + cos 2(ω t + θ ) dθ
2 2 2π
−π
Zπ
1 1
= + cos 2(ω t + θ )d θ
2 4π
−π
1 sin(2ω t + 2θ ) π
1
= +
2 4π 2 −π
78 • Probability and Random Processes for Engineers
1 1
= + [sin(2ω t + 2π ) − sin(2ω t − 2π )]
2 4π
1
=
2
n o 1
∴ E X (t) = E(A )E [cos (ω t + θ )] = E (A )
2 2 2 2
2
1
= E(A2 )E {cos ω (t1 − t2 ) + cos [ω (t1 + t2 ) + 2θ ]}
2
1
= E(A2 ) {E [cos ω (t1 − t2 )] + E (cos[ω (t1 + t2 ) + 2θ ])}
2
Zπ
1 2
= E(A ) cos ω (t1 − t2 ) f (θ )d θ
2
−π
Zπ
+ cos [ω (t1 + t2 ) + 2θ ] f (θ )d θ
−π
Zπ
1 1
= E(A2 ) cos ω (t1 − t2 ) dθ
2 2π
−π
Zπ
1
+ cos [ω (t1 + t2 ) + 2θ ] dθ
2π
−π
1 1
= E(A2 ) cos ω (t1 − t2 ) − |sin [ω (t1 + t2 ) + 2θ ]|π−π
2 2π
1 1
= E(A2 ) cos ω (t1 − t2 ) = E(A2 ) cos ωτ
2 2
Introduction to Random Processes • 79
Note that
|sin [ω (t1 + t2 ) + 2θ ]|π−π = sin [ω (t1 + t2 ) + 2π ] − sin [ω (t1 + t2 ) −2π ] = 0
1
∴ Rxx (t1 ,t2 ) = E(A2 ) cos ωτ
2
We know that covariance is given by
1
Cxx (t1 ,t2 ) = Rxx (t1 ,t2 ) − E {X(t1 )} E {X(t2 )} = E(A2 ) cos ωτ − (0)(0)
2
1
= E(A2 ) cos ωτ
2
Problem 11. A random process {X(t)} has the sample functions of the form
X(t) = A cos(ω t + θ ) where ω is a constant, A is a random variable that has
magnitude +1 and −1 with equal probabilities, and θ is a random variable that
is uniformly distributed in (0, 2π ). Assume that A and θ are independent. Find
the probability density functions of X(t) when A = ±1. Also plot the probability
density functions when A = +1, t = 1, ω = 2, θ = (0, 2π ) and A = +1, t = 1,
ω = 2, θ = (0, 2π ).
S OLUTION :
+ cos(ω t + θ ),
(
A = +1
For A = ±1, we have X(t) =
− cos(ω t + θ ), A = −1
We know that
dθ
fX(t) (x) = f [w (x)]
dx
dθ −1 dθ −1 1
⇒ =p ⇒ = p =p
dx 1 − (x/A)2 dx 1 − (x/A) 2 1 − (x/A)2
dθ 1
For A = ±1, we have =p
dx 1 − (x)2
However, when A = 1, the limit becomes + cos ω t ≤ x ≤ + cos(ω t + 2π )
A = −1, the limit becomes − cos(ω t + 2π ) ≤ x ≤ − cos ω t
1 1
2π √
, cos ω t ≤ x ≤ cos(ω t + 2π ) or
1 − x2
∴ fX(t) (x) = − cos(ω t + 2π ) ≤ x ≤ − cos ω t
0, otherwise
80 • Probability and Random Processes for Engineers
f X (1 ) ( x )
x
− − −
EXERCISE PROBLEMS
1. A random process {X(t)} has the sample functions of the form X(t) =
Y cos ω t where ω is a constant and Y is a random variable that is uniformly
distributed in (0, 1). Sketch three sample functions for Y = 0.25, 0.5, 1 by
fixing ω = 2 without loss of generality. Assume 0 ≤ t ≤ 10.
2. Let the receiver carrier signal of an AM radio be a random process {X(t)}
has the received carrier signal X(t) = A cos(2π f t + θ ) where f is the carrier
frequency with a random phase θ which is a uniform random variable in the
interval (0, 2π ) then what is the expected value of the process{X(t)}.
3. If a random process {X(t)} is sinusoid with a random frequency X(t) =
A sin(ω0t) where A is random variable uniformly distributed over some
interval (0, 1). Then obtain the mean and variance of the process.
4. In an experiment of throwing a fair six-faced dice, the random process
{X(t)} is defined as
sin π t,
(
if odd number shows up
X(t) =
2t, if even number shows up
(i) Find E {X(t)} at t = 0.25 and (ii) Find the probability distribution func-
tion F(x, t) at t = 0.25.
5. Let {X(t)} be a random process with X(t) = cos(ω t + θ ), t ≥ 0 where
ω is a constant and θ is a random variable uniformly distributed in the
interval(−π , π ). Then (i) show that the first and second order moments of
{X(t)} are independent of time and (ii) if θ is constant will the ensemble
mean of {X(t)} be independent of time?
Introduction to Random Processes • 81
3.0 INTRODUCTION
In the previous chapter the concept of random process is explained. In fact, one can
understand from the definition that, a random process is a collection of
random variables, each at different points of time. Due to this reason, random pro-
cesses have all the distributional properties of random variables such as mean, vari-
ance, moments and correlation. When dealing with groups of signals or sequences
(ensembles) it will be important for us to be able to show whether or not these
statistical properties hold good for the entire random process. For this purpose,
and to study the nature of a random process, the concept of stationarity has been
developed. Stationary refers to time invariance of some, or all, statistics of a ran-
dom process, for example, mean, variance, autocorrelation, m th order distribution,
etc. Otherwise, that is, if any of these statistics is not time invariant, then the pro-
cess is said to be nonstationary. Stationarity of a random process can be classi-
fied into two categories: (i) strict sense stationary (SSS) and wide sense stationary
(WSS).
since t1 ≤ t2 we can write t2 = t1 + τ for some τ > 0, therefore, the above equations
can be written as
For the given time points t1 and t2 in the time period (0, t) such that t1 ≤ t2 and for
some τ > 0, a random process {X(t)} is said to be SSS process of order two, if
Similarly, for the given time points t1 ,t2 , · · · · · · , tm in the time period (0, t) such
that t1 ≤ t2 ≤ · · · · · · ≤ tm and for some τ > 0, random process {X(t)} is said to be
SSS process of order m, if
in discrete case or
in continuous case.
It is clear that the distribution of SSS process {X(t)} is independent of time t and
hence it depends only on the time difference τ = (ti + τ ) − ti for 1 ≤ i ≤ m .
By virtue of the property of probability distributions, it may be noted that all
moments E {X r (t)} , r = 1, 2, · · · · · · of SSS process are time independent.
where t1 and t2 are two time points in the time period (0, t) such that 0 < t1 < t2 < t
and τ = t2 − t1
Note:
And in particular, Rxx (0) = E X 2 (t) , which is called the average power
of the process.
(ii) Since τ is the distance between t and t + τ , the function R(τ ) can be written
in the symmetrical form as follows:
n τ τ o
Rxx (τ ) = E X t − X t+ (3.6)
2 2
From the definitions of SSS and WSS processes, it is clear that SSS implies WSS
but the converse is not necessarily true. Therefore, a random process {X(t)} is
said to be stationary (SSS or WSS) if the autocorrelation function is time invari-
ant. That is, Rxx (t1 , t2 ) = Rxx (τ ). Also, in case of a SSS process, all the statistical
properties are independent of time.Due to the time invariance property, in case of
SSS process we have E {X(t)} , E X 2 (t) , and hence V {X(t)} as constants, free
from time (time invariance).
If the ECG recorded at time point t2 is as shown in Figure 3.2 then it is clear that
the distributions of ECG in Figures 3.1 and 3.2 are different and hence we can
conclude that the process is not stationary as it has changed over time.
Stationarity of Random Processes • 85
in discrete case or
fX1 X2 (x1 , x2 ;t1 ,t2 ) = fX1 X2 (x1 , x2 ;t1 + τ , t2 + τ ) in continuous case. (3.7)
(i) E(Xn ) = 0, ∀n
(
E(Xn Xn+s ), for s 6= 0 (3.8)
(ii) Rn (n, n + s) =
E(Xn2 ), for s=0
That is, mean of the sequence {Xn , n ≥ 0} is constant and the autocorrelation func-
tion denoted by Rn (n, n + s) depends only on s, the step (time points) difference.
Rx1 x1 (t1 , t2 ) = E {X1 (t1 )X1 (t2 )} = Rx1 x1 (τ ) for {X1 (t)} to be WSS
Rx2 x2 (t1 , t2 ) = E {X2 (t1 )X2 (t2 )} = Rx2 x2 (τ ) for {X2 (t)} to be WSS (3.9)
It may be noted that the only solution to the equation g (t + s) = g (t) + g (s) can
be given by g (t) = ct with c as constant. That is,
Now, clearly since g (1) = c, we have c = g (1) = E {X(1)} and hence E {X(t)} =
g (t) = ct = E {X(1)}t = µ1t where µ1 = E {X(1)}.
Theorem 3.2: Let {X(t)} be a random process with stationary independent incre-
ments. If X(0) = 0, then V {X(t)} = σ12t and V {X(t) − X(s)} = σ12 (t − s) where
V {X(1)} = σ12 and t > s.
⇒ h (t + s) = V {X(t + s) − X(0)}
Note:
From the results of the Theorems 3.1 and 3.2, it may be noted that, if we assume
that X(0) = 0, then we have
Theorem 3.3: Let {X(t)} be a random process with stationary independent incre-
ments. If X(0) = 0,V {X(t)} = σ12t and V {X(s)} = σ12 s for some t and s, then
C {X(t), X(s)} = Cxx (t, s) = σ12 min(t, s) where V {X(1)} = σ12 .
+ E {X(s) − E [X(s)]}2
1
⇒ Cxx (t, s) = (V {X(t)} +V {X(s)} −V {X(t) − X(s)})
2
1 2
⇒ Cxx (t, s) = σ1 t + σ12 s − σ12 (t − s) = σ12 s
2
σ1 s,
( 2
if t > s
∴ Cxx (t, s) = ⇒ Cxx (t, s) = σ12 min (t, s)
σ12t, if s > t
E {X(t)} = 0
1
R(τ ) = cos ωτ
2
We can plot both X(t) = (+1) cos(2t + π /2) and R(τ ) = (0.5) cos ωτ as shown in
Figure 3.3 and Figure 3.4 respectively. It may be noted that if X(t) = A cos(ω t + θ )
is chosen from any time window, there is no change in the pattern of the plot,
meaning the process is stationary.
Stationarity of Random Processes • 89
0.6 R (τ)
0.4
0.2
τ
0
−15 −10 −5 0 5 10 15
−0.2
−0.4
R(τ) = (0.5)cos2τ
−0.6
SOLVED PROBLEMS
Problem 1. If {X(t)} is a random process with X(t) = Y cost + Z sint for all t
where Y and Z are independent random variables, each of which assumes the val-
ues −2 and 1 with probabilities 1/3 and 2/3 respectively. Prove that {X(t)} is a
stationary process in wide sense but not stationary in strict sense.
S OLUTION :
Since Y and Z are discrete random variables, the probability distribution of random
variable Y can be represented as
Y =y −2 1
P(Y = y) 1/3 2/3
Z=z −2 1
P(Z = z) 1/3 2/3
Since Y and Z are independent random variables, we have the joint probability
distribution as
Y =y P(Y = y)
−2 1
Z=z −2 1/9 2/9 1/3
1 2/9 4/9 2/3
P(Z = z) 1/3 2/3
Consider
1 2
E(Y ) = E(Z) = (−2) + (1) = 0
3 3
1 2
E(Y 2 ) = E(Z 2 ) = (−2)2 + (1)2 = 2
3 3
E(Y Z) = ∑ ∑ yzP (Y = y, Z = z)
y=−2,1 z=−2,1
= ∑ ∑ yzP (Y = y)P (Z = z)
y=−2,1 z=−2,1
1 1 1 2
E(Y Z) = (−2)(−2) + (−2)(1) + (1)(−2)
3 3 3 3
2 1 2 2
+ (1)(1) =0
3 3 3 3
Or
Consider
V {X(t)} = V (Y cost + Z sint) = cos2 tV (Y ) + sin2 tV (Z) = 2 (a constant)
Now, consider
n o
E X 3 (t) = E(Y cost + Z sint)3
n o
= E Y 3 cos3 t +Y 2 Z cos2 t sint +Y Z 2 cost sin2 t + Z 3 sin3 t
+ sin3 tE(Z 3 )
But
1 2
E(Y 3 ) = E(Z 3 ) = (−2)3 + (1)3 = −2
3 3
E(Y 2 Z) = E(Y 2 )E(Z) = 0 and E(Y Z 2 ) = E(Y )E(Z 2 ) = 0
Which is not time invariant as it depends on the time t. But by definition, for a ran-
dom process to be stationary in strict sense, all the moments must be independent
of time. Therefore, the process {X(t)} is not an SSS process.
Problem 2. If {Xn , n ≥ 0}is a sequence of identically and independently dis-
tributed (iid) random variables, each with mean 0 and variance 1, then show that
the sequence {Xn , n ≥ 0} is wide sense stationary.
S OLUTION :
It is given that E(Xn ) = 0, ∀ n
The autocorrelation function is given by
(
E(Xn Xn+s ), for s 6= 0
Rn (n, n + s) =
E(Xn2 ), for s=0
(at)x−1
, x = 1, 2, 3....
P {X(t) = x)} = (1 + at)x+1 .
at
, x=0
1 + at
S OLUTION :
The probability distribution of {X(t)} can be represented as follows:
X(t) = x 0 1 2 3 ..........
2
at 1 at (at)
P {X(t) = x} ·········
1 + at (1 + at)2 (1 + at)3 (1 + at)4
Stationarity of Random Processes • 93
1 2at 3 (at)2
= 0+ + + +······
(1 + at)2 (1 + at)3 (1 + at)3
( 2 )
1 at at
= 1+2 +3 +······
(1 + at)2 1 + at 1 + at
−2
1 at
= 1− =1
(1 + at)2 1 + at
∞
∑ x 2 P {X(t) = x}
n o
E X 2 (t) =
x=0
∞ ∞
= ∑ x(x + 1)P {X(t) = x} − ∑ xP {X(t) = x}
x=0 x=0
Now, consider
∞
1 at
∑ x(x + 1)P {X(t) = x} = 0 + (1)(2) (1 + at)2 + (2)(3) (1 + at)3
x=0
(at)2
+ (3)(4) +······
(1 + at)3
( 2 )
1 at at
= (1)(2) + (2)(3) + (3)(4) +······
(1 + at)2 1 + at 1 + at
−3
1 at
= (2) 1 − = 2(1 + at)
(1 + at)2 1 + at
n o
∴ E X 2 (t) = 2(1 + at) − 1 = 1 + 2at
S OLUTION :
We know that the second moment of the random variable Z = X(t + τ ) − X(t) is
given by
n o
E(Z 2 ) = E [X(t + τ ) − X(t)] 2
Now consider
n o n o
E [X(t + τ ) − X(t)] 2 = E X 2 (t + τ ) + X 2 (t) − 2X(t + τ )X(t)
Since the given random process {X(t)} is a WSS process we know that
n o n o
E X 2 (t + τ ) = E X 2 (t) = R (0)
and
E {X(t + τ ) X (t)} = R (τ )
Therefore, we have
n o
E [X(t + τ ) − X(t)] 2 = R(0) + R(0) − 2R(τ )
= 4 + 4 − 2 4 e−2|τ | = 8 1 − e−2|τ |
S OLUTION :
Given X(t) = A cos(ω t + θ )
Since θ is a uniform random variable distributed in the interval (−π , π ), we have
the PDF of θ as
1
f (θ ) = , −π ≤ θ ≤ π
2π
Stationarity of Random Processes • 95
Consider
Zπ
E {X(t)} = E [A cos(ω t + θ )] = A cos(ω t + θ ) f (θ )d θ
−π
Zπ
1
= A cos(ω t + θ ) dθ
2π
−π
Zπ
A
= cos(ω t + θ )d θ
2π
−π
A
= [sin(ω t + θ )]π−π
2π
A
= [sin(ω t + π ) − sin(ω t − π )] = 0
2π
Consider
= A2 E {cos(ω t1 + θ ) cos(ω t2 + θ )}
A2
= E {cos ω (t1 − t2 ) + cos [ω (t1 + t2 ) + 2θ ]}
2
A2
= {E {cos ω (t1 − t2 )} + E {cos [ω (t1 + t2 ) + 2θ ]}}
2
Zπ Zπ
A2
= cos ω (t1 − t2 ) f (θ )d θ + cos [ω (t1 + t2 ) + 2θ ] f (θ )d θ
2
−π −π
Zπ Zπ
A2 1 1
= cos ω (t1 − t2 ) d θ + cos [ω (t1 + t2 ) + 2θ ] d θ
2 2π 2π
−π −π
A2 A2
= cos ω (t1 − t2 ) − |sin [ω (t1 + t2 ) + 2θ ]|π−π
2 8π
A2 A2
= cos ω (t1 − t2 ) = cos ωτ
2 2
96 • Probability and Random Processes for Engineers
Note that
|sin [ω (t1 + t2 ) + 2θ ]|π−π = sin [ω (t1 + t2 ) + 2π ]
− sin [ω (t1 + t2 ) − 2π ] = 0
A2
∴ Rxx (t1 ,t2 ) = cos ωτ = R(τ )
2
Since mean of the random process {X(t)} is constant and autocorrelation function
is invariant of time, R(t1 ,t2 ) = R(τ ), the process {X(t)} is stationary in wide sense.
Problem 6. If R(τ ) is the autocorrelation function of a wide sense stationary pro-
cess {X(t)} with zero mean, then using Chebyshev’s inequality show that
P {|X(t + τ ) − X(t)| ≥ ε } ≤ 2 {R(0) − R(τ )} /ε 2 for some ε > 0.
S OLUTION :
If X is a random variable, then we know that by Chebyshev’s theorem,
P {|X − E(X)| ≥ ε } ≤ V (X)/ε 2
for some ε > 0. Accordingly, we have
P {|[X(t + τ ) − X(t)] − E [X(t + τ ) − X(t)]| ≥ ε } ≤ V [X(t + τ ) − X(t)] /ε 2
Consider V [X(t + τ ) − X(t)] = V {X(t + τ )} +V {X(t)} − 2Cov(t + τ , t)
Since {X(t)} is a wide sense stationary process with zero mean, we have
n o n o
V [X(t + τ ) − X(t)] = E X 2 (t + τ ) + E X 2 (t) − 2R(τ )
S OLUTION :
(i) Mean:
If S a random variable, then its mean is nothing but the expected value, that
is, E(S).
Rb Rb
Consider E(S) = E X (t) dt = µ dt = (b − a)µ ∵ E {X(t)} = µ
a a
Stationarity of Random Processes • 97
(ii) Variance:
Variance of the random variable S is given by V (S) = E(S2 ) − {E(S)}2
Consider
2
Zb Zb Zb
E(S2 ) = E X (t) dt = E X (t1 ) dt1 X (t2 ) dt2
a a a
Zb Zb
=E X (t1 ) X(t2 )dt1 dt2
a a
Zb Zb
= E {X (t1 ) X(t2 )} dt1 dt2
a a
Zb Zb
= Rxx (t1 ,t2 )dt1 dt2
a a
Z2b
= Rxx (τ )[(b − a) − |τ |]d τ
2a
Zb
V (S) = Rxx (τ )[(b − a) − |τ ]| d τ − [(b − a)µ ]2
a
S OLUTION :
Case (i): When X and Y are not independent.
It is given that
Cov(X, Y )
= t 2V (X) +V (Y ) + 2ρXY σX σY ∵ ρXY =
σX σY
= t 2 σX2 + σY2 + 2ρXY σX σY
+ (σY2 + µY2 )
+ (σY2 + µY2 )
Rzz (t1 ,t2 ) = t1t2 (σX2 + µX2 ) + (t1 + t2 )µX µY + (σY2 + µY2 )
Since mean, variance, autocorrelation and covariance are all not time invariant, the
random process {Z(t)} is not a stationary process.
S OLUTION :
Consider
E {X(t)} = E {cos(at +Y )}
Consider
1 1
E {cos [a (t1 + t2 ) + 2Y ]} = E {cos a(t1 + t2 ) cos 2Y − sin a(t1 + t2 ) sin 2Y }
2 2
1
= {cos a (t1 + t2 )E (cos 2Y ) − sin a(t1 + t2 )E(sin 2Y )}
2
100 • Probability and Random Processes for Engineers
1
⇒ {cos a(t1 + t2 )E(cos 2Y ) − sin a(t1 + t2 )E(sin 2Y )} = 0
2
Consider
1 1
E {cos [a(t2 − t1 )]} = cos [a(t2 − t1 )]
2 2
1 1
∴ R(t1 ,t2 ) = cos [a(t2 − t1 )] = cos [a(τ )] [∵ τ = t2 − t1 ]
2 2
Since mean of the random process {X(t)} is constant and autocorrelation is time
invariant we conclude that the process is a stationary process in wide sense.
Problem 10. Two random processes {X(t)} and {Y (t)} are defined by
Then show that {X(t)} and {Y (t)} are jointly wide-sense stationary, if A and B
are uncorrelated random variables with zero means and equal variances and ω is a
constant.
S OLUTION :
E(A) = E(B) = 0
Consider
= σ 2 cos ωτ [with τ = t1 − t2 ]
Since E {X(t)} = 0 is constant and Rxx (t1 , t2 ) is time invariant, the random pro-
cesses {X(t)} and {Y (t)} are individually wide sense processes.
Now consider
= σ 2 cos ωτ [with τ = t1 − t2 ]
EXERCISE PROBLEMS
1. The random process {X(t)} is defined as X(t) = 2e−At sin(ω t + B)u(t),
where u(t) is the unit step function and the random variables A and B are
independent where A is uniformly distributed in (0, 2) and B is uniformly
distributed in (−π , π ). Verify whether the process is wide sense stationary.
2. Let {X(t)} be random process such that X(t) = sin(ω t + θ ) is a sinusoidal
signal with random phase θ which is a uniform random variable in the inter-
val (−π , π ). If both time t and the radial frequency ω are constants, then
find the probability density function of the random variable X(t) at t = t0 .
Also comment on the dependence of the probability density function of X(t)
and on the stationarity of the process {X(t)}.
3. Consider the random process {X(t)} with X(t) = A(t) cos(2π + θ ), where
the amplitude A(t) is a zero-mean wide sense stationary process with auto-
correlation function RA (τ ) = e−|τ |/2 , the phase θ is a uniform random vari-
able in the interval (0, 2π ), and A(t) and θ are independent. Is {X(t)} a
wide sense stationary process? Justify you answer.
102 • Probability and Random Processes for Engineers
4. A random process {X(t)} has the sample functions of the form X(t) =
A cos(ω t + θ ) where ω is a constant, A is a random variable that has mag-
nitude +1 and −1 with equal probabilities, and θ is a random variable that
is uniformly distributed in (0, 2π ). Assume that A and θ are independent.
(i) Find mean and variance of the random process {X(t)}.
(ii) Is {X(t)} first order strict sense stationary? Give reason for your
answer.
(iii) Find the autocorrelation function of {X(t)}.
(iv) Is {X(t)} wide-sense stationary? Give reasons for your answer.
(v) Plot the sample functions of {X(t)} when A = ±1, t = 1, ω = 2,
θ = 2π .
5. Consider a random process {Y (t)} such that Y (t) = X(t) cos(ω t + θ ), where
ω is a constant, {X(t)} is a wide sense stationary random process, θ is a
uniform random variable in the interval (−π , π ) and is independent of X(t).
Show that {Y (t)} is also a wide sense stationary process.
6. If {X(t)} is a random process with X(t) = A cos(ω t + θ ) where ω is a con-
stant, A is a random variable that has magnitude +1 and −1 with equal
probabilities, and θ is a random variable that is uniformly distributed in
(0, 2π ). Assume that A and θ are independent. The autocorrelation of the
A2
process {X(t)} is given as R(τ ) = cos ωτ . Plot the sample function and
2
autocorrelation when A = +1, t = (0, 10), ω = 2, θ = π .
7. If {X(t)} is a stationary random process and {Y (t)} is another random pro-
cess such that Y (t) = X(t + a) where a is an arbitrary constant, then verify
whether the process {Y (t)} is stationary.
8. Let {X(t)} and {Y (t)} be two independent wide sense stationary processes
with expected values µx and µy and autocorrelation functions Rxx (τ ) and
Ryy (τ ) respectively. Let W (t) = X(t)Y (t), then find µw and Rww (t,t + τ ).
Also, verify whether {X(t)} and {W (t)} are jointly wide sense stationary.
9. Let {X(t)} be a wide sense stationary random process. Verify whether the
processes {Y (t)} and {Z(t)} defined below are wide sense stationary. Also,
determine whether {Z(t)} and each of the other two processes are jointly
wide sense stationary.
(i) Y (t) = X(t + a)
(ii) Z(t) = X(at)
10. Consider the random process {X(t)} such that X(t) = cos(ω t + θ ) where
θ is a uniform random variable in the interval (−π , π ). Show that first and
second order moments of {X(t)} are independent of time. Also find variance
of {X(t)}.
C HAPTER 4
AUTOCORRELATION AND ITS
PROPERTIES
4.0 INTRODUCTION
Autocorrelation function of a random process plays a major role in knowing
whether the process is stationary. In particular, for a stationary random process, the
autocorrelation function is independent of time and hence it becomes dependent
only on time difference. Hence, autocorrelation of a stationary process also helps
to determine the average of the process as the time difference becomes infinite.
Apart from this, the autocorrelation function of a stationary process shows some-
thing about how rapidly one can expect a random process to change as a function of
time. If the autocorrelation function changes slowly (i.e., decays rapidly) then it is
an indication that the corresponding process can be expected to change slowly and
vice-versa. Further, if the autocorrelation function has periodic components, then
the corresponding process is also expected to have periodic components. There-
fore, there is a clear indication that the autocorrelation function contains informa-
tion about the expected frequency content of the random process.
For example, let us assume that the random process {X(t)} represents voltage
in waveform across a resistance of unit ohms. Then, the ensemble average of the
second order moment of {X(t)}, that is E X 2 (t) gives the average power deli-
vered to the resistance of unit ohms by {X(t)} as shown below. That is, the average
power of {X(t)} is
4.1 AUTOCORRELATION
In Chapter 2, we have defined autocorrelation and in Chapter 3 we have studied the
importance of autocorrelation in establishing the stationarity of a random process.
Let us recall that if {X(t)} is a random process and X(t1 ) and X(t2 ) are the two
104 • Probability and Random Processes for Engineers
random variables of the process at two time points t1 and t2 , then the autocorrela-
tion of the process {X(t)} denoted by Rxx (t1 , t2 ) is obtained as the expected value
of the product of X(t1 ) and X(t2 ). That is,
Also, if {X1 (t)} and {X2 (t)} are two random processes observed over a period of
time (0, t) and X1 (t1 ) and X2 (t2 ) are the two random variables respectively of the
process {X1 (t)} at the time point t1 and X2 (t2 ) at the time point t2 then (4.1) is
given as
Rx1 x2 (t1 , t2 ) = E {X1 (t1 )X2 (t2 )} (4.2)
And for two stationary processes {X1 (t)} and {X2 (t)}, we have
a2
= E {cos ω t − cos [ω (t + τ ) + θ ]}
2
a2 a2
⇒ Rxx (τ ) = E (cos ωτ ) − E {cos [ω (2t + τ ) + 2θ ] }
2 2
1
f (θ ) = , 0 ≤ θ ≤ 2π
2π
Z2π Z2π
a2 1 a2 1
∴ Rxx (τ ) = cos ωτ dθ − cos [ω (2t + τ ) + 2θ ] dθ
2 2π 2 2π
0 o
Z2π
a2 a2 1
= cos ωτ − cos [ω (2t + τ ) + 2θ ] dθ
2 2 2π
o
a2
= cos ωτ
2
2
X(t)
t
0
−7 −6 −5 −4 −3 −2 −1 0 1 2 3 4 5 6 7
−1
−2
1 Rxx(τ)
0.5
0
τ
−7 −6 −5 −4 −3 −2 −1 0 1 2 3 4 5 6 7
−0.5
−1
1
Figure 4.2. Graphical representation of Rxx (τ ) = cos 2τ
2
It may be noted that if the autocorrelation function Rxx (τ ) of the random pro-
cess {X(t)} drops (decays) quickly (for example, refer to Figure 4.4), then the
samples of the process (signal) are less correlated which, in turn, means that the
signal has lot of changes over time (for example, refer to Figure 4.3). Such a sig-
nal has high frequency components. If the autocorrelation function Rxx (τ ) drops
slowly (for example, refer to Figure 4.6), then the signal samples are highly cor-
related and such a signal has less high frequency components (for example, refer
to Figure 4.5). Obviously, the autocorrelation function Rxx (τ ) is directly related to
the frequency domain representation of the random process. Note that in Figures
4.4 and 4.6, the autocorrelation Rxx (τ ) is maximum when τ = 0.
X(t)
R(τ)
τ
0
X(t)
R(τ)
τ
0
process, this autocorrelation function holds some important properties which are
given below:
Proof. It is known that given two time points t and t + τ , then the autocorrelation
of the stationary random process {X(t)} is given by
Rxx (τ ) = E{X(t)X(t + τ )}
= E {X(t + τ )X(t)} = Rxx (−τ ) (4.8)
Proof. This can be proved with the help of Cauchy-Schwarz inequality (Refer to
Equation 1.15 of Chapter 1). That is if X and Y are two random variables, then
Now, consider
µx = lim Rxx (τ )
q
τ →∞
Rxx (τ ) = E {X(t)X(t + τ )}
Autocorrelation and Its Properties • 109
It may be noted that as τ → ∞, X(t) and X(t + τ ) become independent and there-
fore, we have
lim Rxx (τ ) = E {X(t)} E {X(t + τ )}
τ →∞
µx = lim Rxx (τ )
q
⇒ (4.11)
τ →∞
On simplification we have
If we let Rxx (h) = Rxx (0), then the right-hand side of (4.12) becomes zero and
obviously, the left-hand size of (4.12) also becomes zero for every τ which yields
the result that
Rxx (τ + h) = Rxx (τ ) (4.13)
Note:
If {X(t)} is a wide sense stationary process with autocorrelation function Rxx (τ )
then
Rxx (0) ≥ 0, Rx (τ ) = Rxx (−τ ), Rxx (0) ≥ Rxx (τ )
Proof. It is known that given two time points with t and t + τ the cross-correlation
of the stationary random processes {X1 (t)} and {X2 (t)} is given by
Property 4.7: If {X1 (t)} and {X2 (t)} are two stationary random processes with
autocorrelation functions Rx1 x1 (τ ) and Rx2 x2 (τ ) respectively and let Rx1 x2 (τ ) be
q
their cross-correlation function, then Rx1 x2 (τ ) ≤ Rx1 x1 (0) Rx2 x2 (0).
Proof. This can be proved with the help of Schwarz’s inequality. That is, if X and
Y are two random variables, then {E(XY )}2 ≤ E(X 2 )E(Y 2 ).
Now, consider
Since E [X12 (t)] = Rx1 x1 (0) and E [X22 (t)] = Rx2 x2 (0)
q
∴ Rx1 x2 (τ ) ≤ Rx1 x1 (0) Rx2 x2 (0) (4.16)
Autocorrelation and Its Properties • 111
Property 4.8: If {X1 (t)} and {X2 (t)} are two stationary random processes, then
1
Rx1 x2 (τ ) ≤ Rx1 x1 (0) + Rx2 x2 (0) .
2
Proof. Let {X1 (t)} and {X2 (t)} be the two stationary random processes with
autocorrelation functions Rx1 x1 (τ ) and Rx2 x2 (τ ) respectively and let Rx1 x2 (τ ) be
their cross-correlation function. Now consider
E {X1 (t) − X2 (t + τ )}2 = E [X12 (t)] + E [X22 (t + τ )] − 2X1 (t)X2 (t + τ )
= Rx1 x1 (0) + Rx2 x2 (0) − 2Rx1 x2 (τ )
Eliminating the third term −E {X1 (t) − X2 (t + τ }2 from right-hand side, we have
2Rx1 x2 (τ ) ≤ Rx1 x1 (0) + Rx2 x2 (0)
1
⇒ Rx1 x2 (τ ) ≤ Rx1 x1 (0) + Rx2 x2 (0) (4.17)
2
Now, consider
E {X1 (t) + X2 (t + τ )}2 = E [X12 (t)] + E [X22 (t + τ )] + 2X1 (t)X2 (t + τ )
= Rx1 x1 (0) + Rx2 x2 (0) + 2Rx1 x2 (τ )
Eliminating the third term −E {X1 (t) − X2 (t + τ }2 from right-hand side, we have
−2Rx1 x2 (τ ) ≤ Rx1 x1 (0) + Rx2 x2 (0)
1
⇒ −Rx1 x2 (τ ) ≤ Rx1 x1 (0) + Rx2 x2 (0) (4.18)
2
From (4.17) and (4.18), we have
1
Rx1 x2 (τ ) ≤ Rx1 x1 (0) + Rx2 x2 (0) (4.19)
2
Property 4.9: If {X1 (t)} and {X2 (t)} are two independent stationary random
processes with mean values E [X1 (t)] = µx1 and E [X2 (t)] = µx2 , then Rx1 x2 (τ ) =
µx1 µx2 .
Proof. If Rx1 x2 (τ ) is the cross-correlation function of the two stationary random
processes {X1 (t)} and {X2 (t)}, then
Rx1 x2 (τ ) = E {X1 (t)X2 (t + τ )}
112 • Probability and Random Processes for Engineers
Property 4.10: If two stationary random processes {X1 (t)} and {X2 (t)} are
orthogonal, then Rx1 x2 (τ ) = 0.
E {X1 (t)X2 (t + τ )} = 0
⇒ Rx1 x2 (τ ) = 0 (4.21)
Hence, the correlation coefficient between X(t1 ) and X(t2 ), denoted by ρxx (τ ), can
be given as
Theorem 4.1: If {X(t)} is a wide sense stationary random process and S is a ran-
ZT ZT ZT
dom variable such that S = X (t) dt, then the variance σS =
2 C(t1 , t2 )dt1 dt2
−T −T −T
Z2T
= C(τ )(2T − |τ |) d τ for some T > o where τ = t1 − t2 or τ = t1 − t2 .
−2T
ZT ZT
= C(t1 , t2 )dt1 dt2
−T −T
ZT ZT Z2T
σS2 = C(t1 , t2 )dt1 dt2 = C(τ ) (2T − |τ |) d τ (4.25)
−T −T −2T
Note
(i) For any arbitrary white noise process (Refer to Section 2.3.4 of Chapter
2 for definition) the autocovariance is given as C(t1 , t2 ) = b(t1 )δ (t1 − t2 )
for b(t) ≥ 0. Therefore, in case of stationary white noise process, we have
C(τ ) = bδ (τ ) with b as constant. In fact, if we can express C(τ ) = bδ (τ ),
where b is constant, then we have the variance as
Z2T
σS2 =b δ (τ ) (2T − |τ |) d τ
−2T
Z2T Z2T
=b δ (τ ) (2T )d τ − b δ (τ ) |τ | d τ = 2T b (4.26)
−2T −2T
Z2T
σS2 = C(τ ) (2T − |τ |) d τ
−2T
Z2T Z2T Za
= C(τ ) (2T )d τ − C(τ ) |τ | d τ ≈ 2T C(τ ) d τ
−2T −2T −a
Autocorrelation and Its Properties • 115
Za Za Za
⇒ C(τ ) (2T )d τ − C(τ ) |τ | d τ ≈ 2T C(τ ) d τ (4.27)
−a −a −a
Za
b= C(τ ) d τ (4.28)
−a
SOLVED PROBLEMS
1 + τ4
Problem 1. Can the function Rxx (τ ) = serve as a valid autocorrela-
1 + τ6
tion function for a continuous time real valued wide sense stationary process
X(t)? Justify.
S OLUTION :
If {X(t)} is a wide sense stationary process with autocorrelation function
Rxx (τ ) then
Consider
1 + τ4
Rxx (τ ) =
1 + τ6
1 + 04
R(0) = =1≥0
1 + 06
1 + (−τ )4 1 + τ 4
R(−τ ) = = = R(τ )
1 + (−τ )6 1 + τ 6
1 + 04 1 + τ4
R(0) = = 1, |R(τ )| = ≤ 1, ∵ τ 4 ≥ τ 6,
1 + 06 1 + τ6
⇒ R(0) ≥ |R(τ )|
1 + τ4
Therefore, Rxx (τ ) = is a valid autocorrelation function.
1 + τ6
116 • Probability and Random Processes for Engineers
Problem 2. Let {X(t)} be a random process and X(t1 ) and X(t2 ) are the two
random variables of the process at two time points t1 and t2 with autocor-
relation function Rxx (t1 , t2 ). If {Y (t)} is another random process such that
Y (t) = X(t1 )+X(t2 ) with autocorrelation function Ryy (t1 , t2 ), then show that
Ryy (t, t) = Rxx (t1 , t1 ) + Rxx (t2 , t2 ) + 2Rxx (t1 , t2 )
S OLUTION :
Consider
E {Y (t)}2 = E {X(t1 ) + X(t2 )}2
n o
= E X 2 (t1 ) + X 2 (t2 ) + 2X(t1 )X(t2 )
This implies that for the determination of E[Y 2 (t)], the average power of
the output process {Y (t)}, the average power of the input process E [X 2 (t)
alone is not sufficient, but the knowledge of the autocorrelation function
Rxx (τ ) is also required.
Since the given random process {X(t)} is a WSS process, we know that
n o n o
E X 2 (t + τ ) = E X 2 (t) = R (0)
and
E {X(t + τ ) X (t)} = R (τ )
Therefore, we have
n o
E [X(t + τ ) − X(t)] 2 = R(0) + R(0) − 2R(τ )
= 4 + 4 − 2 4 e−2|τ | = 8 1 − e−2|τ |
S OLUTION :
We know that since {X(t)} is a wide sense stationary process, and {Y (t)} is
another wide sense stationary random process such that Y (t) = X(t + a) −
X(t − a), the autocorrelation function Ryy (τ ) can be given as
Ryy (τ ) = E {Y (t)Y (t + τ )}
Problem 5. Given that {X(t)} and {Y (t)} are two independent and sta-
tionary random processes. If {Z(t)} is another process such that Z(t) =
aX(t)Y (t), then find Rzz (t, t + τ ).
118 • Probability and Random Processes for Engineers
S OLUTION :
We know that since {X(t)} and {Y (t)} are independent stationary processes,
and Z(t) = aX(t)Y (t), the autocorrelation function Rzz (t, t + τ ) can be given
as
= a 2 Rxx (τ )Ryy (τ )
Problem 6. If there are two stationary random processes {X(t)} and {Y (t)}
such that Z(t) = X(t) +Y (t) then find Rx+y (τ ).
S OLUTION :
We know that since {X(t)} and {Y (t)} are stationary processes, and Z(t) =
X(t) +Y (t), the autocorrelation function Rx+y (τ ) can be given as
= E {[X(t)X(t + τ ) + X(t)Y (t + τ )
+ E {Y (t)Y (t + τ )}
S OLUTION :
It is given X(t) = A sin(ω t + θ )
Since θ is uniformly distributed between −π and π , we have its PDF as
1
f (θ ) = , −π ≤ θ ≤ π
2π
= A2 E {sin (ω t1 + θ ) sin (ω t2 + θ )}
Zπ
A2
= cos ω (t1 − t2 ) f (θ )d θ
2
−π
Zπ
− cos [ω (t1 + t2 ) + 2θ ] f (θ )d θ
−π
Zπ
A2 1
= cos ω (t1 − t2 ) dθ
2 2π
−π
Zπ
A2 1
− cos [ω (t1 + t2 ) + 2θ ] dθ
2 2π
−π
A2 A2
= cos ω (t1 − t2 ) − |sin [ω (t1 + t2 ) + 2θ ]|π−π
2 8π
A2 A2
= cos ω (t1 − t2 ) = cos ωτ
2 2
A2
∴ Rxx (t1 ,t2 ) = cos ωτ = Rxx (τ )
2
120 • Probability and Random Processes for Engineers
(ii) Mean of the random process {Y (t)} where Y (t) = X 2 (t) is given by
n o
E {Y (t)} = E X 2 (t) = Rxx (0)
For our convenience and without loss of generality, we can also write as
follows:
2
Z10 Z10Z10
2
E(S ) = E X(t)dt = E {X(t1 )X(t2 )} dt1 dt2
0 0 0
Autocorrelation and Its Properties • 121
Z10Z10
= R(t1 ,t2 )dt1 dt2
0 0
Z10
= (10 − |τ |)Rxx (τ )d τ
−10
n o Z10
∴ E S2 = (10 − |τ |)Rxx (τ )d τ
−10
Z10 Z10
(10 − |τ )|) 64 + 10e−2|τ | d τ
2
E(S ) = (10 − |τ |) Rxx (τ ) d τ =
−10 −10
Z10
(10 − τ ) 64 + 10e−2|τ | d τ
=2
0
Z10
640 − 64τ + 100e−2τ − 10τ e−2τ d τ
=2
0
= 6495
Problem 9. A stationary zero mean random process {X(t)} has the auto-
2
correlation function Rxx (τ ) = 10e−0.1τ . Find the mean and variance of
1 R5
S= X (t) dt.
50
S OLUTION :
Given {X(t)} is a stationary zero mean random process
2
Autocorrelation function Rxx (τ ) = 10 e−0.1τ
Consider
1 R5 1 R5
E(S) = E X (t) dt = E {X (t)} dt = 0 ∵ E {X(t)} = 0
50 50
122 • Probability and Random Processes for Engineers
Consider
2
1 Z5 1
Z5 Z5
E(S2 ) = E X (t) dt = E {X (t1 ) X(t2 )} dt1 dt2
5 5
0 0 0
Z5 Z5
1
= Rxx (t1 ,t2 )dt1 dt2
5
0 0
Z5
1
= [5 − |τ |] Rxx (τ )d τ
5
−5
RT RT RT
Since Rxx (t1 ,t2 )dt1 dt2 = (T − |τ |)R (τ )d τ (Refer Result A.4.1
0 0 −T
in Appendix A), we have
Z5
1 2
2
E(S ) = (5 − |τ |)(10e−0.1τ )d τ
5
−5
Z5
2
=2 (5 − |τ |)e−0.1τ d τ
−5
Z5
2
=4 (5 − τ )e−0.1τ d τ
0
Z5 Z5
−0.1τ 2 2
= 20 e dτ − 4 τ e−0.1τ d τ
0 0
= I1 − I2
R5 2
It may be noted that the integral part I1 = 20 e−0.1τ d τ can be evaluated
0
using any numerical integration methods.
R5 2
Now, consider, the integral part I2 = 4 τ e−0.1τ d τ
0
τ2 τ
Let u = ⇒ du = d τ ⇒ τ d τ = 5du
10 5
Z 2.5
⇒ I2 = 4 e−u du = 4(0.9179) = 3.6716
0
Z5
2
∴ 2
E(S ) = I1 − I2 = 20 e−0.1τ d τ − 3.6716
0
Autocorrelation and Its Properties • 123
Z5
2
= 20 e−0.1τ d τ − 3.6716 ∵ E(S) = 0
0
S OLUTION :
We know that if {X(t)} is a stationary random process with autocorrelation
function Rxx (τ ), then the mean of the process, say E [X(t)] = µx can be
obtained as
µx = lim Rxx (τ )
q
τ →∞
s s
25τ 2 + 36 25τ 2 (1 + 36/25τ 2 )
µx = lim = lim
τ →∞ 6.25τ 2 + 4 τ →∞ 6.25τ 2 (1 + 4/6.25τ 2 )
s
4(1 + 36/25τ 2 ) √
µx = lim = 4=2
τ →∞ (1 + 4/6.25τ 2 )
= R(0) − {µx }2
25(0) + 36
= − 22 = 5
6.25(0) + 4
Problem 11. Let {X(t)} and {Y (t)} be two random stationary random pro-
cess such that X(t) = 3 cos(ω t + θ ) and Y (t) = 2 cos(ω t + θ − π /2) where
θ is a random variable uniformly distributed in (0, 2π ). Then prove that
q
Rxx (0) Ryy (0) ≥ Rxy (τ )
124 • Probability and Random Processes for Engineers
S OLUTION :
Consider
= {3 cos ω (t + τ ) + (θ )} {3 cos(ω t + θ )}
cos [ω (t + τ ) + ω (t + 2θ )] + cos ωτ
= 9E
2
9 9
= E {cos [ω (t + τ ) + ω (t + 2θ )]} + E {cos ωτ }
2 2
1
f (θ ) = , 0 ≤ θ ≤ 2π
2π
Consider
Z2π
1
E {cos [ω (t + τ ) + ω t + 2θ )]} = cos [ω (t + τ ) + ω t + 2θ )] dθ
2π
0
2π
sin(2ω t + ωτ + 2θ )
1
=
2π 2 0
1
= {sin(2ω t+ωτ + 4π )−sin(2ω t + ωτ )}
4π
1
= {sin(2ω t + ωτ ) − sin(2ω t + ωτ )} = 0
4π
Consider
Z2π
9 9 1
E {cos ω t} = cos ωτ dθ
2 2 2π
0
9
= cos ωτ
2
9 9
∴ Rxx (τ ) = cos ωτ ⇒ Rxx (0) =
2 2
Autocorrelation and Its Properties • 125
Now consider
Ryy (τ ) = E {Y (t + τ )Y (t)}
1
= {sin(2ω t+ωτ +3π )−sin(2ω t+ωτ −π )}
4π
1
= {sin(2ω t + ωτ ) − sin(2ω t + ωτ )} = 0
4π
Consider
Z 2π
1
2E {cos ω t} = 2 cos ωτ dθ
0 2π
= 2 cos ωτ
q
∴ Rxx (0)Ryy (0) ≥ Rxy (τ )
Since,
τ = t1 − t2 or τ = t2 − t1 we have Rxx (0)Ryy (0) ≥ Rxy (τ ) .
p
EXERCISE PROBLEMS
1. Which of the following functions are valid autocorrelation functions for the
respective wide sense stationary processes?
5.0 INTRODUCTION
It is known that random process is associated with probability and probability dis-
tributions. That is, every outcome (i.e., member function) of a random process
is associated to a probability of its happening. For example, as shown in Illus-
trative Example 2.2 in Chapter 2, two member functions X(t) = − sin(1 + t) and
X(t) = sin(1 + t) of a random process {X(t)} can happen as follows:
− sin(1 + t) if tail turns up
X(t) =
sin(1 + t) if head turns up
The random process {X(t)} is called binomial process if X(t) represents the
number of successes, say x, observed by the time t in a sequence of Bernoulli
trials. A Bernoulli trial can be represented by
0 if failure is observed at time t
X(t) =
1 if success is observed at time t
Then we have
p if x observed at time t is success
P {X(t) = x} =
q = 1 − p if x observed at time t is failure
Since the trials are conducted discretely over a period of time, we can also denote
this probability as
for showing the probability of x successes out of n trials conducted at kth step. That
is, X(t), t > 0 is represented by Xk , k = 1, 2, · · · · · · .
Let us suppose that we observe a sequence of random variables assuming value
+1 with probability p and value −1 with probability q = 1 − p then a natural
example is the sequence of Bernoulli trials, say X1 , X2 , X3 , · · · · · · , Xn , · · · · · · ,
each with probability of success equal to p (similar to the probability p of getting
+1) and with probability of failure equal to q = 1 − p (similar to the probability
q = 1 − p of getting −1). Here the partial sum, in fact, Sn = X1 + X2 + X3 + · · · · · · +
Xn , n ≥ 0 with S0 = 0 follows binomial process. This can be thought of as a random
walk of a particle that takes a unit step up and down randomly with Sn = X1 + X2 +
X3 + · · · · · · + Xn representing the position after n th step. Refer to Figure 5.1 for one
of the realizations of Sn = X1 + X2 + X3 + · · · · · · + Xn .
Sn
4
3
2
0 n
0 2 4 6 8 10 12
−1
−2
e−λ λ x
P {X = x} = , x = 0, 1, 2, · · · · · · (5.3)
x!
where the parameter λ > 0 represents the rate of occurrence of events (points).
e−λ ti (λ ti )x
P {X(ti ) = x} = , x = 0, 1, 2, · · · · · · (5.4)
x!
Clearly, X(ti ) is a Poisson random variable and hence {X(t)} is a Poisson process.
Notationally, x occurrences in the time interval (0, ti ) are denoted by X(ti ) = x or
n (0,ti ) = x. Therefore, the number of occurrences, X(ti ) = x or n (0,ti ) = x, in an
interval of length ti − 0 = ti follows Poisson distribution with parameter λ ti > 0
where λ > 0 is the rate of occurrence of events. Obviously, for given two time
points t1 and t2 in the interval (0, t) such that t1 < t2 , if the number of occurrences
is m up to time t1 and the number of occurrences is n up to time t2 , then the number
Binomial and Poisson Processes • 131
It may be noted that while X(t1 ) = m or n (0, t1 ) = m represents there are m occur-
rences in the time interval (0, t1 ), X(t2 ) = n or n(0, t2 ) = n represents there are n
occurrences in the time interval (0, t2 ) and so on. And hence we have
e λ t (λ t)x
P {X(t + s) − X(s) = x} = , x = 0, 1, 2, · · · · · · (5.7)
x!
X(t)
x − occurrences in (0,ti )
or n(0,ti ) = X(ti ) = x
t
0 t1 t2 ... ti ...
Poisson points
X(t)
n(t1,t2 ) = X(t2 ) − X(t1 ) = n − m
n(0,t2 ) = X(t2 ) = n
n(0, t1 ) = X(t1 ) = m
t
0 t1 t2
e λ t (λ t)x
P {X(ti+1 ) − X(ti ) = x} = , x = 0, 1, 2, · · · · · ·
x!
e λ t (λ t)x
P {X(t + s) − X(s) = x} = , x = 0, 1, 2, · · · · · ·
x!
Property 5.2: If the intervals (t1 , t2 ) and (t2 , t3 ) are non-overlapping, then the
random variables n(t1 , t2 ) = X(t2 ) − X(t1 ) and n(t2 , t3 ) = X(t3 ) − X(t2 ) are inde-
pendent. This is true in case of Poisson process and hence Poisson process is a
process with independent increments.
Binomial and Poisson Processes • 133
Property 5.3: For a specific t, it is known that {X(t)} a Poisson random variable
with parameter λ t > 0. Therefore, we have
Mean: E {X(t)} = λ t
n o
E X 2 (t) = λ t + (λ t)2
n o
Variance: V {X(t)} = E X 2 (t) − {E[X(t)]}2 = λ t
λ t1 + λ 2t1t2 , if t1 < t2
(
Autocorrelation: R(t1 ,t2 ) = E {X(t1 )X(t2 )} =
λ t2 + λ 2t1t2 , if t1 > t2
λ t1 , if t1 < t2
Autocovariance: Cxx (t1 , t2 ) = Rxx (t1 , t2 ) − E {X(t1 )X(t2 )} =
λ t2 , if t1 > t2
Theorem 5.1: If {X1 (t)} and {X2 (t)} represent two independent Poisson pro-
cesses with parameters λ1t and λ2t respectively, then the process {Y (t)}, where
Y (t) = X1 (t) + X2 (t), is a Poisson process with parameter (λ1 + λ2 ) t. (That is, the
sum of two independent Poisson processes is also a Poisson process.)
Proof. It is given that {X1 (t)} and {X2 (t)} are two independent Poisson pro-
cesses with parameters λ1t and λ2t respectively, and Y (t) = X1 (t) + X2 (t) therefore
we have
e−λ1t (λ1t)x
P {X1 (t) = x} = , x = 0, 1, 2, · · · · · ·
x!
e−λ2t (λ2t)x
P {X2 (t) = x} = , x = 0, 1, 2, · · · · · ·
x!
n
Consider P {Y (t) = n} = ∑ P {X1 (t) = r} P {X2 (t) = n − r}
r=0
n
e−λ1t (λ1t)r e−λ2t (λ2t)n−r
= ∑ r! (n − r)!
r=0
134 • Probability and Random Processes for Engineers
n
n! (λ1t)r (λ2t)n−r
= e−(λ1 +λ2 )t ∑ (n − r)!
r=0 n! r!
e−(λ1 +λ2 )t n
=
n! ∑ nCr (λ1t)r (λ2t)n−r
r=0
Alternative proof:
Consider Y (t) = X1 (t) + X2 (t)
E {Y (t)} = E {X1 (t) + X2 (t)} = E {X1 (t)} + E {X2 (t)} = λ1t + λ2t = (λ1 + λ2 ) t
V {Y (t)} = V {X1 (t) + X2 (t)} = V {X1 (t)} +V {X2 (t)} = λ1t + λ2t = (λ1 + λ2 ) t
Since mean and variance are equal, we conclude that the sum of two independent
Poisson processes is also a Poisson process with parameter (λ1 + λ2 ) t.
Theorem 5.2: If {X1 (t)} and {X2 (t)} represent two independent Poisson pro-
cesses with parameters λ1t and λ2t respectively, then the process {Y (t)}, where
Y (t) = X1 (t) − X2 (t), is not a Poisson process. (That is, the difference of two inde-
pendent Poisson processes is not a Poisson process.)
Proof. It is given that {X1 (t)} and {X2 (t)} are two independent Poisson processes
with parameters λ1 and λ2 respectively, and Y (t) = X1 (t)−X2 (t) therefore we have
e−λ1t (λ1t)x
P {X1 (t) = x} = , x = 0, 1, 2 · · · · · ·
x!
e−λ2t (λ2t)x
P {X2 (t) = x} = , x = 0, 1, 2, · · · · · ·
x!
n
Consider P {Y (t) = n} = ∑ P {X1 (t) = n + r} P {X2 (t) = r}
r=0
n
e−λ1t (λ1t)n+r e−λ2t (λ2t)r
= ∑ (n + r)! r!
r=0
p n+2r
λ1
n/2 n t λ1 λ2
= e−(λ1 +λ2 )t
λ2 ∑ r! (n + r) !
r=0
Binomial and Poisson Processes • 135
This is not in the form of a probability mass function of Poisson distribution which
implies Y (t) = X1 (t) − X2 (t) is not a Poisson process. Therefore, the difference of
two independent Poisson processes is not a Poisson process.
Alternative proof:
Consider Y (t) = X1 (t) − X2 (t)
E {Y (t)} = E {X1 (t) − X2 (t)} = E {X1 (t)} − E {X2 (t)} = λ1t − λ2t = (λ1 − λ2 ) t
V {Y (t)} = V {X1 (t) − X2 (t)} = V {X1 (t)} +V {X2 (t)} = λ1t + λ2t = (λ1 + λ2 ) t
Since mean and variance are not equal, we conclude that the difference of two
independent Poisson processes is not a Poisson process.
Theorem 5.3: If {X1 (t)} and {X2 (t)} represent two independent Poisson
processes with parameters λ1t and λ2t respectively, then P [X1 (t) = x/ {X1 (t)
λ1
+X2 (t) = n}] is binomial with parameters n and p where p = . That is,
λ1 + λ2
P [X1 (t) = x/{X1 (t) + X2 (t) = n}] = nCx px qn−x
λ1 λ2
where p = and q = 1 − p = .
λ1 + λ2 λ1 + λ2
Proof.
P {(X1 (t) = x) ∩ (X1 (t) + X2 (t) = n)}
Consider P [X1 (t) = x/{X1 (t) + X2 (t) = n}] =
P {X1 (t) + X2 (t) = n}
n! (λ1t)x (λ2t)n−x
=
x! (n − x)! {(λ1 + λ2 )t}n
λ1 λ2
x n−x
= nCx
λ1 + λ2 λ1 + λ2
= nCx p x q n−x
136 • Probability and Random Processes for Engineers
λ1 λ2
where p = and q = 1 − p = .
λ1 + λ2 λ1 + λ2
Therefore, P [X1 (t) = x/{X1 (t) + X2 (t) = n}] is binomial with parameters n
and p
λ1
where p = .
λ1 + λ2
Theorem 5.4: If {X(t)} is a Poisson process with parameter λ t then
t1
P [X(t1 ) = x/{X(t2 ) = n}] is binomial with parameters n and p = . That is, the
t2
conditional probability of a subset of two Poisson events is, in fact, binomial.
Proof. Let t1 and t2 be two time points and let X(t1 ) and X(t2 ) be two random
variables at these time points forming a subset of the Poisson process {X(t)}. Let
t1 < t2 , and consider
P {X(t1 ) = x, n (t1 , t2 ) = n − x}
P [X(t1 ) = x/{X(t2 ) = n}] =
P {X(t2 ) = n}
P {X(t1 ) = x} P {n (t1 ,t2 ) = n − x}
=
P {X(t2 ) = n}
t1
where p = and q = 1 − p
t2
Therefore, P [X(t1 ) = x/{X(t2 ) = n}] is binomial with parameters n and
t1
p = . That is, the conditional probability of a subset of two Poisson events is
t2
binomial.
Proof. Let Ex be the event that “x occurrences are tagged out of n occurrences”.
Then we have
P(Ex )=P {x tagged occurrences/n occurrences in (0, t)} P {n occurrences in (0, t)}
= P {x tagged and (n − x) untagged occurrences out of n occurrences}
P {X(t) = n}
e λ t (λ t)n
( )
n
= Cx p x q n−x , x = 0, 1, 2, · · · · · · n
n!
It may be noted that the event {Y (t) = x} represents the mutually exclusive union
of the events Ex , Ex+1 , Ex+2 , · · · · · · meaning that there should be a minimum of x
occurrences out of which all are tagged, that is the minimum value of n is x.
∞ ∞
e λ t (λ t)n
∴ P {Y (t) = x} = ∑ En = ∑ nCx p x q n−x n!
, x = 0, 1, 2, · · · · · ·
n=x n=x
∞
n! (λ t)n
= eλt ∑ x! (n − x)!
p x q n−x
n!
n=x
∞
(λ t)n
= eλt ∑ x! (n − x)!
p x q n−x
n=x
Theorem 5.6: The time X (waiting time or service time) between the occurrences
of events in a Poisson process with parameter λ x is an exponential.
Or, if there is an arrival at time point t0 and the next arrival is at time point
t1 then the time between these two Poisson points, t0 and t1 given by X = t1 − t0
follows exponential distribution with probability density function
f (x) = λ e−λ x , x > 0.
Proof. We know that the probability of getting k occurrences in the interval (t0 ,t1 )
of length say x = t1 − t0 is Poisson with parameter λ x > 0 and is given by
e−λ x (λ x)k
P {n(t0 , t1 ) = k} = , k = 0, 1, 2, · · · · · · ; x = t1 − t0
k!
138 • Probability and Random Processes for Engineers
P {n(t0 , t1 ) = 0} = e−λ x , x = t1 − t0
Let the first occurrence happen only beyond time point t1 then we have X >
t1 − t0 = x, then clearly X > x. This implies that there are no occurrences in the
interval (t0 , t1 ) (Refer to Figure 5.4). Hence,
= P {n (t0 , t1 ) = 0} = e−λ x , x = t1 − t0
Now, consider
where x = t1 − t0 .
Therefore, the time between arrivals, represented by the random variable X
follows exponential distribution with parameter λ .
X(t)
an occurrence at t1
an occurrence beyond t1
an occurrence at t0
t
t0 t1
SOLVED PROBLEMS
Problem 1. A random process {Yn } is defined by Yn = 3Xn + 1, where {Xn } is a
2 1
Bernoulli process that assumes 1 with probability and 0 with probability . Find
3 3
the mean and variance of {Yn }.
S OLUTION :
Since {Xn } is a Bernoulli process, we have
1 with probability 2/3
Xn =
0 with probability 1/3
∴ E {Xn } = 1(2/3) + 0(1/3) = 2/3
n o
E Xn2 = (1)2 (2/3) + (0)2 (1/3) = 2/3
n o
∴ V {Xn } = E Xn2 − {E[Xn ]}2 = 2/3 − (2/3)2 = (2/3(1 − 2/3) = 2/9
2
Now consider E {Yn } = E {3Xn + 1} = 3E {Xn } + 1 = (3) + 1 = 3.
3
2
V {Yn } = V {3Xn + 1} = 9V {Xn } = (9) =2
9
Problem 2. Let {Xn , n ≥ 1} denote the presence or absence of a pulse at the nth
time instance in a digital communication system or digital data processing system.
If x = 1 represents the presence of a pulse with probability p and x = 0 represents
the absence of a pulse with probability q = 1 − p, then {Xn , n ≥ 1} is a Bernoulli
process {Xn , n ≥ 1} with probabilities defined below
p if x = 1
P {Xx = x} =
q = 1 − p if x = 0
Show that {Xn , n ≥ 1} is a strict sense stationary process. Or otherwise show that
a Bernoulli process {Xn , n ≥ 1} is a strict sense stationary process.
S OLUTION :
In order to prove that {Xn , n ≥ 1} is strict sense stationary, it is enough to
show that the probability distributions of {Xn , n ≥ 1} of different orders are same.
Consider the first order probability distribution of Xn as
{Xn = x} 1 0
P {Xn = x} p q
140 • Probability and Random Processes for Engineers
It may be noted that this first order distribution is same for Xn+τ for some τ > 0
also. That is the first order distribution is time invariant.
Let us consider the second order joint probability distribution of the process
{Xn , n ≥ 1} for some n = r and n = s. Then we have the second order joint prob-
ability distribution P {Xr = r, Xs = s} of Xr and Xs as
Xs
1 0
1 p2 pq
Xr
0 pq q2
It may be noted that this first order joint probability distribution is same for Xr+τ
and Xs+τ also. That is, the second order distribution is time invariant.
Similarly, consider the third order distribution P {Xr = r, Xs = s, Xt = t} of
the process {Xn , n ≥ 1} for some n = r, n = s and n = t as follows:
Xr Xs Xt P {Xr = r, Xs = s, Xt = t}
0 0 0 q3
0 0 1 pq 2
0 1 0 pq 2
1 0 0 pq 2
0 1 1 p 2q
1 0 1 p 2q
1 1 0 p 2q
1 1 1 p3
We can show that the third order joint probability distribution of Xr , Xs and Xt and
the third order joint probability distribution of Xr+τ , Xs+τ and Xt+τ are same. That
is, the third order distribution is time invariant.
Continuing this way, we can prove that the distributions of all orders are time
invariant. Therefore, we conclude that the process {Xn , n ≥ 1} is stationary in
strict sense.
Problem 3. At a service counter customers arrive according to Poisson process
with mean rate of 3 per minute. Find the probabilities that during a time interval
of 2 minutes, (i) exactly 4 customers arrive and (ii) more than 4 customers arrive.
Binomial and Poisson Processes • 141
S OLUTION :
Let {X(t)} be Poisson process with parameter λ t. Then the probability of x
arrivals at time t (or x arrivals in the time interval (0, t)) is given by
e−λ t (λ t)x
P {X(t) = x} = , x = 0, 1, 2, · · · · · ·
x!
It is given that mean arrival rate λ = 3 per minute, that is, in every time interval of
(ti , ti+1 ) = 1 min, i = 0, 1, 2, · · · · · · with t0 = 0, the average arrival rate is λ = 3.
This implies that in the time interval (0, t1 ) = (0, 1) there are λ t1 = (3)(1) = 3
arrivals on the average and in the time interval (0, t2 ) = (0, 2) there are λ t2 =
(3)(2) = 6 arrivals on the average and in the time interval (0, t3 ) = (0, 3) there
are λ t3 = (3)(3) = 9 arrivals on the average and so on (Refer to Figure 5.5).
Therefore, the probability that x customers arrive in the interval of (0, 2)
minutes is
e−3(2) [(3)(2)]x e−6 6x
P {X(2) = x} = P {n (0, 2) = x} = = ,
x! x!
x = 0, 1, 2, · · · · · ·
E{X(t)} = λt =3t
... ti ... t
0 t1 = 1 t2 = 2 t3 = 3
(i) Hence, the probability that exactly 4 customers arrive during a time interval
of (0, 2) minutes is given as
e−6 64
P {X(2) = 4} = = 0.1339
4!
(ii) The probability that more than 4 customers arrive during a time interval of
(0, 2) minutes is given as
P {X(2) ≥ 5} = 1 − P {X(2) ≤ 4}
4
e−6 6x
= 1− ∑
x=0 x!
6 62 63 64
−6
= 1−e 1+ + + + = 0.7174
1! 2! 3! 4!
142 • Probability and Random Processes for Engineers
Problem 4. Suppose that customers arrive at a counter from town A at the rate
of 1 per minute and from town B at the rate of 2 per minute according to two
independent Poisson processes. Find the probability that the interval between two
successive arrivals is more than 1 minute.
S OLUTION :
It is given that {X1 (t)} is Poisson process with parameter λ1t = (1)(1) = 1 and
{X2 (t)} is an independent Poisson process with parameter λ2t = (2)(1) = 2 since
t = 1 minute. Therefore, the average arrival, say λ , of customers at the counter is
also Poisson with parameter
λ = E {X1 (t) + X2 (t)} = E {X1 (t)} + E {X2 (t)}
= λ1t + λ2t = (λ1 + λ2 ) t = (1 + 2)(1) = 3
It may be noted that the interval between two successive Poisson arrivals follows
an exponential distribution with parameter λ . If X is an exponential random vari-
able representing the interval between two successive arrivals, then the required
probability is
Z∞ Z∞ −3x ∞
−λ x e e−3
P {X > 1} = λ e dx = 3e dx = 3 −3x
= 3 0−
−3 1 −3
1 1
= e−3 = 0.0498
S OLUTION :
Let t1 and t2 be two time points and let X(t1 ) and X(t2 ) be two random vari-
ables at these time points forming a subset of the Poisson process {X(t)}. Let
t1 < t2 , and consider
P {X(t1 ) = x, n (t1 ,t2 ) = n − x}
P [X(t1 ) = x/{X(t2 ) = n}] =
P {X(t2 ) = n}
P {X(t1 ) = x} P {n (t1 ,t2 ) = n − x}
=
P {X(t2 ) = n}
(Since {X(t)} and n (t1 , t2 ) = {X(t2 ) − X(t1 )} are independent)
n on o
e−λ t1 (λ t1 )x /x! e−λ (t2 −t1 ) {λ (t2 − t1 )}n−x /(n − x)!
=
e−λ t2 (λ t2 )n /n!
Binomial and Poisson Processes • 143
x
t1 n−x
t1
= nCx 1−
t2 t2
= nCx p x q n−x
t1
where p = and q = 1 − p
t2
Therefore, P [X(t1 ) = x/{X(t2 ) = n}] is binomial with parameters n and
t1
p= .
t2
Consider P {X(2) = 2/X(6) = 6}
⇒ x = 2, n = 6, t1 = 2, t2 = 6
t1 2 1 2
⇒ p= = = , ⇒ q = 1− p =
t2 6 3 3
2 6−2
1 2
∴ P {X(2) = 2/X(6) = 6} = 6C2
3 3
(6)(5) 1 2 2 6−2 24
= = (15) 6
(1)(2) 3 3 3
(15)(16)
= = 0.3292
729
Problem 6. If {X(t)} is a Poisson process such that E {X(9)} = 6 then (a) find the
mean and variance of X(8), (b) find P {X(4) ≤ 5/X(2) = 3} and (c) P {X(4) ≤ 5/
X(2) ≤ 3}.
S OLUTION :
2 16
E {X(t)} = λ t ⇒ E {X(8)} = λ (8) = (8) =
3 3
144 • Probability and Random Processes for Engineers
Similarly,
2 16
V {X(t)} = λ t ⇒ V {X(8)} = λ (8) = (8) =
3 3
P {X(2) = 3, X(4) ≤ 5}
(b) We know that P {X(4) ≤ 5/X(2) = 3} =
P {X(2) = 3}
Now consider P {X(4) ≤ 5/X(2) = 3} which implies that less than or equal
to 5 occurrences have occurred in the interval (0, 4) given that a maximum
of 3 occurrences have occurred in the interval (0, 2). This follows that there
have to be utmost 2 occurrences only in the interval (2, 4). Therefore, the
required probability becomes
P {X(2) = 3} P {n (2, 4) ≤ 2}
P {X(4) ≤ 5/X(2) = 3} =
P {X(2) = 3}
= P {n (2, 4) ≤ 2}
e−λ t (λ t)x
P {n (t1 , t2 ) = x} = , x = 0, 1, 2, 3, · · · · · ·
x!
e−2λ (2λ )x
⇒ P {n (2, 4) = x} = , x = 0, 1, 2, 3, · · · · · ·
x!
2
e−2λ (2λ )x
∴ P {n (2, 4) ≤ 2} = ∑ x!
x=0
2
e−4/3 (4/3)x
= ∑ x!
x=0
Problem 7. Let {X(t)} be the Poisson process with parameter λ t such that
X(t) = 1 if the number of occurrences (Poisson points) is even in the interval (0, t)
and X(t) = −1 if the number of occurrences is odd (this process is known as semi
random process in telegraphic signal studies) obtain the mean and autocorrelation
of the process.
S OLUTION :
Since {X(t)} is the Poisson process with parameter λ t we have
e−λ t (λ t)x
P {X(t) = x} = , x = 0, 1, 2, 3, · · · · · ·
x!
146 • Probability and Random Processes for Engineers
Also we know that the number of occurrences denoted by n (t1 , t2 ) in the interval
(t1 , t2 ) of length t = t2 − t1 > 0 is Poisson with probability mass function
e−λ t (λ t)x
P {n (t1 , t2 ) = x} = , x = 0, 1, 2, 3, · · · · · ·
x!
Now, we have
P {X(t) = 1} = P {n (0, t) = 0} + P {n (0, t) = 2} + P {n (0, t) = 4} + · · · · · ·
Similarly,
P {X(t) = −1} = P {n (0, t) = 1} + P {n (0, t) = 3} + P {n (0, t) = 5} + · · · · · ·
Therefore, the autocorrection Rxx (t1 , t2 )of the process {X(t)} can be obtained as
follows:
Rxx (t1 , t2 ) = ∑ ∑ x1 x2 P {X(t1 ) = x1 , X(t2 ) = x2 }
x1 =1, −1 x2 =1, −1
If we let t = t1 − t2 > 0 and X(t2 ) = 1 then for the even number of Poisson points
in the interval (t1 , t2 ), we have X(t2 ) = 1. This gives
Consider
Consider
Consider
Letting t = t1 − t2 , we have
Problem 8. Let {X(t)} be a Poisson process with parameter λ t and let us sup-
2
pose that each occurrence gets tagged independently with probability p = . If the
3
average rate of occurrence is 3 per minute then obtain the probability that exactly
4 occurrences are tagged in the time interval (0, 2).
S OLUTION :
If we let {Y (t)} as the number of tagged occurrences then we know that
e−λ pt (λ pt)x
P {Y (t) = x} =
x!
2
It is given that λ = 3, p = , t = 2
3
Therefore, the probability that exactly 4 occurrences are tagged in the time
interval (0, 2) can be obtained as
4
−(3) 23 (2) 2
e (3) (2)
3 e−4 44
P {Y (2) = 4} = = = 0.1954
4! 4!
S OLUTION :
Let {X(t)} be a Poisson process with parameter λ t where t is the time between
arrivals. If we let X as the random variable representing the time between arrivals
then it follows exponential distribution whose probability density function is
given by
f (x) = λ e−λ x , x>0
(i) The probability that the time interval between two consecutive arrivals is
more than 1 minute is obtained as
Z∞ Z∞ ∞
−λ x e−2x
P {X > 1} = λe dx = 2e −2x
dx = 2
−2 1
1 1
e−2
= 2 0− = e−2 = 0.1353
−2
(ii) The probability that the time interval between two consecutive arrivals is
between 1 and 2 minutes
Z2 Z2 2
−λ x e−2x
P {1 < X < 2} = λe dx = 2e−2x
dx = 2
−2 1
1 1
= −e−4 + e−2 = 0.1170
(iii) The probability that the time interval between two consecutive arrivals is
less than or equal to 4 minutes
Z4 Z4 4
−λ x e−2x
P {X ≤ 4} = λe dx = 2e −2x
dx = 2
−2 0
0 0
= −e−8 + 1 = 0.9997
Problem 10. Consider a random telegraph signal process {X(t)} in which the
sample function is X(t) = 0 or X(t) = 1. It is supposed that the process starts at
time t = 0 in the zero state X(t) = 0 and then it remains there for a time interval
equal to T1 at which point it switches to the state X(t) = 1 and remains there for
a time interval equal to T2 then switches state again and so on. Find the first order
probability mass function of the process and hence find the mean of the process.
150 • Probability and Random Processes for Engineers
S OLUTION :
Since the given telegraph signal process {X(t)} is binary valued, any sample
will be a Bernoulli random variable. That is, if Xk = X(tk ) is a Bernoulli variable,
then it is required to find the probabilities of Xk = X(tk ) = 0 and Xk = X(tk ) = 1.
Let us suppose that there are exactly n switches in the time interval (0, tk ).
Then Sn = T1 + T2 + · · · · · · + Tn is the random variable representing the time taken
for n switches. We know that
e−λ tk (λ tk )n
P(n switches in (0, tk ) = P {X(tk ) = n} = , n = 0, 1, 2, · · · · · ·
n!
This is Poisson with parameter λ tk .
Therefore, the number of switches in the time interval (0, tk ) follows a Poisson
distribution. Since the sample member X(tk ) = 0 of the random process will be
equal to 0 if the number of switches is even, we have
e−λ tk (λ tk )n
P {X(tk ) = 0} = ∑ P(n switches in (0, tk ) = ∑ n!
n is even n is even
1
= e−λ tk cosh(λ tk ) = 1 + e−2λ tk
2
Similarly, the sample member X(tk ) = 1 of the random process will be equal to 1
if the number of switches is odd, we have
e−λ tk (λ tk )n
P {X(tk ) = 1} = ∑ P(n switches in (0, tk ) = ∑ n!
n is odd n is odd
1
= e−λ tk sinh(λ tk ) = 1 − e−2λ tk
2
Therefore, the probability mass function of the telegraphic signal process can be
described by a Bernoulli distribution given by
X(tk ) = n 0 1
1 1
P {X(tk ) = n} 1 + e−2λ tk 1 − e−2λ tk
2 2
EXERCISE PROBLEMS
1. A random process {Yn } is defined by Yn = aXn +b, where {Xn } is a Bernoulli
process that assumes 1 or 0 with equal probabilities. Find the mean and
variance of {Yn }.
2. If {X1 (t)} is a Poisson process with rate of occurrence λ1 = 2 and {X2 (t)} is
another independent Poisson process with rate of occurrence λ2 = 3. Then
obtain (i) the probability mass function of the random process {Y (t)} where
Y (t) = X1 (t) + X2 (t), (ii) Find P {Y (2) = 5} and (iii) the mean and variance
of {Y (t)} and also (iv) obtain the parameters under these processes when
t = 2.
3. Suppose that customers are arriving at a ticket counter according to a Pois-
son process with a mean rate of 2 per minute. Then, in an interval of 5 min-
utes, find the probabilities that the number of customers arriving is
(i) exactly 3, (ii) greater than 3 and (iii) less than 3.
4. Patients arrive at the doctor’s clinic according to a Poisson process with rate
parameter λ = 1/10 minutes. The doctor will not attend a patient until at
least three patients are in the waiting room. Then
(i) find the expected waiting time until the first patient is admitted to see
the doctor; and
(ii) what is the probability that no patient is admitted to see the doctor in
the first one hour.
5. Let Tn denote the time taken for the occurrence of the n th event of a Poisson
process with rate parameter λ . Let us suppose that one event has occurred
in the time interval (0, t). Then obtain the conditional distribution of arrival
time T1 over (0, t).
6. Let Tn denote the time taken for the occurrence of the n th event of a Poisson
process with rate parameter λ . Let us suppose that one event has occurred
in the time (0, 10) interval. Then obtain P {T1 ≤ 4/X(10) ≤ 1}.
7. It is given that {X1 (t)} and {X2 (t)} represent two independent Poisson
processes and X1 (2) and X2 (2) are random variables observed from these
processes at t = 2 with parameters 6 and 8 respectively. Then obtain
P {Y (2) = 1}, where Y (2) = X1 (2) + X2 (2).
8. Suppose the arrival of calls at a switch board is modeled as a Poisson process
with the rate of calls per minute being λ = 0.1. Then
(i) What is the probability that the number of calls arriving in a 10 minutes
interval is less than 3?
(ii) What is the probability that one call arrives during the first 10 minutes
interval and two calls arrive during the second 10 minutes interval?
152 • Probability and Random Processes for Engineers
9. If {X(t)} is a Poisson process such that E {X(8)} = 6 then (a) find the mean
and variance of X(7), (b) find P {X(3) ≤ 3/X(1) ≤ 1}.
10. Let {X(t)} be a Poisson process with parameter λ t and let us suppose that
3
each occurrence gets tagged independently with probability p = . If the
4
average rate of occurrence is 4 per minute then obtain the probability that
exactly 3 occurrences are tagged in the time interval (0, 3).
C HAPTER 6
NORMAL PROCESS
(GAUSSIAN PROCESS)
6.0 INTRODUCTION
Due to the nature of normal distribution, all processes can be approximated to
normal process (also called Gaussian Process). In fact, as discussed in the previous
chapter, the random processes following any standard statistical distributions are
known as special random processes. The Gaussian process plays an important
role in random process because it is a convenient starting point for many studies
related to electrical and computer engineering. Also, in most of the situations, the
Gaussian process is useful in modeling the white noise signal observed in practice
which can be further interpreted as a filtered white Gaussian noise signal. For a
definition of white noise process, readers are referred to Section 2.3.4 in Chapter
2. In this chapter, we study in detail the aspects of normal process. Throughout
this book, normal process and Gaussian process are interchangeably used. Some
processes depending on stationary normal process are also studied. In addition, in
this chapter, the processes such as random walk process and Weiner process are
also considered.
f (x1 , x2 , · · · , xi , · · · , xn ; t1 , t2 , · · ·ti , · · · , tn )
n n
T
− 1 ∑ ∑ |Σ|i j (xi −µ (ti )) x j −µ (t j )
2|Σ| i=1 j=1
e
= , −∞ < xi < ∞, ∀ i (6.1)
(2π ) n/2 |Σ|1/2
154 • Probability and Random Processes for Engineers
where µ (ti ) = E {X(ti )}, Σ is the nth order square matrix (called variance-
covariance matrix) with elements σ (ti ,t j ) = Cxx (ti ,t j ) and |Σ|i j is the cofactor
of σ (ti , t j ) in |Σ|. Refer to Section 1.8 of Chapter 1 for derivation of probability
density function of n–dimensional normal random variables.
It may be noted that the covariance of two random variables X(ti ) and X(t j )
observed at time points ti and t j respectively is given by
It is clear that the normal process involves mean, variance and correlation as
parameters that may or may not depend on time. Therefore, if mean and variance
of a normal process are constants and correlation coefficient is time invariant, then
we can conclude that the process is stationary in strict sense.
where µ (t) is the mean and σ (t) is the standard deviation of the normal process.
It is clear that the normal process involves mean and variance as parameters
that may or may not depend on time. Therefore, if mean and variance of a normal
process are constants then we can conclude that the process is stationary in strict
sense.
Normal Process (Gaussian Process) • 155
X (t)
t
t0 t1
Figure 6.1 represents the normal distribution of the random process {X(t)} at time
point t = t1 . This means that at a particular time point, say t = t1 , we have a ran-
dom variable X(t1 ) which is normally distributed with some mean, say µ (t1 ) and
variance, say σ 2 (t1 ). From any given process, if we observe that µ (t1 ) = µ (t2 ) =
µ (t3 ) = · · · · · · = µ and σ 2 (t1 ) = σ 2 (t2 ) = σ 2 (t3 ) = · · · · · · = σ 2 , meaning that the
mean and variance are constants being independent of time, we say that the process
is stationary in strict sense. It may be noted that as many samples of the process
{X(t)} will be close enough to the mean of the normal distribution.
(6.4)
− ∞ < x1 , x2 < ∞
where µ (t1 ) and µ (t2 ) are the means and σ (t1 ) and σ (t2 ) are the standard devia-
tions of the random variables X(t1 ) and X(t2 ) respectively. Refer to Section 1.8.4
of Chapter 1 for derivation of probability density function of two-dimensional nor-
mal random variables.
In this case the normal process involves mean, variance and correlation as
parameters that may or may not depend on time. Therefore, if mean and variance
of a normal process are constants and correlation coefficient is time invariant, then
we can conclude that the process is stationary in strict sense.
156 • Probability and Random Processes for Engineers
If X(t1 ) and X(t2 ) are independent, then ρ (t1 , t2 ) = 0, and hence we have
( )
x1 −µ (t1 ) 2 x2 −µ (t2 ) 2
− 12 +
σ (t1 ) σ (t2 )
e
f (x1 , x2 ; t1 , t2 ) = √ ,
2π σ (t1 ) σ (t2 )
− ∞ < x1 , x2 < ∞ (6.5)
is called standard normal process with mean 0 and variance 1. That is, E {Z(t)}=0
and V {Z(t)} = 1.
(iv) If X(t) and Y (t) are two random variables of the normal processes {X(t)}
and {Y (t)} with zero means, then we have
n o
E X 2 (t)Y 2 (t) = E[X 2 (t)]E[Y 2 (t)] + 2 {E[X(t)Y (t)]}2 (6.8)
(v) If X(t) and Y (t) are two random variables of the normal processes {X(t)}
and {Y (t)} with zero means, variances σx2 and σy2 , and correlation coeffi-
cient ρxy (t1 ,t2 ), then we have
1 1 −1
P {X(t1 )Y (t2 ) = positive} = + sin ρxy (t1 ,t2 ) (6.9)
2 π
which gives the probability that X(t) and Y (t) are of same signs and,
1 1 −1
P {X(t1 )Y (t2 ) = negative} = − sin ρxy (t1 ,t2 ) (6.10)
2 π
which gives the probability that X(t) and Y (t) are of different signs.
(vi) If X(t) and Y (t) are two random variables of the normal processes {X(t)}
and {Y (t)} with zero means, variances σx2 and σy2 , and correlation coeffi-
cient, ρxy (t1 ,t2 ) then we have
2
E {|X(t1Y (t2 ))|} = σx σy (cos α + α sin α ) (6.11)
π
where
Rxy (t1 ,t2 )
sin α = ρxy (t1 ,t2 ) = (6.12)
σx σy
Note:
If X(t1 ) and X(t2 ) are two random variables of the same normal process
observed at time points t1 and t2 , with zero means, variances σx2 and σy2 ,
and correlation coefficient, ρxy (t1 ,t2 ), then we have
Rxx (τ ) Rxx (τ )
ρxx (τ ) = = = sin α (6.13)
σx2 Rxx (0)
158 • Probability and Random Processes for Engineers
If {X(t)} is a zero mean stationary normal process and if Y (t) = X 2 (t) then the
process {Y (t)} is called a square-law detector process.
Let the normal process {X(t)} be stationary with mean E {X(t)} = 0, variance
V {X(t)} = σx2 (say) and autocorrelation function Rxx (τ ). Now,
n o
E {X(t)} = 0, ⇒ V {X(t)} = σx2 = E X 2 (t) = Rxx (0)
n o
∴ E {Y (t)} = E X 2 (t) = Rxx (0)
n o
∴ V {Y (t)} = σy2 = E Y 2 (t) − {E [Y (t)]}2 = 3R2xx (0) − R2xx (0) = 2R2xx (0)
Also Cyy (τ ) = Ryy (τ )−E {Y (t1 )} E {Y (t2 )} = R2xx (0)+2R2xx (τ )−R2xx (0) = 2R2xx (τ )
Therefore, {Y (t)} is a wide sense stationary process with mean and autocor-
relation as given below:
Z∞ Z∞
1 2 2 2 2 /2σ 2
∴ E {Y (t)} = E {|X(t)|} = |x| √ e−x /2σx dx = √ xe−x x dx
2πσx 2πσx
−∞ 0
x2
Let v = ⇒ x dx = σx2 dv
2σx2
Z∞ r Z∞ r Z∞
2 −x2 /2σx2 2 2
⇒ √ xe dx = σx −v
e dv = σx ∵ e−v dv = 1
2πσx π π
0 0 0
r r r
2 2 2 2
∴ E {Y (t)} = σx = σ = Rxx (0)
π π x π
Consider Ryy (t1 , t2 ) = E {Y (t1 )Y (t2 )}
= E {|X(t1 )| |X(t2 )|} = E {|X(t1 )X(t2 )|}
2 2
= σ (cos α + α sin α ) (Refer to Equation (6.11))
π x
where sin α = ρxx (t1 , t2 )
Cov {X(t1 )X(t2 )} E {X(t1 )X(t2 )}
We know that ρ (t1 , t2 ) = p =
σx2
p
V {X(t1 )} V {X(t2 )}
∵ E {X(t)} = 0
E {X(t1 )X(t2 )} Rxx (t1 , t2 )
⇒ ρxx (t1 , t2 ) = =
σx2 σx2
Since the process {X(t)} is stationary, we have
Rxx (τ ) Rxx (τ )
⇒ ρxx (τ ) = = = sin α (Refer to Equation (6.13))
σx2 Rxx (0)
Therefore, {Y (t)} is wide sense stationary process with autocorrelation function
2
Ryy (τ ) = Rxx (0)(cos α + α sin α )
π
160 • Probability and Random Processes for Engineers
2 n π o
Consider E Y 2 (t) = Ryy (0) = Rxx (0) 0 + (1) = Rxx (0)
π 2
π
Because when τ = 0 we have sin α = 1 ⇒ α =
2 !2
r
2 2 2 2
Hence, V {Y (t)}=E Y (t) −{E[Y (t)]} =Rxx (0)− Rxx (0) = 1− Rxx (0)
π π
Therefore, {Y (t)} is wide sense stationary process with mean and autocorrelation
as given below:
r r r
2 2 2 2
E {Y (t)} = σx = σ = Rxx (0) (constant)
π π x π
2
Ryy (τ ) = Rxx (0)(cos α + α sin α ), (time invariant), (6.15)
π
Let the normal process {X(t)} be stationary with mean E {X(t)} = 0, variance
V {X(t)} = σx2 (say) and autocorrelation function Rxx (τ ). Now,
n o
E {X(t)} = 0, ⇒ V {X(t)} = σx2 = E X 2 (t) = Rxx (0)
We know that
X(t1 ) + + − −
X(t2 ) + − + −
X(t1 )X(t2 ) + − − +
1 1
∴ P {X(t1 )X(t2 ) = +} = , P {X(t1 )X(t2 ) = −} =
2 2
1 1 1
⇒ E {X(t1 )Y (t2 )/X(t1 )X(t2 )} = {X(t1 )X(t2 ) + |X(t1 )X(t2 )|} + (0)
2 2 2
1
= {X(t1 )X(t2 ) + |X(t1 )X(t2 )|}
4
1
∴ E {Z(t1 )Z(t2 )} = {E {X(t1 )X(t2 )} + E {|X(t1 )X(t2 )|}}
4
1
⇒ Rzz (t1 , t2 ) = Rxx (t1 , t2 ) + Ryy (t1 ,t2 )
4
where Ryy (t1t2 ) is the autocorrelation of the full-wave linear detector process
Y (t) = |X(t)|.
Since the processes {X(t)} and {Y (t)} are stationary, we have
Rxx (τ )
1 2
∴ Rzz (τ ) = Rxx (τ ) + Rxx (0)(cos α + α sin α , sin α =
4 π Rxx (0)
π
τ = 0 ⇒ sin α = 1 but cos2 α = 1 − sin2 α ⇒ cos α = 0 ⇒ α =
2
π
1 2 1
∴ Rzz (0) = Rxx (0) + Rxx (0) = Rxx (0)
4 π 2 2
n o 1
⇒ E Z 2 (t) = Rzz (0) = Rxx (0)
2
1
Consider E {Z(t)} = {E {X(t)} + E {|X(t)|}}
2
r
1 1 2
= {0 + E {Y (t)}} = Rxx (0)
2 2 π
r
1
= Rxx (0) ∵ Y (t) = |X(t)|
2π
n o
∴ V {Z(t)} = E Z 2 (t) − {E[Z(t)]}2
r !2
1 1 1 1
= Rxx (0) − Rxx (0) = 1− Rxx (0)
2 2π 2 π
162 • Probability and Random Processes for Engineers
Z∞ Z∞
1 2 2 1 2 /2σ 2
E {Z(t)} = E {X(t)} = x√ e−x /2σx dx = √ xe−x x dx
2πσx 2πσx
0 0
x2
Let v = ⇒ xdx = σx2 dv
2σx2
Z∞ Z∞ Z∞
2 −x2 /2σx2 1 1
⇒ √ xe dx = √ σx
e dv = √ σx −v
e−v dv = 1
2πσx 2π 2π
0 0 0
r r
1 1 2 1
∴ E {Z(t)} = √ σx = σ = Rxx (0)
2π 2π x 2π
o Z∞ 1 2 2
e−x /2σx dx
n o n
∴ E Z 2 (t) = E X 2 (t) = x2 √
2πσx
0
Z∞
1 2 /2σ 2
=√ x2 e−x x dx
2πσx
0
x2
Let v = ⇒ xdx = σx2 dv
2σx2
Z∞ Z∞
!
2 2 −x2 /2σx2 1 σ2
⇒ √ x e dx = √ (2vσx2 )e−v p x dv
2πσx 2πσx 2σx2 v
0 0
Z∞
σ2 √ −v 1
= √x v e dv = σx2
π 2
0
Z∞
3 1 1√
∵ v 2 −1 e−v dv = Γ(3/2) = Γ(1/2) = π
2 2
0
n o 1 1
∴ E Z 2 (t) = σx2 = Rxx (0)
2 2
Hence,
n o
V {Z(t)} = E Z 2 (t) − {E[Z(t)]}2
r !2
1 1 1 1
= Rxx (0) − Rxx (0) = 1− Rxx (0)
2 2π 2 π
Normal Process (Gaussian Process) • 163
Therefore, {Z(t)} is wide sense stationary process with mean and autocorrelation
as given below:
r
1
E {Z(t)} = Rxx (0) (constant)
2π
1 2
Rzz (τ ) = Rxx (τ ) + Rxx (0)(cos α + α sin α (time invariant) (6.16)
4 π
1 1
∴ E {Y (t)} = E {X(t)} = (+1) + (−1) = 0
2 2
n o n o 1 1
E Y 2 (t) = E X 2 (t) = (+1)2 + (−1)2 = 1
2 2
Meaning that the white noise has infinite average power, which is physically not
possible. Notwithstanding, it is useful because any Gaussian noise signal observed
in a real system can be interpreted as a filtered white Gaussian noise signal with
finite power.
1
⇒ P {X(t = +d} = P {X(t = −d} =
2
If the coin is tossed n times, then there could be k heads and n − k tails in the total
time of t = nT . Therefore, the distance covered by the process is
kd, ahead for heads
X(nT ) =
(n − k)d, backward for tails
That is, after n tosses, the total distance between the origin to the present position
of the process is
X(nT ) = kd − (n − k)d = (2k − n)d
P {X = k} = n Ck pk qn−k , k = 0, 1, 2, · · · · · · , n
whose mean is np and variance is npq the same can be approximated as a normal
distribution with mean µ = np and variance σ 2 = npq. That is, the random variable
X ∼ N(np, npq) which implies
2
k−np
1 −1 √
n
Ck pk qn−k ∼
=√ e 2 npq
2π npq
But
!2
k−n/2
k n−k − 21 √
∼ n 1 1 1 n/4
P {X(nT ) = (2k − n)d} = Ck =√ p e ,
2 2 2π n/4
for (2k − n)d = −nd, (−n + 2)d, (−4 − n)d, · · · · · · , (−2 + n)d, nd
⇒ Rxx (t2 , t1 ) = ω t1
Rxx (t2 , t1 ) = ω t2
E {X(t1 )} = E {X(t2 )} = 0
168 • Probability and Random Processes for Engineers
SOLVED PROBLEMS
Problem 1. Given a normal process with mean 0, autocorrelation function R(τ ) =
4 e−3|τ | , where τ = t1 −t2 , and the random variables Y = X(t +1) and W = X(t −1)
then find
(a) (i) E(YW ), (ii) E (Y +W )2 and (iii) the correlation coefficient
between Y and W .
(b) (i) Find probability density function f (y), (ii) cumulative probability
P(Y < 1) and (iii) the joint probability density function f (y, w).
(a) S OLUTION :
(i) Consider
E(YW ) = E [X(t + 1)X(t − 1)]
= R(t + 1, t − 1)
= 2 [R(0) + R(2)]
n o
= 2 4 + 4e−6 = 8.0198
(iii) Consider
Cov (Y,W ))
ρZW = p p
V (Y ) V (W )
E (YW ) − E(Y ) E(W )
=q q
E(Y ) − [E(Y )] E(W 2 ) − [E(W )]2
2 2
E (YW )
=p p (Since E(Y ) = E(W ) = 0)
E(Y 2 ) E(W 2 )
E (YW ) E(YW ) 0.0099
=p p = = = 0.00248
R(0) R(0) R(0) 4
Normal Process (Gaussian Process) • 169
(b) S OLUTION :
(i) It is given that mean E(Y ) = µy = 0
√
We know that standard deviation σy = R(0) = 4 = 2
p
This implies that the random variable Y is normal with mean 0 and standard
deviation 2.
Therefore, the probability density function of Y is given as
( )
1 y − µy 2
1
f (y) = √ exp −
2π σ 2 σy
2
1 −y
= √ exp for − ∞ < y < ∞
2 2π 8
(ii) Consider
Therefore, we have
!
1 − E {X(t + 1)} 1
P(Z < z) = P Z < p =P Z< = 0.6915
E {X 2 (t + 1))} 2
(iii) We have shown that Y is standard normal variable with mean 0 and standard
deviation 2. Similarly, we can show that W is also standard normal variable
with mean 0 and standard deviation 2. Since it is not known that Z and W
are uncorrelated random variables, the correlation, r, between these random
variables can be obtained as follows:
Problem 2. Suppose that {X(t)} is a random process with µ (t) = 3 and C(τ ) =
4e−0.2|τ | , where τ = t1 − t2 . Find (i) P [X(5) ≤ 2] and (ii) P [|X(8) − X(5)| ≤ 1]
using central limit theorem.
S OLUTION :
It is given that E {X(t)} = µ (t) = 3 and V {X(t)} = C(0) = 4e−0.2|0| = 4
(i) Consider
!
X(5) − E {X(5)} 2 − E {X(5)}
P [X(5) ≤ 2] = P p < p
V {X(5)} V {X(5)}
!
2 − µ (5)
2−3
=P Z< p =P Z< √
C(0) 4
(ii) Consider
= P (−0.526 ≤ Z ≤ 0.526)
= 0.40
S OLUTION :
It is given that {X(t)} is a Gaussian process with E {X(t)} = µ (t) = 10 and
V {X(t)} = C(0) = 16e−0.2|0| = 16
(i) Consider
!
X(10) − E {X(10)} 8 − E {X(10)}
P [X(10) ≤ 8] = P p < p
V {X(10)} V {X(10)}
!
8 − µ (10)
8 − 10
=P Z< p =P Z< √
C(0) 16
(ii) Consider
S OLUTION :
It is given that Y (t) = X 2 (t) and E{X(t)} = 0
We know that Cyy (t1 , t2 ) = Ryy (t1 , t2 ) − E {Y (t1 )} E {Y (t2 )}
= E {Y (t1 )Y (t2 )} − E {Y (t1 )} E {Y (t2 )}
n o n o n o
= E X 2 (t1 )X 2 (t2 ) − E X 2 (t1 ) E X 2 (t2 )
n o n o n o
But E X 2 (t1 )X 2 (t2 ) = E X 2 (t1 ) E X 2 (t2 ) + 2 {E[X(t1 )X(t2 )]}2
n o n o
∴ Cyy (t1 , t2 ) = E X 2 (t1 ) E X 2 (t2 ) + 2 {E[X(t1 )X(t2 )]}2
n o n o
− E X 2 (t1 ) E X 2 (t2 )
Problem 5. If {X(t)} is a zero mean stationary Gaussian process with mean µ (t) =
0 and autocorrelation function Rxx (τ ) = 4e−3|τ | then find a system g(x) such that
the first order density f (y;t) of the resulting output Y (t) = g {X(t)} is uniform in
the interval (6, 9).
Normal Process (Gaussian Process) • 173
S OLUTION :
It is given that E{X(t)} = µ (t) = 0 and Rxx (τ ) = 4e−3|τ |
Therefore, {X(t)} is normal with mean 0 and variance 4 and hence the first order
probability density function becomes
1 x2
f (x, t) = √ e− 8 , −∞ < x < ∞
2 2π
Since Y (t) is uniform in the interval (6, 9), we have
1
f (y, t) = , 6<y<9
3
It may be noted that x and y are the realizations of X(t) and Y (t) respectively.
We know that the probability density function f (y, t) of Y (t) can also be
expressed as
f (g(y);t)
f (y;t) =
|J|
Where x = g(y) and J = g′ (y)
f (g(y);t) 3 2
⇒ g′ (x) = = (3) f (g(y);t) = √ e−x /8
f (y;t) 2 2π
3 1
Z
2
⇒ g(x) = √ e−x /8 +C (1)
2 2π
Alternative proof:
It is given that Y (t) = g {X(t)} ⇒ X(t) = g−1 {Y (t)}
Consider the cumulative distribution function
n o
F(y; t) = P {Y (t) ≤ y} = P {g[X(t)] ≤ y} = P X(t) ≤ g−1 (y)
= P {X(t) ≤ x} = F(x; t)
Ry
We know that F(y;t) = f (y, t)dy = 13 (y − 6)
6
1
∴ {g(x) − 6} = F(x;t) ⇒ g(x) = 6 + 3F(x;t)
3
Zx
3 1 2
= 6+ √ e−x /8 dx (2)
2 2π
−∞
174 • Probability and Random Processes for Engineers
Problem 6. It is given that {X(t)} is a random process such that X(t) = Y cos ω t +
W sin ω t, where ω is constant and Y and W are two independent normal random
variables with E(Y ) = E(W ) = 0 and E(Y 2 ) = E(W 2 ), then prove that {X(t)} is a
stationary process of order 2.
S OLUTION :
It is given that E(Y ) = E(W ) = 0 and E(Y 2 ) = E(W 2 ) ⇒ V (Y ) = E(W )
Let V (Y ) = E(W ) = σ 2
Since X(t) is a linear combination of two independent random variables Y and
W , we know that X(t) is also a normal random variable with mean and variance
given as
E {X(t)} = E {Y cos ω t +W sin ω t} = cos ω tE(Y ) + sin ω tE(W ) = 0
∵ E(Y ) = E(W ) = 0
− ∞ < x1 , x2 < ∞
where ρ (t1 , t2 ) is the correlation coefficient between X(t1 ) and X(t2 ) which is
given by
Cxx (t1 , t2 ) E {X(t1 )X(t2 )}
ρ (t1 , t2 ) = p =
σ2
p
V {X(t1 )} V {X(t2 )}
Sine mean and variance of the random process {X(t)} are constants, the joint
density probability function f (x1 , x2 ;t1 , t2 ) of X(t1 ) and X(t2 ) depends only on
ρ (t1 , t2 ) as it is a function of the time points t1 and t2 . We know that the random
process {X(t)} is stationary in second order if
Normal Process (Gaussian Process) • 175
f (x1 , x2 ; t1 , t2 ) = f (x1 , x2 ; t1 + τ , t2 + τ )
Which implies that the second order probability density function is time invariant.
Clearly, it is true if ρ (t1 , t2 ) = ρ (t1 + τ , t2 + τ ) = ρ (τ ), where τ = t1 − t2 or
τ = t2 −t1 . Therefore, in order to show that the random process {X(t)} is stationary
in second order, if is sufficient to show that the correlation coefficient between
X(t1 ) and X(t2 ) is time invariant. That is,
ρ (t1 , t2 ) = ρ (t1 + τ , t2 + τ )
Now, consider
1
ρ (t1 , t2 ) = E {(Y cos ω t1 +W sin ω t1 ) (Y cos ω t2 +W sin ω t2 )}
σ2
1 cos ω t1 cos ω t2 E(Y 2 ) + sin ω t sin ω t E(W 2 )
1 2
= 2
σ + (cos ω t1 sin ω t2 + sin ω t1 cos ω t2 ) E(YW )
S OLUTION :
Since {X(t)} is a zero mean Gaussian random process, we have
Given E {X(t)} = 0 and E {X(t + τ )} = 0
Rxx (τ ) = 4e−2|τ |
n o
∴ V {X(t)} = E X 2 (t) − {E[X(t)]}2 = 4 − 02 = 4
V (Y ) = V (W ) = V {X(t)} = 4
Consider
Now we know that the joint density function f (y, w) can be obtained as
1 1
2 2
f (y, w) = √ exp − y − 2ryw + w
2π σ 2 1 − r2 2σ 2 (1 − r2 )
1 1
2 −2|τ | 2
= exp − y −2(e )yw+w
2π (4) 1−e−4|τ | 2(4)(1−e−4|τ |)
p
1 1
2 −2|τ | 2
= exp − y − 2(e )yw + w
8π 1 − e−4|τ | 8(1 − e−4|τ | )
p
1 1 2 1 1 2
∴ f (y, w) = √ e− 8 y √ e− 8 w
2π (2) 2π (2)
This shows that Y = X(t) and W = X(t + τ ) are two independent normal random
variables as τ → ∞.
Normal Process (Gaussian Process) • 177
Problem 8. Find the mean and variance of the simple random walk given by
{Xn , n ≥ 0} where Xn = ∑ni=1 Yi , n = 1, 2, 3, · · · · · ·, X0 = 0 and Y1 , Y2, ······ are
independently identically distributed random variables with P {Yn = 1} = p and
P {Yn = −1} = q, p + q = 1 for all n.
S OLUTION :
It is given that
n
Xn = ∑ Yi , n = 1, 2, 3, · · · · · ·
i=1
⇒ Xn = Xn−1 +Yn , n = 1, 2, · · · · · ·
Here, X0 = 0 and Y1 , Y2, ······ are independently identically distributed random vari-
ables with
P {Yn = 1} = p and P {Yn = −1} = q, p + q = 1 for all n.
Now, from Xn = Xn−1 +Yn , n = 1, 2, · · · · · · we have
X1 = X0 +Y1 = Y1
X2 = X1 +Y2 = Y1 +Y2
And so on
Xn = Y1 +Y2 + · · · · · · +Yn
Now,
E(Xn ) = E(Y1 +Y2 + · · · · · · +Yn ) = nE(Yi ), i = 1 or 2 or 3 · · · · · ·
Consider
E(Yi ) = {(1)(p) + (−1)q} = (p − q) = (2p − 1)
∴ E {Xn } = nE(Yi ) = n(2p − 1)
Consider
E(Yi2 ) = (1)2 (p) + (−1)2 q = p + q = 1
Problem 9. Let {X(t)} be a Gaussian white noise process and {Y (t)} is another
Rt
process such that Y (t) = X(α )d α then
0
(i) Find the autocorrelation function.
(ii) Show that {Y (t)} is a Wiener process.
S OLUTION :
(i) We know that the autocorrelation of a Gaussian white noise process is given
by
Rxx (t1 , t2 ) = b(t1 )δ (t1 − t2 ) = b0 δ (τ )
Z t Zs
Ryy (t, s) = Rxx (α , β )d β d α
0 0
Z t Zs
= b0 δ (α − β )d β d α
0 0
Zs Zt
= b0 u(t − β )d β or b0 u(s − α )d α
0 0
min(t,
Z s)
∴ Ryy (t, s) = b0 d β = b0 min(t, s)
0
(ii) By definition we know that the autocorrelation function of the Wiener pro-
cess is same as the one obtained in (i). Also Y (0) = 0 and since {X(t)} is a
Gaussian white noise process we have E {Y (t)} = 0, and hence we conclude
that {Y (t)} is a Wiener process.
Normal Process (Gaussian Process) • 179
Problem 10. Let {X(t)} be a Wiener process with parameter b0 and {Y (t)} is
Rt
another process such that Y (t) = X(α )d α , then find the mean and variance of
0
{Y (t)}.
S OLUTION :
Since {X(t)} is a Wiener process, we have E {X(t)} = 0
Zt Zt
⇒ E {Y (t)} = E X(α )d α = E {X(α )} d α = 0
0 0
n o Zt Zt
⇒ E Y 2 (t) = E {X(α )X(β )} d α d β
0 0
Zt Zt
= E {X(α )X(β )} d α d β
0 0
Zt Zt
= Rxx (α , β )d β d α
0 0
Rxx (α , β ) = b0 min(α , β )
n o Zt Zt
∴ V {Y (t)} = E Y 2 (t) − {E[Y (t)]}2 = b0 min(α , β )d α d β
0 0
Zt Zβ Zt Zα
= b0 α d α d β + b0 β d β d α (Refer to the Figure below)
0 0 0 0
b0t 3
=
3
180 • Probability and Random Processes for Engineers
α=β
β>α
α>β
0 t α
EXERCISE PROBLEMS
1. If {X(t)} is a Gaussian process with mean 0, autocorrelation function
Rxx (τ ) = 2−|τ | , where τ = t1 − t2 , then obtain P {|X(t)| ≤ 0.5}.
2. If {X(t)} is a random process whose sample path is given by X(t) = W
sin(π t) + Y , where Y is a positive random variable with mean µ and
variance σ 2 and W is the standard normal random variable independent of
Y . Obtain mean and variance of {X(t)}. Comment on the stationarity of the
process {X(t)}.
3. If {X(t)} is a Gaussian process with mean 0 and autocorrelation function
R(τ ) = 4 e−|τ | , where τ = t1 − t2 , then find P(W > 2) where the random
variable W = X(t − 1).
4. Given a normal process {X(t)} with mean 0, autocorrelation function
Rxx (τ ) = 2−|τ | , where τ = t1 − t2 , then what is the joint probability den-
sity function of the random variables Y = X(t) and W = X(t + 1).
5. Suppose {X(t)} is a Gaussian random process with mean E {X(t)} = 0 and
autocorrelation function Rxx (τ ) = e−|τ | . If A is a random variable such that
R1
A = X(t)dt. Then determine expectation and variance of A.
0
6. If {X(t)} is a zero mean stationary Gaussian process with Rxx (τ ) = cos(τ )
find the mean, variance and autocorrelation of the square law detector pro-
cess of {X(t)}.
7. If {X(t)} is a zero mean stationary Gaussian process with autocorrelation
function Rxx (τ ) = cos(τ ) and if {Z(t)} is a half-wave linear detector process
then obtain mean and variance of {Z(t)}.
8. If {X(t)} is a zero mean stationary Gaussian process with autocorrelation
function R(τ ) = 4 e−3|τ | and if {Y (t)} is a hard limiter process then obtain
mean, variance and autocorrelation of {Y (t)} when the time points are t and
t + 2.
Normal Process (Gaussian Process) • 181
7.0 INTRODUCTION
In random process studies, a central problem in the application of such processes
is the estimation of various statistical parameters in terms of real data which are
nothing but signals. Similar to that of estimation of parameters in statistical dis-
tribution, the parameter estimation in random process is mostly related to find
the expected values of some functional forms of the given process. The real chal-
lenge in this estimation is the limited availability of data (signals). For example,
one may observe only one signal during a time interval out of as many possible
signals. Obviously, in this chapter, the problem of estimating the mean and, of
course, the variance of a given process is considered. If entire spectrum of signals
(i.e., ensemble) is available, ensemble average can be obtained which is similar
to that of population parameter in statistical studies. However, as discussed above,
only one signal (single realization of the process) can be observed during a time
interval from which time average can be obtained as an estimate of ensemble aver-
age. Ergodicity is, in fact, related to the estimation of ensemble average using time
average. Ergodicity is related to correlation and distribution as well. As a result,
one can verify whether a given process is mean ergodic or correlation ergodic or
distribution ergodic.
Under this circumstance, if the random variable X(t j , ξ ) is discrete with probabil-
ity mass function P[X(t j , ξ ) = x] then its expected value (average) can be statisti-
cally obtained as
Spectrum Estimation: Ergodicity • 183
n
E X(t j , ξ ) = ∑ xi P X t j , ξi = xi
for j = 1, 2, · · · · · · (7.1)
i=1
For example, refer to Figure 7.1 taken from Chapter 2 for illustration. In this figure,
if we assume that at time point t6 the process X(t, ξ ) will assume the values
X (t, ξ) 50
X (t6, ξ)
X (t, ξ2)
Outcome (gain in rupees)
40
x2
30 X (t, ξi)
x1
20
X (t, ξn)
10 X (t, ξ1)
xn
xi
0 t
0 1 2 3 4 5 6 7 8 9 10
Time t6
1 n 1
E[X(t6 , ξ )] = ∑ xi
n i=1
∵ P[X (t6 , ξi ) = xi ] =
n
1 n
= ∑ X (t6 , ξi )
n i=1
0≤ξ ≤1
(
1,
and f (ξ , t) =
0, otherwise
Therefore,
Z1
E[X(t)] = 1.5 cos(2.5t + ξ )d ξ = 1.5 [sin(2.5t + 1) − sin(2.5t)]
0
ZT
1
XT = XT (t) dt
2T
−T
Since, the functions XT (t) and X(t) are same over a period of time, in general, we
present the time average simply as
ZT
1
XT = X (t) dt (7.6)
2T
−T
However, if the process is observed in the time interval (0, T ), then we have
ZT
1
XT = X (t) dt (7.7)
T
0
Again, consider the Example 2.4 of Chapter 2, in which one of the member func-
tions of the random process {X(t)} is given by X(t, ξ1 ) = 1.5 cos (2.5t + 0.05) or
186 • Probability and Random Processes for Engineers
simply X(t) = 1.5 cos (2.5t + 0.05). Then, in the interval (0, t) = (0, T ) = (0, 2),
we have the time average as
Z2
1
XT = 1.5 cos (2.5t + 0.05) dt
2
0
1.5 sin (2.5t + 0.05) 2 1.5
= = [sin (5.05) − sin (0.05)] = −0.2981
2 2.5 0 5
Therefore, the value −0.2681 can be taken as the time average of the process
X(t) = 1.5 cos (2.5t + 0.05). Refer to Figure 7.2.
X (t)
2
X T = 0.2981
1
−1
−2 t
−1 −T = 0 1 +T = 2 3
Figure 7.2. Time average of the process X(t) = 1.5 cos (2.5t + 0.05)
ZT
1
E {Y (t)} = lim Y T = Y (t) dt → F(x,t)
T →∞ 2T
−T
This is true because, we know that the time average of {X(t)} is given by
ZT
1
X T = µT = X(t) dt
2T
−T
ZT
1
⇒ E {µT } = E {X(t)} dt = µ ∵ E {X(t)} = µ
2T
−T
Proof.
Consider
ZT
1
XT = X(t) dt
2T
−T
ZT
1
E {X(t)} dt = µ E {X(t)} = µ
⇒ E XT = ∵
2T
−T
⇒ lim X T = µ = E {X(t)}
T →∞
Z2T
|τ |
1
lim C(τ ) 1 − dτ = 0
T →∞ 2T 2T
−2T
Or
Z2T
1 τ
lim C(τ ) 1 − dτ = 0
T →∞ T 2T
0
SOLVED PROBLEMS
Problem 1. If {X1 (t)} and {X2 (t)} are two mean ergodic processes with means
µ1 and µ2 respectively and {X(t)} is another random process such that X(t) =
X1 (t) + AX2 (t) where A is a random variable independent of X2 (t). The random
variable A assumes 0 and 1 with equal probabilities. Then show that the process
{X(t)} is not mean ergodic.
S OLUTION :
Consider E {X(t)} = E {X1 (t) + AX2 (t)}
= E {X1 (t)} + E {AX2 (t)}
= E {X1 (t)} + E(A) {X2 (t)}
1 1 1
But E(A) = (0) + (1) =
2 2 2
1
∴ E {X(t)} = E {X1 (t)} + {X2 (t)}
2
Let ξ = {ξ1 , ξ2 } = {0, 1} be the set of all possible outcomes of the random vari-
able A. It is possible that A(ξ ) = 0 for a particular ξ . Then we have
where E {X(t)} = µt
Though E {X(t)} is constant in both the cases, the values are different, therefore,
the process {X(t)} is not mean ergodic.
Problem 2. Show that the random process {X(t)} with constant mean is mean
ergodic, if
1 ZT ZT
lim C(t1 , t2 ) dt1 dt2 =0
T →0 4T 2
−T −T
S OLUTION :
We know that according to mean ergodic theorem, the stationary random pro-
cess {X(t)} with a constant ensemble average µ and time average given by X T =
ZT
1
X(t) dt is mean ergodic if lim V X T = 0
2T T →∞
−T
ZT
1
Consider X T = X(t) dt
2T
−T
ZT
1
⇒ E(X T ) = E {X(t)} dt = E {X(t)}
2T
−T
Now consider
2
1 ZT 1
ZT
1
ZT
2
XT = X(t) dt = X(t) dt X(t) dt
2T 2T 2T
−T −T −T
Since the process {X(t)} is stationary, given two time points t1 and t2 the time
averages at these time points are same. Therefore, we can write
ZT ZT
2 1 1
XT = X(t1 ) dt1 X(t2 ) dt2
2T 2T
−T −T
Spectrum Estimation: Ergodicity • 191
ZT ZT
1
= X(t1 )X(t2 ) dt1 dt2
4T 2
−T −T
ZT ZT
2
1
⇒ E XT = E {X(t1 )X(t2 )} dt1 dt2
4T 2
−T −T
ZT ZT
1
= R(t1 , t2 ) dt1 dt2
4T 2
−T −T
We know that
2 2
V (X T ) = E(X T ) − E(X T
2
ZT ZT ZT
1 1
= R(t1 , t2 ) dt1 dt2 − E X(t)dt
4T 2 2T
−T −T −T
ZT ZT ZT
1 1
= R(t1 , t2 ) dt1 dt2 − E X(t1 )dt1
4T 2 2T
−T −T −T
ZT
1
E X(t2 )dt2
2T
−T
ZT ZT ZT
1 1
= R(t1 , t2 ) dt1 dt2 − E {X(t1 )} dt1
4T 2 2T
−T −T −T
ZT
1 E {X(t2 )} dt2
2T
−T
ZT ZT ZT ZT
1 1
= R(t1 , t2 ) dt1 dt2 − 2 E {X(t2 )} E {X(t1 )} dt1 dt2
4T 2 4T
−T −T −T −T
ZT ZT
1
= [R(t1 , t2 ) − E {X(t1 )} E {X(t2 )}] dt1 dt2
4T 2
−T −T
192 • Probability and Random Processes for Engineers
ZT ZT
1
= C(t1 , t2 )dt1 dt2
4T 2
−T −T
1 ZT ZT
∴ lim V (X T = 0 ⇒ lim C(t1 , t2 )dt1 dt2 =0
T →∞ T →∞ 4T 2
−T −T
Therefore, the random process {X(t)} with constant mean is mean ergodic, if
1 ZT ZT
lim C(t1 , t2 ) dt1 dt2 =0
T →0 4T 2
−T −T
Z2T
1 τ
C(τ ) 1 − dτ
V XT =
T 2T
0
And hence prove that the sufficient condition for the mean-ergodicity of the pro-
cess {X(t)} is
Z2T
1 τ
lim C(τ ) 1 − dτ = 0
T →∞ T 2T
0
R∞
which in turn implies that |C(τ )| d τ < ∞.
−∞
S OLUTION :
It is known that V X T can be expressed as (See Problem 2)
ZT ZT
1
V XT = C(t1 ,t2 ) dt1 dt2
4T 2
−T −T
ZT ZT Z2T
C(t1 ,t2 ) dt1 dt2 = C(τ ) (2T − |τ |) d τ
−T −T −2T
Spectrum Estimation: Ergodicity • 193
Z2T
1
C(τ )(2T − |τ |)d τ
∴ V XT =
4T 2
−2T
Z2T
|τ |
1
= C(τ ) 1 − dτ
2T 2T
−2T
Z2T
1 τ
= C(τ ) 1 − dτ
T 2T
0
Z2T
1 τ
lim C(τ ) 1 − dτ = 0
T →∞ T 2T
0
Obviously, when lim V X T = 0, we have
T →∞
Z2T
|τ |
1
lim C(τ ) 1 − dτ = 0
T →∞ 2T 2T
−2T
We know that τ varies from −2T to +2T and hence |τ | ≤ 2T . Therefore, we have
|τ |
1− ≤ 1 and which follows that
2T
Z2T Z2T
|τ |
1 1
C(τ ) 1 − dτ ≤ |C(τ )|d τ
2T 2T 2T
−2T −2T
Here, |C(τ )|is considered to ensure that the inequality is valid. Now,
Z2T Z2T
|τ |
1 1
lim C(τ ) 1 − d τ = 0 is true only if lim |C(τ )|d τ = 0
T →∞ 2T 2T T →∞ 2T
−2T −2T
Z∞
This leads to the conclusion that |C(τ )|d τ < ∞ (finite). Therefore, the sufficient
−∞
194 • Probability and Random Processes for Engineers
condition for mean-ergodicity of the stationary process {X(t)} can also be given
Z∞
as |C(τ )|d τ < ∞.
−∞
Problem 4. If {X(t)} is a wide sense stationary process with constant mean and
autocovariance function
ω 1 − ττ ,
(
for 0 ≤ τ ≤ τ0
C(τ ) = 0
0, otherwise
where ω is a constant, then find the variance of the time average of {X(t)} over
the interval (0, T ). Also examine if the process {X(t)} is mean ergodic.
S OLUTION :
We know that the time average X T of the stationary random process {X(t)} in
ZT
1
the interval (−T, T ) is given by X T = X(t) dt and its variance is given by
2T
−T
Z2T
|τ |
1
V (X T ) = C(τ ) 1 − dτ .
2T 2T
−2T
Accordingly, we can write the time average and variance of time average in
the interval (0, T ) as
ZT
1
XT = X(t) dt
T
0
ZT
|τ |
1
V (X T ) = C(τ ) 1 − dτ
T T
−T
ZT
2 τ
= C(τ ) 1 − dτ
T T
0
It may be noted that given the value τ0 and the interval (0, T ), it is possible to have
either Case (i) τ0 > T or Case (ii) τ0 < T (Refer the Figure below).
Spectrum Estimation: Ergodicity • 195
0 T τ0
0 τ0 T
ZT
τ τ
2
V (X T ) = ω 1− 1− dτ
T τ0 T
0
ZT
2ω τ τ τ2
= 1− − + dτ
T T τ0 τ0 T
0
T
2ω τ2 τ2 τ3
= τ− − +
T 2T 2τ0 3τ0 T 0
2ω T2 T2
T T
= T− − + = ω 1− (1)
T 2 2τ0 3τ 0 3τ0
Zτ0
τ τ
2
V (X T ) = ω 1− 1− dτ
T τ0 T
0
Zτ0
2ω τ τ τ2
= 1− − + dτ
T T τ0 τ0 T
0
196 • Probability and Random Processes for Engineers
τ0
2ω τ2 τ2 τ3
= τ− − +
T 2T 2τ0 3τ0 T 0
!
2ω τ02 τ02 τ03 ωτ0 τ0
= τ0 − − + = 1− (2)
T 2T 2τ0 3τ0 T T 3T
It may be noted that as T → ∞ only (2) holds good since in this case T > τ0 .
n ωτ τ0 o
0
Therefore, lim V (X T ) = lim 1− =0
T →∞ T →∞ T 3T
Hence, the process {X(t)} is mean ergodic.
Problem 5. A binary transmission process {X(t)} has zero mean and autocorre-
|τ |
lation function R(τ ) = 1 − . Find the mean and variance of the time average of
T
the process {X(t)} over the interval (0, T ) and verify whether the process is mean
ergodic.
S OLUTION :
We know that the time average X T of the stationary random process {X(t)} in
ZT
1
the interval (0, T ) is given by X T = X(t) dt.
T
0
Therefore, the mean of the time average denoted by E X T is obtained as
ZT
1
E(X T ) = E {X(t)} dt = E {X(t)} = 0
2T
−T
In the interval (0, T ) the variance of the time average, denoted by V X T , is
given by
ZT
|τ |
1
V (X T ) = C(τ ) 1 − dτ
T T
−T
ZT
|τ |
1
= R(τ ) 1 − dτ
T T
−T
Spectrum Estimation: Ergodicity • 197
ZT
|τ |
2
1
= 1− dτ
T T
−T
ZT
2 τ 2
= 1− dτ
T T
0
" #T
(1 − τ /T )3
2 2
= =
T3(−1/T ) 3
0
2 2
Therefore, lim V (X T ) = lim = 6= 0
T →∞ T →∞ 3 3
Since lim V (X T ) = 6 0 the process is not mean ergodic.
T →∞
Problem 6. If {X(t)} is a wide sense stationary process such that X(t) = 10 cos
(100t + θ ) where θ is a random variable uniformly distributed over (−π , π ) then
show that the process {X(t)} is correlation ergodic.
S OLUTION :
Since θ is a random variable uniformly distributed over (−π , π ), we have the
probability density function of θ as
1
f (θ ) = , −π ≤ θ ≤ π
2π
We know that autocorrelation function of the stationary random process {X(t)} is
given by
R(τ ) = E {X(t + τ )X(t)}
= E {10 cos[100(t + τ ) + θ ]10 cos[100t + θ }
= 100E {cos[100(t + τ ) + θ ] cos[100t + θ }
= 50E {cos(200t + 100τ + 2θ ) + cos 100τ }
= 50E {cos(200t + 100τ + 2θ ) } + 50E {cos 100τ }
Zπ Zπ
= 50 cos(200t + 100τ + 2θ ) f (θ )d θ + 50 cos 100τ f (θ )d θ
−π −π
198 • Probability and Random Processes for Engineers
Zπ Zπ
25 25
= cos(200t + 100τ + 2θ )d θ + cos 100τ d θ
π π
−π −π
25
= (0) + 50 cos 100τ (See Appendix B)
π
= 50 cos 100τ
We know that the time average of the process {Z(t)} in the interval (−T, T ) is
given by
ZT
1
ZT = X(t + τ )X(t) dt
2T
−T
ZT
1
= 100 cos[100(t + τ ) + θ ] cos[100t + θ ]dt
2T
−T
Zπ
50 cos(200t + 100τ + 2θ ) + cos 100τ
= dt
T 2
−π
Zπ ZT
25 25
= [cos(200t + 100τ + 2θ )] dt + cos 100τ dt
T T
−π −T
25
= (0) + 50 cos 100τ = 50 cos 100τ
T
= 50 cos 100τ
∴ lim Z T = lim {50 cos 100τ } = 50 cos 100τ = R(τ )
T →∞ T →∞
Problem 7. If {X(t)} is a zero mean wide sense stationary process with Rxx (τ ) =
e−2|τ | , show that {X(t)} is mean ergodic.
S OLUTION :
It is given that E{X(t)} = µ = 0 (that is, the ensemble average is a constant).
Now consider the time average of the process {X(t)} in the interval (−T, T )
given by
Spectrum Estimation: Ergodicity • 199
ZT ZT
1 1
XT = X(t) dt ⇒ E(X T ) = E {X(t)} dt = µ = 0
2T 2T
−T −T
2 2
We know that V (X T ) = E(X T ) − E(X T )
ZT ZT ZT
2 1 1 1
Consider X T = X 2 (t) dt = X(t) dt X(t) dt
2T 2T 2T
−T −T −T
Since the process {X(t)} is stationary, given two time points t1 and t2 the time
averages at these time points are same. Therefore, we can write
ZT ZT
2 1 1
XT = X(t1 ) dt1 X(t2 ) dt2
2T 2T
−T −T
ZT ZT
1
= X(t1 )X(t2 ) dt1 dt2
4T 2
−T −T
ZT ZT
2 1
⇒ E(X T ) = E {X(t1 )X(t2 )} dt1 dt2
4T 2
−T −T
ZT ZT
1
= R(t1 , t2 ) dt1 dt2
4T 2
−T −T
ZT ZT
1
= C(t1 , t2 ) dt1 dt2 ∵ E {X(t)} = 0
4T 2
−T −T
Z2T
1 τ
V (X T ) = C(τ ) 1 − dτ
T 2T
0
Z2T
1 τ
= R(τ ) 1 − dτ ∵ E {X(t)} = 0, C(τ ) = R(τ )
T 2T
0
Z2T
1 τ
e−2|τ | 1 −
= dτ
T 2T
0
200 • Probability and Random Processes for Engineers
Z2T
1 τ
e−2τ 1 −
= dτ
T 2T
0
Z2T Z2T
1 −2τ 1
= e dτ − 2 e−2τ τ d τ
T 2T
0 0
1 e−2τ
−2τ
e−2τ
2T 2T
1 e
= − 2 τ − (1)
T −2 0 2T −2 4 0
1 e−4T 1 1
= − − 2
T −2 −2 2T
e−4T e−4T 1
2T − (1) − (0) −
−2 4 4
1 e−4T 1
= 1+ −
2T 4T 4T
1 1 − e−4T
∴ lim V (X T ) = lim 1− =0
T →∞ T →∞ 2T 4T
Problem 8. A random process {X(t)} has the sample functions of the form X(t) =
A cos(ω t + θ ) where ω is a constant, A is a random variable that has magnitude
+1 and −1 with equal probabilities, and θ is a random variable that is uniformly
distributed between 0 and 2π . Assume that A and θ are independent. Is {X(t)} a
mean ergodic process?
S OLUTION :
In order to show that a random process {X(t)} defined in the time interval
(−T, T ) is mean ergodic we have to show that
(i) Ensemble average: E {X(t)} = µ (constant).
ZT
1
(ii) Time average: lim X T = X(t) dt = µ .
T →∞ 2T
−T
Since A and θ are independent we have
2Rπ
Also E[cos(ω t + θ )] = cos(ω t + θ ) f (θ )d θ
0
Since θ is a random variable that is uniformly distributed between 0 and 2π ,
we have
1
f (θ ) = , 0 ≤ θ ≤ 2π
2π
Z2π
1
∴ E[cos(ω t + θ )] = cos(ω t + θ ) dθ
2π
0
1
= [sin(ω t + θ )]02π
2π
1
= [sin(ω t + 2π ) − sin ω t] = 0
2π
∴ E {X(t)} = E(A)E[cos(ω t + θ )] = 0
ZT
A sin(ω t + θ ) T
1
XT = A cos(ω t + θ ) dt =
2T 2T ω −T
−T
A
= [sin(θ + ω T ) − sin(θ − ω T )]
2T ω
A
= [2 cos θ sin ω T ]
2T ω
sin ω T
= A cos θ
ωT
sin ω T
lim X T = A cos θ lim = (A cos θ ) (0) = 0
T →∞ T →∞ ωT
Problem 9. Consider the sinusoid with random phase X(t) = a sin(ω t + θ ) where
a is constant and θ is a random variable uniformly distributed over (0, 2π ). Show
that the process {X(t)} is correlation ergodic.
202 • Probability and Random Processes for Engineers
S OLUTION :
In order to show that a stationary random process {X(t)} defined in the time
interval (−T, T ) is correlation ergodic we have to show that the process {Z(t)} is
mean ergodic, where Z(t) = X(t)X(t + τ ) or Z(t) = X(t + τ )X(t). That is,
(i) E {Z(t)} = E {X(t + τ )X(t)} = R(τ ) (the autocorrelation function).
ZT
(ii) 1
lim Z T = lim 2T X(t + τ )X(t) dt = R(τ ).
T →∞ T →∞
−T
Consider
= a2 E {sin[ω t + ωτ + θ ] sin(ω t + θ )}
cos ωτ − cos(2ω t + ωτ + 2θ )
= a2 E
2
a2 a2
= E(cos ωτ ) − E {cos(2ω t + ωτ + 2θ )}
2 2
Z2π Z2π
a2 1 a2 1
= cos ωτ dθ − cos(2ω t + ωτ + 2θ ) dθ
2 2π 2 2π
0 0
Z2π
a2 a2
= cos ωτ − cos(2ω t + ωτ + 2θ )d θ
2 4π
0
Consider
Z2π
a2 a2
cos(2ω t + ωτ + 2θ )d θ = [sin(2ω t + ωτ + 2θ )]02π
4π 8π
0
a2
= [sin(2ω t + ωτ + 4π ) − sin(2ω t + ωτ )] = 0
8π
a2
∴ E {Z(t)} = R(τ ) = cos ωτ
2
Now, consider
ZT ZT
1 1
ZT = X(t + τ )X(t) dt = a sin[ω (t + τ ) + θ ]a sin(ω t + θ )dt
2T 2T
−T −T
Spectrum Estimation: Ergodicity • 203
ZT
a2
= sin[ω (t + τ ) + θ ] sin(ω t + θ )dt
2T
−T
ZT
a2 cos ωτ − cos(2ω t + ωτ + 2θ )
= dt
2T 2
−T
ZT ZT
a2 a2
= cos ωτ dt − cos(2ω t + ωτ + 2θ ) dt
4T 4T
−T −T
ZT
a2 a2
= cos ωτ − cos(2ω t + ωτ + 2θ ) dt
2 4T
−T
Consider
ZT
a2 a2
cos(2ω t + ωτ + 2θ ) dt = [sin(2ω t + ωτ + 2θ )]T−T
4T 8ω T
−T
a2
= [sin(ωτ + 2θ + 2ω T ) − sin(ωτ + 2θ − 2ω T )]
8ω T
a2
= [2 cos(ωτ + 2θ ) sin(2ω T )]
8ω T
a2
= [cos(ωτ + 2θ ) sin(2ω T )]
4ω T
a2 a2
∴ lim Z T = lim cos ωτ − [cos(ωτ + 2θ ) sin(2ω T )]
T →∞ T →∞ 2 4ω T
a2
= cos ωτ = R(τ )
2
Since E {Z(t)} = lim Z T = R(τ ) the process {X(t)} is correlation ergodic.
T →∞
Problem 10. If {X(t)} is a wide sense stationary process such that X(t) = 4 cos
(50t + θ ) where θ is a random variable uniformly distributed over (−π , π ) then
show that the process {X(t)} is correlation ergodic.
S OLUTION :
Since θ is a random variable uniformly distributed over (−π , π ), we have the
probability density function of θ as
1
f (θ ) = , −π ≤ θ ≤ π
2π
204 • Probability and Random Processes for Engineers
= E {4 cos[50(t + τ ) + θ ]4 cos[50t + θ ] }
Zπ Zπ
=8 cos(100t + 50τ + 2θ ) f (θ )d θ + 8 cos 50τ f (θ )d θ
−π −π
Zπ Zπ
4 4
= cos(100t + 50τ + 2θ )d θ + cos 50τ d θ
π π
−π −π
4
= (0) + 8 cos 50τ (See Appendix B)
π
= 8 cos 50τ
Now, consider ZT = X(t + τ )X(t)
⇒ E(ZT ) = E {X(t + τ )X(t)} = R(τ ) = 8 cos 50τ
We know that the time average of the process {Z(t)} in the interval (−T, T ) is
given by
ZT
1
ZT = X(t + τ )X(t) dt
2T
−T
ZT
1
= 4 cos[100(t + τ ) + θ ]4 cos[100t + θ ] dt
2T
−T
Zπ
8 cos(100t + 50τ + 2θ ) + cos 50τ
= dt
T 2
−π
Zπ ZT
4 4
= [cos(100t + 50τ + 2θ )] dt + cos 50τ dt
T T
−π −T
Spectrum Estimation: Ergodicity • 205
4
= (0) + 8 cos 50τ = 8 cos 50τ
T
∴ lim Z T = lim {8 cos 50τ } = 8 cos 50τ = R(τ )
T →∞ T →∞
EXERCISE PROBLEMS
1. If {X(t)} is a random process such that X(t) = A where A is a random vari-
able with mean µA then show that the process {X(t)} is not mean ergodic.
2. If {X(t)} is a zero mean wide sense stationary process with Rxx (τ ) = 4e−|τ | ,
show that {X(t)} is mean ergodic.
3. Consider the sinusoid with random phase X(t) = a sin(ω t + θ ) where a
is constant and θ is a random variable uniformly distributed over (0, 2π ).
Show that the process {X(t)} is mean ergodic.
4. If {X(t)} is a wide sense stationary process such that X(t) = cos(t + θ )
where θ is a random variable uniformly distributed over (−π , π ) then show
that the process {X(t)} is correlation ergodic.
5. Show that the stationary random process {X(t)} whose autocovariance
function is given by C(τ ) = qe−ατ , where q and α are constants, is mean
ergodic.
6. If {X(t)} is a stationary random process such that X(t) = η + W (t) where
W (t) is a white noise process with autocorrelation function Rww (τ ) = qδ (τ ),
where q is constant and δ is the unit impulse function, then show that {X(t)}
is mean ergodic.
7. If {X(t)} is a stationary process with X(t) = A, where A is a random vari-
able, having an arbitrary probability density function. Check whether the
process {X(t)} is mean ergodic.
8. If {X(t)} is a stationary random telegraph signal with mean 0 and autocor-
relation function R(τ ) = e−2λ τ , where λ is constant, then find the mean and
variance of the time average of {X(t)}. Also verify whether {X(t)} is mean
ergodic.
9. If {X(t)} is a stationary Gaussian process with zero mean and autocorre-
lation function R(τ ) = 10 e−|τ | then show that the process {X(t)} is mean
ergodic.
10. If {X(t)} is a stationary random process with autocorrelation function R(τ )
then show that the process {X(t)} is correlation ergodic if and only if
Z2T
1 τ
lim φ (τ ) 1 − d τ → {R(τ )}2
T →∞ 2T 2T
0
where
φ (τ ) = E {X(t1 + τ )X(t1 )X(t2 + τ )X(t2 ) } .
C HAPTER 8
POWER SPECTRUM: POWER SPECTRAL
DENSITY FUNCTIONS
8.0 INTRODUCTION
In case of stationary random process {X(t)} observed in a time interval (0,t), the
autocorrelation function, denoted by Rxx (τ ), where τ = t1 − t2 or τ = t2 − t1 with
t1 , t2 ∈ (0,t), plays an important role in determining the strength of a process that
is, particularly, observed in the form of a signal. In fact, the autocorrelation shows
how rapidly one can expect the random signal represented by X(t) of a station-
ary process {X(t)} to change as a function of time, t. That is, if Rxx (τ ) decays
rapidly (deteriorates fast), then it indicates that the signal is expected to change
rapidly (fast). In other sense, if Rxx (τ ) decays slowly (deteriorates slowly), then it
indicates that the signal is expected to change slowly. Further, if the autocorrela-
tion function has periodic components, then such a periodicity will be reflected on
the corresponding process as well. Apparently, one can understand the fact that the
autocorrelation function Rxx (τ ) contains information about the expected frequency
content in the signal of the stationary process of interest.
Power spectral density (PSD) describes how the power (or variance or ampli-
tudes) of a time series (a time dependent signal) is distributed with frequency. In
other simple terms, the PSD captures the frequency content in a signal. Or other-
wise, the PSD refers to the amount of power per unit of frequency as a function
of frequency. For example, if a realization X(t) of a stationary random process
{X(t)} represents a voltage waveform across a one ohm (1 Ω) resistance, then the
ensemble average of the square of X(t) is nothing but the average of power deliv-
ered to the 1 Ω resistance by X(t). That is, E X 2 (t) = Rxx (t, t) = Rxx (0) gives
the average power of {X(t)}. Therefore, the autocorrelation function at τ = 0 gives
the average power of the process {X(t)} .
It may be noted that in the theory of signals, spectra are associated with Fourier
transforms. The Fourier transforms are used to represent a function as a superposi-
tion of exponentials for determining signals. In fact, for random signals, the notion
of a spectrum has two interpretations: the first one involves transforms of aver-
ages and is essentially deterministic and the second one leads to the representation
of the process under consideration as superposition of exponentials with random
Power Spectrum: Power Spectral Density Functions • 207
amplitude, X(t)
3
2
1
0
-1
-2
-3
-4
time, t
Figure 8.1. Magnified amplitudes of a single realization of the random process {X(t)}
Frequency
4
−4 −3 −2 −1 0 1 2 3 4
Amplitude
Figure 8.2. Frequency distribution of amplitudes and the fitted normal curve as spectral density
function
However, in practice, one cannot manually count the frequencies for each
amplitude as signal may vary rapidly. This job is, in fact, done by the Fourier trans-
form of autocorrelation of the signal under study since autocorrelation captures the
changes in the signal over time. As discussed earlier, since autocorrelation is used,
this transformation will result in power spectral density function of the stationary
random process.
208 • Probability and Random Processes for Engineers
Z∞
Sxx (ω ) = Rxx (τ ) e−i ωτ d τ (8.1)
−∞
Z∞
Sxx ( f ) = Rx (τ ) e−i2π f τ d τ (8.2)
−∞
Accordingly, if the PSD, Sxx (ω ), is known, then the autocorrelation function Rxx (τ )
can be obtained as the Fourier inverse transform of Sxx (ω ) and is given by
Z∞
1
Rxx (τ ) = Sxx (ω )e+iτ ω d ω (8.3)
2π
−∞
Z∞
Rxx (τ ) = Sxx ( f )e+i 2π τ f d f (8.4)
−∞
Z∞
Sxy (ω ) = Rxy (τ ) e−i ωτ d τ (8.5)
−∞
Z∞
Sxy ( f ) = Rxy (τ ) e−i2π f τ d τ (8.6)
−∞
Power Spectrum: Power Spectral Density Functions • 209
Z∞
Rxy (τ ) = Sxy ( f ) e+i 2π τ f d f (8.8)
−∞
R∞
Proof. We know that Sxx (ω ) = Rxx (τ ) e−i ωτ d τ
−∞
R∞
When ω = 0, we have Sxx (0) = Rxx (τ ) d τ
−∞
Proof. We know that the second order moment of the process {X(t)} is given by
n o
E X 2 (t) = Rxx (t, t) = Rxx (0)
Z∞
1
⇒ Rxx (0) = Sxx (ω ) d ω
2π
−∞
210 • Probability and Random Processes for Engineers
Or with, ω = 2π f , we have
n o Z∞
2
E X (t) = Rxx (0) = Sxx ( f ) d f
−∞
Property 8.3: The PSD function Sxx (ω ) of a stationary process {X(t)} with auto-
correlation function Rxx (τ ) is an even function. That is, Sxx (ω ) = Sxx (−ω ). Also,
Sxy (ω ) = Syx (−ω ).
Z∞
Proof. We know that Sxx (ω ) = Rxx (τ ) e−i ωτ d τ
−∞
Z∞
Now, consider Sxx (−ω ) = Rxx (τ ) e+i ωτ d τ
−∞
Let τ = −ν
Z∞
⇒ Sxx (−ω ) = Rxx (−v) e−i ω v dv
−∞
Z∞
= Rxx (v) e−i ω v dv = Sxx (ω ) ∵ R(v) = R(−v)
−∞
Property 8.4: The PSD function Sxx (ω ) and autocorrelation function Rxx (τ ) of a
stationary process {X(t)} form a Fourier cosine transform pair.
Z∞
Sxx (ω ) = Rxx (τ ) e−i ωτ d τ
−∞
Z∞
= Rxx (τ ) (cos ωτ − i sin ωτ )d τ
−∞
Z∞ Z∞
= Rxx (τ ) cos ωτ d τ − Rxx (τ )i sin ωτ d τ
−∞ −∞
Z∞ Z∞
=2 Rxx (τ ) cos ωτ d τ ∵ Rxx (τ )i sin ωτ d τ = 0
0 −∞
Power Spectrum: Power Spectral Density Functions • 211
Z∞
∴ Sxx (ω ) = 2Rxx (τ ) cos ωτ d τ (8.11)
0
Z∞
1
= Sxx (ω ) (cos τω + i sin τω ) d ω
2π
−∞
Z∞ Z∞
= Sxx (ω ) cos τω d ω + Sxx (ω )i sin τω d ω
−∞ −∞
Z∞ Z∞
1
= Sxx (ω ) cos τω d ω ∵ Sxx (ω )i sin τω d ω = 0
π
0 −∞
Z∞
1
∴ Rxx (τ ) = Sxx (ω ) cos τω d ω (8.12)
π
0
1
which is a Fourier inverse cosine transform of Sxx (ω ).
π
Property 8.5: The PSD function of the output process {Y (t)} corresponding
to the input process {X(t)} in the system that has an impulse response h (t) =
e−β t U(t), where U(t) is the unit step function, is given as
Property 8.6: The PSD function Sxx (ω ) of a stationary process {X(t)} (whether
real or complex) with autocorrelation function Rxx (τ ) is a real and non-negative
∗ (ω ) = S (ω ). Also, S∗ (ω ) = S (ω ), where S∗ is the com-
function. That is, Sxx xx xy yx
plex conjugate.
Proof. Consider
Similarly,
Rxx (−τ ) = E {X(t)X ∗ (t + τ )}
R∗xx (−τ ) = E {X(t + τ )X ∗ (t)} = Rxx (τ )
Now consider
Z∞
Sxx (ω ) = Rxx (τ ) e−i ωτ d τ
−∞
Z∞ Z∞
∗
Sxx (ω ) = R∗xx (τ ) ei ωτ d τ = Rxx (−τ ) ei ωτ d τ
−∞ −∞
Let ν = −τ
Z∞
⇒ ∗
Sxx (ω ) = Rxx (v) e−i ω v dv = Sxx (ω )
−∞
Z∞ ω0Z+ε /2
n o 1 1 ε
⇒ E Y 2 (t) = Ryy (0) = Syy (ω ) d ω = Syy (ω )d ω = Sxx (ω0 )
2π 2π 2π
−∞ ω0 −ε /2
This is true because Sxx (ω ) is constant as the band is narrow and is equal to
Sxx (ω0 ).
Since E Y 2 (t) ≥ 0 ⇒ Sxx (ω0 ) ≥ 0
This is contrary to the assumption made that Sxx (ω0 ) < 0. Therefore, Sxx (ω ) is
non-negative function.
Power Spectrum: Power Spectral Density Functions • 213
If {XT (t)} is a truncated random process of the original real stationary random
process {X(t)} such that
(
X(t), |t | ≤ T
XT (t) =
0, otherwise
o
1 n
lim E |XT (ω )| 2 = Sxx (ω )
T →∞ 2T
X (t)
XT (t) =X (t)
XT (t) = 0 XT (t) = 0
t
−T 0 +T
Z∞ Z+T
−iω t
XT (ω ) = XT (t) e dt = X(t) e−iω t dt
−∞ −T
|XT (ω )| 2 = XT (ω )XT (− ω )
214 • Probability and Random Processes for Engineers
Since {X(t)} is stationary, for a given two time points t1 , t2 ∈ (−T, +T ), we have
Z+T Z+T
−iω t1
|XT (ω )| =
2
X(t1 )e dt1 X(t2 ) eiω t2 dt2
−T −T
Z+T Z+T
= X(t1 )X(t2 ) e−iω (t1 −t2 ) dt1 dt2
−T −T
o Z+T Z+T
E {X(t1 )X(t2 )} e−iω (t1 −t2 ) dt1 dt2
n
⇒ E |XT (ω )| =
2
−T −T
Z+T Z+T
= R(t1 , t2 ) e−iω (t1 −t2 ) dt1 dt2
−T −T
o Z+T Z+T
R(t1 − t2 ) e−iω (t1 −t2 ) dt1 dt2
n
E |XT (ω )| =
2
−T −T
Z+T Z+T
= g(t1 − t2 ) dt1 dt2
−T −T
n o +2T
Z
E |XT (ω )| 2
= g(τ ) (2T − | τ |) d τ
−2T
+2T +2T
1 n 1
o Z Z
∴ lim E |XT (ω )| 2 = lim g(τ ) d τ − lim g(τ ) | τ |d τ
T →∞ 2T T →∞ T →∞ 2T
−2T −2T
Power Spectrum: Power Spectral Density Functions • 215
Z+∞
= g(τ ) d τ
−∞
Z+∞
1 n
R(τ ) e−iω τ d τ = Sxx (ω )
o
⇒ lim E |XT (ω )| 2 =
T →∞ 2T
−∞
Therefore, in simple terms the relation between {X(t)} and {Y (t)} can be as
shown as
Y (t) = f {X(t)} (8.14)
Here f stands for an appropriate functional operator. Now, the process {Y (t)}
formed so is considered the output of a system whose input is {X(t)}. It may be
noted that the sample functions X(t, ξi ), i = 1, 2, . . . . . . are random in nature and
hence the sample functions Y (t, ξi ), i = 1, 2, . . . . . . too are quite random. Such
systems are called the systems with random inputs. If the relationship given in
(8.14) is linear then we have the systems with linear inputs otherwise we have
systems with non-linear inputs. Y (t) = cX(t), where c is constant, is an example
for linear case whereas Y (t) = X 2 (t) is an example for non-linear case.
If a linear system with input process {X(t)} and output process {Y (t)} is given
then by which we mean that
Y (t) = f {X(t)}
= f {a1 X1 (t) + a2 X2 (t) + . . . . . .}
= a1 f {X1 (t)} + a2 f {X2 (t)} + . . . . . .
we have
E {Y (t)} = E { f [X(t)]} = f {E[X(t)]}
Similarly, we have
Z∞
Rxy (t1 ,t2 ) = Rxx (t1 , t2 − a)h(a)da
−∞
Z∞
Ryy (t1 ,t2 ) = Rxy (t1 − a, t2 )h(a)da
−∞
That is,
Rxx (t1 ,t2 ) → Rxy (t1 ,t2 ) → Ryy (t1 ,t2 )
h(t2 ) h(t1 )
SOLVED PROBLEMS
Problem 1. Given that the random process {X(t)} is a wide sense stationary
process whose autocorrelation function traps an area of 6.25 square units in the
first quadrant. Find the value of the power spectral density function at zero fre-
quency.
S OLUTION :
Let Rxx (τ ) be the autocorrelation function of a stationary random process {X(t)}.
It is given that
Z∞
Rxx (τ ) d τ = 6.25 Sq. units
0
S OLUTION :
It is given that
b
(a − | ω |) , | ω | ≤ a
S(ω ) = a
0, |ω | > a
We know that the autocorrelation function of a stationary random process {X(t)}
is given by
Z∞
1
Rxx (τ ) = Sxx (ω ) ei τω d ω
2π
−∞
Za
1 b
= (a − | ω |) ei τω d ω
2π a
−a
Za
b
= (a − | ω |) (cos τω + i sin ωτ ) d τ
2π a
−a
Za Za
b b
= (a − | ω |) cos τω d ω + (a − | ω |) i sin τω d ω
2π a 2π a
−a −a
Za
b
= (a − |ω |) cos τω d ω − 0 ∵ (a − |ω |) sin ωτ is an odd function
2π a
−a
Za
b
= (a − ω ) cos τω d ω ∵ (a − |ω | ) cos τω is an even function
πa
0
218 • Probability and Random Processes for Engineers
sin τω cos τω a
b
= (a − ω ) −
πa τ τ2 0
cos aτ
b 1
= − 2 + 2
πa τ τ
b 1 − cos aτ
=
πa τ2
b 2 sin2 aτ /2
=
πa τ2
b sin2 aτ /2 sin aτ /2
2
ab
= =
π a τ 2 /2 2π aτ /2
Problem 3. Find the power spectral density function of the random process whose
2
autocorrelation function is given by R(τ ) = e−aτ .
S OLUTION :
It is given that
2
Rxx (τ ) = e−aτ
We know that the power spectral density function of a stationary random process
{X(t)} is given by
Z∞
Sxx (ω ) = Rxx (τ ) e−i ωτ d τ
−∞
Z∞
2
= e−aτ e−i ωτ d τ
−∞
Z∞
2 +iωτ )
= e−(aτ dτ
−∞
Add and subtract (iω /2a)2 to make the exponent a perfect square. This gives
Z∞ n o
−a τ 2 +2(iω /2a)τ +(iω /2a)2 −(iω /2a)2
Sxx (ω ) = e dτ
−∞
Z∞
2 /4a 2
= e−ω e−a{τ +iω /2a} d τ
−∞
Z∞
−ω 2 /4a −a{τ +iω /2a}2
= 2e e dτ
0
Power Spectrum: Power Spectral Density Functions • 219
√ ω
a τ + i2a dτ = du
Let u = ⇒ √
a
2 Z∞
e−ω /4a 2
∴ Sxx (ω ) = √ e−u du
a
−∞
2 Z∞
2e−ω /4a 2
= √ e−u du
a
0
√ 1
Now let v = u2 ⇒ dv = 2udu = 2 vdu ⇒ du = v−1/2 dv
2
2 Z∞
e−ω /4a 1
∴ Sxx (ω ) = √ e−v v 2 −1 dv
a
0
−ω 2 /4a
e e−ω /4a √
2
= √ Γ(1/2) = √ π
a a
π −ω 2 /4a
r
∴ Sxx (ω ) = e
a
S OLUTION :
(i) We know that since {X(t)}is a wide sense stationary process, and {Y (t)} is
another wide sense stationary random process such that Y (t) = X(t + a) −
X(t − a), the autocorrelation function Ryy (τ ) can be given as
Ryy (τ ) = E {Y (t)Y (t + τ )}
Ryy (τ ) = E {[X(t + a) − X(t − a)] [X(t + τ + a) − X(t + τ − a)]}
= E [X(t + a)X(t + τ + a)] − E [X(t + a)X(t + τ − a)] −
E [X(t − a)X(t + τ + a)] + E [X(t − a)X(t + τ − a)]
= Rxx (τ ) − Rxx (τ + 2a) − Rxx (τ − 2a) + Rxx (τ )
= 2Rxx (τ ) − Rxx (τ + 2a) − Rxx (τ − 2a)
220 • Probability and Random Processes for Engineers
Now let τ + 2a = u ⇒ τ = u − 2a ⇒ d τ = du
Similarly, let τ − 2a = v ⇒ τ = v + 2a ⇒ d τ = dv
Consider
Z∞ Z∞
−iωτ
Rxx (τ + 2a) e dτ = Rxx (u) e−iω (u−2a) du
−∞ −∞
Z∞
2iω a
=e Rxx (u) e−iω u du = e2iω a Sxx (ω )
−∞
Z∞ Z∞
Rxx (τ − 2a) e−iωτ d τ = Rxx (v) e−iω (u+2a) du
−∞ −∞
Z∞
−2iω a
=e Rxx (u) e−iω u du = e−2iω a Sxx (ω )
−∞
= 4 sin2 aω Sxx (ω )
Problem 5. Determine the autocorrelation function of a stationary random process
{X(t)} whose power spectral density function is given by
1, |ω | < a
(
Sxx (ω ) =
0 , otherwise
Power Spectrum: Power Spectral Density Functions • 221
Also show that the member functions X(t) and X t + πa of the process {X(t)} are
uncorrelated.
S OLUTION :
Let Rxx (τ ) be the autocorrelation function of a stationary random process {X(t)}.
Then we know that
Z∞
1
Rxx (τ ) = Sxx (ω ) eiτω d ω
2π
−∞
Za
1
= (1) eiτω d ω
2π
−a
1 ei τ
a
=
2π iτ −a
1 eiτ a − e−iτ a
=
2π iτ
1 2i sin aτ sin aτ
= =
2π iτ πτ
Consider the autocorrelation
π
π o π sin a
a = a sin π = 0
n
E X(t) X t + = Rxx = π
a a πa π2
Again consider the autocovariance
n π o n π o n π o
Cxx X(t) X t + = E X(t) X t + − E {X(t)} E X t +
a a a
π
= Rxx − (0)(0) = 0
a
π
Since covariance of X(t) and X t + is zero, the member functions X(t) and
π a
X t+ of the process {X(t)} are uncorrelated.
a
Problem 6. If {X(t)} and {Y (t)} are two stationary processes with the cross-
power spectral density function given by
a + (ibω /α ), |ω | ≤ α
(
Sxy (ω ) =
0, otherwise
where α > 0, a and b are constants then obtain the crosscorrelation function.
222 • Probability and Random Processes for Engineers
S OLUTION :
If {X(t)} and {Y (t)} are two stationary processes with crosscorrelation function
Rxy (τ ), then crosscorrelation function Rxy (τ ) can be obtained as the Fourier inverse
transform of Sxy (ω ) and is given by
Z∞
1
Rxy (τ ) = Sxy (ω )e+iτ ω d ω
2π
−∞
Zα
ibω
1
= a+ e+iτ ω d ω
2π α
−α
Zα Zα
1 1 ibω +iτ ω
= a e+iτ ω d ω + e dω
2π 2π α
−α −α
Zα Zα
a ib
= e+iτ ω d ω + ω e+iτ ω d ω
2π 2πα
−α −α
α α
a e+iτ ω
+iτ ω
e+iτ ω
ib e
= + ω − (1)
2π iτ −α 2πα iτ (iτ )2 −α
a e+iτ α e−iτ α
+iτ α
e+iτ α e−iτ α e−iτ α
ib e
= − + α − − − α −
2π iτ iτ 2πα iτ (iτ )2 iτ (iτ )2
a e+iτ α − e−iτ α
+iτ α
+ e−iτ α
+iτ α
− e−iτ α
ib e e
= + α −
2π iτ 2πα iτ (iτ )2
Problem 7. Obtain the power spectral density function of the output process {Y (t)}
corresponding to the input process {X(t)} in the system that has an impulse response
h (t) = e−β t U(t).
Power Spectrum: Power Spectral Density Functions • 223
S OLUTION :
Let Sxx (ω ) and Syy (ω ) be the power spectral density functions of the processes
{X(t)} and {Y (t)} respectively. Then we know that by Property 5 of PSD
Z∞
#∞
e−(β +iω )t
"
−(β +iω )t
= e dt =
−(β + iω )
0 0
1
=
(β + iω )
1 1 1
⇒ |H(ω )| = H(ω )H (ω ) =
2 ∗
=
(β + iω ) (β − iω ) β 2 + ω2
1
∴ Syy (ω ) = Sxx (ω )
β 2 + ω2
S OLUTION :
It is given that
1
Sxx (ω ) =
(1 + ω 2 )2
Z∞
1
Rxx (τ ) = Sxx (ω )ei τω d ω
2π
−∞
Z∞
1 1
= ei τω d ω
2π (1 + ω 2 )
2
−∞
224 • Probability and Random Processes for Engineers
Z∞
1 1
= (cos τω + i sin τω ) d ω
2π (1 + ω 2 )
2
−∞
Z∞ Z∞
1 1 1 1
= cos τω d ω + i sin τω d ω
2π (1 + ω 2 )
2 2π (1 + ω 2 )
2
−∞ −∞
Z∞
1 1 1
= cos τω d ω + 0 ∵ i sin τω is odd
2π (1 + ω 2 )
2
(1 + ω 2 )
2
−∞
Z∞
1 1
= cos τω d ω
2π (1 + ω 2 )
2
−∞
Z∞
1 1 (1 + τ ) e−τ
Rxx (τ ) = cos τω d ω =
2π (1 + ω 2 )
2 4
−∞
S OLUTION :
We know that the power spectral density function of a stationary random process
{X(t)} with autocorrelation function Rxx (τ ) is given by
Z∞
Sxx (ω ) = Rxx (τ ) e−i ωτ d τ
−∞
Consider
cos[b(t + τ ) + bt + 2θ )] + cos bτ
2
=a E
2
a2 a2
= E {cos[b(t + τ ) + bt + 2θ )]} + E {cos bτ }
2 2
1
f (θ ) = , 0 ≤ θ ≤ 2π
2π
Consider
Z2π
1
E {cos[b(t + τ ) + bt + 2θ )]} = cos[b(t + τ ) + bt + 2θ )] dθ
2π
0
2π
sin(2bt + bτ + 2θ )
1
=
2π 2 0
1
= {sin(2bt + bτ + 4π ) − sin(2bt + bτ )}
4π
1
= {sin(2bt + bτ ) − sin(2bt + bτ )} = 0
4π
Consider
Z2π
a2 a2 1
E {cos bτ } = cos bτ dθ
2 2 2π
0
a2
= cos bτ
2
a2
Rxx (τ ) = cos bτ
2
Z∞ 2
a
∴ Sxx (ω ) = cos bτ e−i ωτ d τ
2
−∞
a2
= Fourier transform of cos bτ
2
226 • Probability and Random Processes for Engineers
a2 iτ b a2
= e + e−iτ b = cos bτ
4 2
R∞
This is true because φ (x)δ (x − c)dx =φ (c)
−∞
a2 π a2
∴ F cos bτ = [δ (ω − b) + δ (ω + b)]
2 2
a2 π
= {δ (ω − b) + δ (ω + b)}
2
(Refer to Result A.6.1 in Appendix A for more details of this result.)
Problem 10. Find the power spectral density function of the random process
2
whose autocorrelation function is given by R(τ ) = e−aτ cos bτ where a and b
are constants.
S OLUTION :
It is given that
2
R(τ ) = e−aτ cos bτ
We know that the power spectral density function of a stationary random process
{X(t)} is given by
Z∞
Sxx (ω ) = Rxx (τ ) e−i ωτ d τ
−∞
Z∞
2
= e−aτ cos bτ e−i ωτ d τ
−∞
Z∞
eibτ + e−ibτ
−aτ 2
= e e−iωτ d τ
2
−∞
Z∞
1 2
e−aτ e−i(ω −b)τ + e−i(ω +b)τ d τ
=
2
−∞
Power Spectrum: Power Spectral Density Functions • 227
Z∞ h i h i
1 − aτ 2 +i(ω −b)τ − aτ 2 +i(ω +b)τ
= e +e dτ
2
−∞
Z∞ h i Z∞ h i
1 − aτ 2 +i(ω −b)τ 1 − aτ 2 +i(ω +b)τ
= e dτ + e dτ
2 2
−∞ −∞
Consider
Z∞ h i
1 − aτ 2 +i(ω −b)τ
e dτ
2
−∞
Add and subtract [i(ω − b)/2a]2 to make the exponent a perfect square. This gives
Z∞ n o Z∞ n o
1 − aτ 2 +i(ω −b)τ )2 −a τ 2 +2[i(ω −b)/2a]τ +[i(ω −b)/2a]2 −[i(ω −b)/2a]2
e dτ = e dτ
2
−∞ −∞
2 /4a Z∞
e−(ω −b) 2
= e−a{τ +i(ω −b)/2a} d τ
2
−∞
a τ + i(ω2a−b)
√
Let u = ⇒ dτ = du
√
a
Z∞ 2 Z∞
e−(ω −b) /4a
n o
1 − aτ 2 +i(ω −b)τ )2 2
⇒ e dτ = √ e−u du
2 2 a
−∞ −∞
2 /4a Z∞
e−(ω −b) 2
= √ e−u du
a
0
√ dv −1/2
Now let v = u2 ⇒ dv = 2udu = 2 vdu ⇒ du = 2v
2 /4a Z∞ 2 Z∞
e−(ω −b) −u2 e−(ω −b) /4a 1
⇒ √ e du = √ e−v v 2 −1 dv
a 2 a
0 0
2
e−(ω −b) /4a
= √ Γ(1/2)
2 a
1 π −(ω −b)2 /4a
r
= e
2 a
228 • Probability and Random Processes for Engineers
Problem 11. If {X(t)} is a random process such that X(t) = Y (t)Z(t), where Y (t)
and Z(t) are independent wide sense stationary processes. Then show that
(i) Rxx (τ ) = Ryy (τ )Rzz (τ )
∞
(ii) Sxx (ω ) = 1 R
2π Syy (a)Szz (ω − a) da
−∞
S OLUTION :
(i) We know that since {Y (t)} and {Z(t)} are independent stationary processes,
and X(t) = Y (t)Z(t), the autocorrelation function Rxx (t, t + τ ) can be given as
(ii) We know that the power spectral density function of {X(t)} given by Sxx (ω )
is the Fourier transformation of Rxx (τ ). That is,
Z∞
1
= Syy (a)Szz (ω − a) eiωτ dad ω
2π
−∞
∂a ∂a
y ∂ z dydz = 1 0 = dydz
dad ω = ∂∂ ω ∂ω 0 1
∂y ∂z
Z∞ Z∞
1
∴ F −1
Syy (a)Szz (ω − a) da = Syy (y)Szz (z) ei(y+z)τ dydz
2π
−∞ −∞
Z∞
1
= Syy (y)eiyτ dy
2π
−∞
Z∞
Szz (z) eizτ dz
−∞
= F −1 Syy (ω ) 2π F −1 {Szz (ω )} ,
by letting y, z = ω
= 2π Ryy (τ )Rzz (τ )
Z∞
1
F Ryy (τ )Rzz (τ ) = Syy (a)Szz (ω − a) da
∴
2π
−∞
Z∞
1
= Syy (a)Szz (ω − a) da
2π
−∞
S OLUTION :
It is given that the process {X(t)} is stationary Gaussian random process with
mean E {X(t)} = 0, variance V {X(t)} = σx2 (say) and autocorrelation function
Rxx (τ ). Now,
n o
E {X(t)} = 0, ⇒ V {X(t)} = σx2 = E X 2 (t) = Rxx (0)
n o
∴ E {Y (t)} = E X 2 (t) = Rxx (0)
230 • Probability and Random Processes for Engineers
Consider
Z∞
Syy (ω ) = Ryy (τ ) e−i ωτ d τ
−∞
Z∞ n
R2xx (0) + 2R2xx (τ ) e−iωτ d τ
o
=
−∞
Z∞ Z∞
−iωτ
2
= Rxx (0)e d τ + 2 R2xx (τ )e−iωτ d τ
−∞ −∞
where δ (ω ) is the unit impulse function. This is true because we know that F −1
2πα 2 δ (ω ) = α 2 .
Now, consider
Z∞
1
F −1 {Sxx (ω ) ∗ Sxx (ω )} = {Sxx (a)Sxx (ω − a)da} eiωτ d ω
2π
−∞
Z∞
1
= Sxx (a)Sxx (ω − a)eiωτ dad ω
2π
−∞
∂a ∂a
1 0
dad ω = ∂ u ∂ v dudv = = dudv
∂ω ∂ω 0 1
∂u ∂v
Z∞ Z∞
1
F −1
{Sxx (ω ) ∗ Sxx (ω )} = Sxx (u)Sxx (v)ei(u+v)τ du dv
2π
−∞ −∞
Z∞
1
= Sxx (u)eiyτ du
2π
−∞
Z∞
Sxx (v) eizτ dv
−∞
= F −1 Syy (ω ) 2π F −1 {Szz (ω )} ,
by letting y, z = ω
= 2π Rxx (τ )Rxx (τ )
1
∴ F {Rxx (τ )Rxx (τ )} = Sxx (ω ) ∗ Sxx (ω )
2π
1
∴ Syy (ω ) = 2π R2xx (0)δ (ω ) + Sxx (ω ) ∗ Sxx (ω )
π
EXERCISE PROBLEMS
1. Given that the random process {X(t)} is a wide sense stationary process
whose power spectral density function traps an area of 12.5 square units in
the first quadrant. Find the power of the random process.
2. Find the power spectral density function of the random process whose auto-
2
correlation function is given by R(τ ) = e−τ .
3. Determine the autocorrelation function of a stationary random process {X(t)}
whose power spectral density function is given by
k, |ω | < a
(
Sxx (ω ) =
0 , otherwise
where k is a constant.
232 • Probability and Random Processes for Engineers
1 − |τ | , −1 < τ < 1
(
Rxx (τ ) =
0, otherwise
Show that the power spectral density function is given in the form Sxx (ω ) =
sin ω /2 2
.
ω /2
5. If the power spectral density of a stationary random process {X(t)} is given by
1 + ω2 , | ω | ≤ 1
(
S(ω ) =
0, |ω | > 1
a + ibω , |ω | ≤ 1
(
Sxy (ω ) =
0, otherwise
where α > 0, a and b are constants then obtain the crosscorrelation function.
9. Given that a process {X(t)} has the autocorrelation function Rxx (τ ) = A e−a| τ |
cos bτ where A > 0, a > 0 and b are real constants, then find the power spec-
tral density function of {X(t)}.
10. If {X(t)} is a stationary process with power spectral density function
given by
ω , |ω | ≤ 1
( 2
Sxy (ω ) =
0 , otherwise
9.0 INTRODUCTION
Let us assume that a man starts from one of the k locations, say L0,1 , L0,2 , · · · , L0,k ,
at time point t0 and moves to one of the locations, say L1,1 , L1,2 , · · · , L1, k , or
remains at the same location at time point t1 and from there he further moves to
one of the next locations, say L2,1 , L2,2 , · · · , L2, k , or remains at the same location
at time point t2 and so on and reaches one of the locations, say Ln−1,1 , Ln−2,2 , · · · ,
Ln−1, k , or remains at the same location at time point tn−1 from where he moves
finally to one of the locations Ln,1 , Ln,2 , · · · , Ln, k or remains at the same location
at time point tn . This implies that in the given interval of time, say (0, t), at every
point of time t0 , t1 , · · · · · ·tn−1 ,tn ∈ (0, t) the person has the choice of moving to
one of the k − 1 predefined states (locations) or remain at the same location. This
means that at every point of time he has k options (locations) to choose. If we rep-
resent X(t, ξ ) as the person is in state (location) ξ at time point t, then one such a
random move, X(t, ξ1 ) of X(t, ξ ), is shown in Figure 9.1. Now, if being in one of
the states at time point tn depends only on the state (location) where he was at time
point tn−1 , then we say that the man’s process of making random moves towards a
final state is a Markov process. Otherwise, it is not a Markov process.
X(t, ξ)
L0k L1k Ln−1,k Ln,k
• • • •
• • • •
Location (state) • • • •
Let us see another example. In this example, a person involves in a game being
played at time points t0 , t1 , · · · · · ·tn−1 ,tn ∈ (0, t). At every point of time, he gets
Rs. 100 if he wins and he gives Rs. 50 if he loses. It is clear that the amount he owns
(present state) at any point of time depends on what the amount he had (previous
state) in the immediate past time point. Suppose he has Rs. 1000 at a point of time,
then the amount he is going to have after the next game will be either Rs. 1100 or
Rs. 950. Alternatively, the amount he had in the previous game was Rs. 900 and
he won to have Rs. 1000 or he had Rs. 1050 and lost Rs. 50 to have Rs. 1000. In
this example also, the process of man having an amount of money after a game
was played (present state of the process) at a particular time point depends on
how much money the man had after the game (past state of the process) in the
immediate past time. Therefore, this process can be termed a Markov process.
Consider the next example where a programmer writes code for an algorithm.
Every time point he completes the code and the program is run and its ability is
checked to see whether the program works well (state of the process). Next correc-
tions/improvements are done accordingly and the program is run at this time point.
Here, if the state that the program works well at a particular time point depends on
the state that how the program worked in the previous run in the immediate past
time then the process can be termed a Markov process otherwise it cannot be a
Markov process.
It may be noted that in all these examples the happening of the state of the
process at any point of time is quite random and is observed over a period of time.
Therefore, by nature, Markov process is a random process as it is time dependent.
Or otherwise, a random process becomes a Markov process under the condition
that the state of the process at any point of time depends only on the state of the
process at immediate past time.
It may be noted that a Markov process is indexed by state space, ξ , and time
parameter, t.
Here, P {Xn = j/Xn−1 = i} means the probability that the process that was in state
i in (n − 1)th step moved to state j in nth step.
Or
(n)
Pi j = P {Xm+n = j/Xm = i} , m, n = 1, 2, 3, · · · · · · · · · ;
i, j = 1, 2, 3, · · · · · · · · · , k
i, j = 1, 2, 3, · · · · · · · · · , k
(n)
Pi j = P {Xn = j/X0 = i} = P {Xm+n = j/Xm = i} , m, n = 1, 2, 3, · · · · · · · · · ;
i, j = 1, 2, 3, · · · · · · · · · , k
1 2 ··· j ··· k
(1) (1) (1) (1)
1 P11 P12 ··· P1 j ··· P1k
(1) (1) (1) (1)
2 P21 P22 P2i P2k
··· ···
.. .. .. .. .. .. ..
. . . . . . .
P(1) =
(1) (1) (1) (1)
i P
i1 Pi2 · · · Pi j · · · Pik
.. .. .. .. .. .. ..
. . . . . . .
(1) (1) (1) (1)
k Pk1 Pk2 · · · Pk j · · · Pkk
Here
(1)
(i) 0 ≤ Pi j ≤ 1, ∀ i, j
k (1)
(ii) ∑ Pi j = 1, ∀i
j=1
(1)
If Pi j = 0, then state j is not accessible from state i in one step.
Similarly, for n = 2, 3, · · · · · · we get two-step transition probability matrix P(2)
and three-step transition probability matrix P(3) and so on.
Note:
(i) A matrix is said to be stochastic matrix if each row of probabilities adds
to one.
(ii) A stochastic matrix say P, is said to be a regular matrix, if all the elements
of P(n) , for some n, are greater than zero.
(iii) A homogeneous Markov chain is said to be regular if its transition probabi-
lity matrix is regular.
Let us suppose that the one-step transition probability matrix with three states
1, 2 and 3 is given as follows:
1 2 3
1 0.3 0.7 0.0
P(1) = 2 0.6 0.0 0.4
3 0.5 0.5 0.0
Based on the one-step transition probability matrix, the transitions from one state
to one or more available states can be depicted using a transition diagram as shown
in Figure 9.2. In this figure, the possible transitions among three states 1, 2 and 3
are shown.
2
0.7
0.5
0.3 0.4
1 0.6
0.5 3
Figure 9.2. Transition diagram showing transitions among three states 1, 2 and 3 and the
corresponding transition probabilities
In Figure 9.2, state 1 is accessible from states 2, 3 and its own, state 2 is accessible
from 1 and state 3 whereas state 3 is accessible only from state 2 all in one step.
This can be guessed from looking at the columns of the transition matrix. This
(1) (1) (1) (1) (1) (1)
implies that P11 = 0.3, P21 = 0.6, P31 = 0.5; P12 = 0.7, P22 = 0.0, P32 = 0.5;
(1) (1) (1)
P13 = 0.0, P23 = 0.4, P33 = 0.0.
Higher order transition probability matrices
(1)
The first cycle in the transition diagram gives one-step transition probabilities Pi j ,
(2)
second cycle gives two-step transition probabilities Pi j and so on and nth cycle
(n)
gives the n-step transition probabilities Pi j . For example, higher order transition
probabilities and matrices can be obtained as follows.
Two-step transition probabilities (i.e., reachability of states in two steps):
(2) (2)
P11 = (0.3)(0.3) + (0.7)(0.6) = 0.51, P12 = (0.3)(0.7) = 0.21,
(2)
P13 = (0.7)(0.4) = 0.28
Markov Process and Markov Chain • 239
(2) (2)
P21 = (0.6)(0.3) + (0.4)(0.5) = 0.38, P22 = (0.6)(0.7) + (0.4)(0.5) = 0.62,
(2)
P23 = 0
(2) (2) (2)
P31 = (0.5)(0.3) + (0.5)(0.6) = 0.45, P32 = (0.5)(0.7) = 0.35, P33 = 0.20
+ (0.7)(0.4)(0.5) = 0.419,
(3)
P12 = (0.7)(0.4)(0.5) + (0.7)(0.6)(0.7) + (0.3)(0.3)(0.7) = 0.497,
(3)
P13 = (0.3)(0.7)(0.4) = 0.084,
(3)
P21 = (0.4)(0.5)(0.3) + (0.6)(0.7)(0.6) + (0.4)(0.5)(0.6)
+ (0.6)(0.3)(0.3) = 0.486,
(3)
P22 = (0.6)(0.3)(0.7) + (0.4)(0.5)(0.7) = 0.266,
(3)
P23 = (0.4)(0.5)(0.4) + (0.6)(0.7)(0.4) = 0.248,
(3)
P31 = (0.5)(0.4)(0.5) + (0.5)(0.3)(0.3) + (0.5)(0.6)(0.3)
+ (0.5)(0.7)(0.6) = 0.445,
(3)
P32 = (0.5)(0.4)(0.5) + (0.5)(0.3)(0.7) + (0.5)(0.6)(0.7) = 0.415,
(3)
P33 = (0.5)(0.7)(0.4) + (0.5)(0.3)(0.7) + (0.5)(0.6)(0.7) = 0.140
(0) (0)
π1 πi ••••••
(0) (1)
Pi1 (1)
π2 π1 ••••••
(1)
Pi2
•••
•••
(1)
(1) Pij (1)
π2 πj ••••••
•••
•••
(1)
Pik
(0) (1)
πk πk ••••••
(n)
Figure 9.3. State probabilities (πi , i = 1, 2, 3, · · · · · · k, n = 0, 1, 2, 3, · · · · · · ) and transition prob-
(n)
abilities (Pi j , n = 1, 2, 3, · · · · · · · · · ; i, j = 1, 2, 3, · · · · · · · · · , k)
Further, given the one-step transition probability matrix P(1) and if π (0) =
(0) (0) (0) (0)
[π1 , π2 , · · · , πi , · · · , πk ]
is the initial state probability distribution, then the
state probability distribution after one step π (1) , two steps π (2) and so on n steps
π (n) and can be obtained as
where P(n) is the n-step transition probability matrix. The state probability distri-
butions can also be obtained by using the relationship
Note:
A Markov chain {Xn } , n = 1, 2, 3, · · · · · · · · · is said to be completely specified,
if initial probability distribution and transition probability matrix are known.
242 • Probability and Random Processes for Engineers
(n)
Or otherwise, the (i, j)th element Pi j of n-step transition probability matrix P(n)
(1)
is equal to the (i, j)th element of nth power of one-step transition matrix Pi j , that
n on
is the (i, j)th element of P(1) .
Here, state j can be reached from state i in two steps, meaning that the chain starts
initially from state i and moves to an intermediate state, say k, in the first step and
then from k, it moves further to state j in the second step. Therefore, we have
(2)
Pi j = P {X2 = j/X0 = i}
= P {X2 = j/X1 = k} P {X1 = k/X0 = i}
= P {X1 = k/X0 = i} P {X2 = j/X1 = k}
(1) (1)
= Pik Pk j
That is,
(2) (1) (1) (1) (1) (1) (1)
Pi j = Pi1 P1 j + Pi2 P2 j + Pi3 P3 j + · · · · · ·
∞
∑ Pik
(1) (1)
= Pk j
k=1
Markov Process and Markov Chain • 243
Expanding ∑∞
(1) (1) th (2)
k=1 Pik Pk j , we get the (i, j) element Pi j , i = 1, 2, 3, · · · · · · · · · , k;
j = 1, 2, 3, · · · · · · · · · , k, of two-step transition probability matrix P(2) as the (i, j)th
element of the square of the one-step transition probability matrix P(1) , that is, the
n o2
(i, j)th element of P(1) . Hence,
n o2
P(2) = P(1)
Here, state j can be reached from state i in three steps. Let us assume that the chain
starts initially from state i and moves to an intermediate state, say k, in two steps
and then from k, it moves further to state j in the third step. Therefore, we have
(3)
Pi j = P {X3 = j/X0 = i}
= P {X3 = j/X2 = k} P {X2 = k/X0 = i}
= P {X0 = k/X2 = i} P {X3 = j/X2 = k}
(2) (1)
= Pik Pk j
That is,
(3) (2) (1) (2) (1) (2) (1)
Pi j = Pi1 P1 j + Pi2 P2 j + Pi3 P3 j + · · · · · ·
∞
∑ Pik
(2) (1)
= Pk j
k=1
Expanding ∑∞
(2) (1) th (3)
k=1 Pik Pk j , we get the (i, j) element Pi j , i = 1, 2, 3, · · · · · · · · · , k;
j = 1, 2, 3, · · · · · · · · · , k, of the three-step transition probability matrix P(3) as the
(i, j)th element of the cubic power of the one-step transition probability matrix
n o3
P(1) , that is the (i, j)th element of P(1) . Hence,
n o3
P(3) = P(1)
the n-step transition probability matrix P(n) , as the (i, j)th element
n of the
th
o n power
n
of one-step transition matrix P(1) , that is the (i, j)th element of P(1) . Hence,
n on
P(n) = P(1)
(n) 1 q p
lim P = (9.10)
n→∞ p+q q p
Proof. Given the matrix p(1) , we know by matrix analysis, the characteristic equa-
tion of p(1) can be given as
C(λ ) = λ I − P(1) = 0
1 0 1− p p
= λ − =0
0 1 q 1−q
λ − (1 − p) −p
= =0
−q λ − (1 − q)
= (λ − 1)(λ − 1 + p + q) = 0
(n) 1 q p
∴ lim P =
n→∞ p+q q p
n o h i
(n) (n) (n) (n)
π = lim π (n) = lim π1 , lim π2 , · · · , lim πi , · · · , lim πk
n→∞ n→∞ n→∞ n→∞ n→∞
= [π1 , π2 , · · · , πi , · · · , πk ]
which is called the steady-state probability distribution of the Markov chain. This
means that
n o
(n)
lim πi = πi , i = 1, 2, · · · , k (9.11)
n→∞
n o
(n)
Here, while πi is called the limiting probability of the sequence πi , i=
1,
n 2, · ·
o · , k of state probabilities, π is called the limiting distribution of the sequence
π (n) , n = 1, 2, 3, · · · · · · of distributions.
Note:
k
π P(1) = π and ∑ πi = 1 (9.12)
i=1
246 • Probability and Random Processes for Engineers
k
∑
(m) (1) (m−1)
fii = Pi j f ji , m = 2, 3, · · · · · · , i = 1, 2, · · · · · · , k
j=1
j 6= i
Similarly, the probability that the chain goes to state j for the first time having
started from the state i after n steps, that is after making n transitions, is denoted
(n)
by fi j , n = 1, 2, 3, · · · · · · · · · where
k
∑
(m) (1) (m−1)
fi j = Pis fs j , m = 2, 3, · · · · · · , i = 1, 2, · · · · · · , k; j=1, 2, · · · · · · , k,
s=1
s 6= j
Nii = n 1 2 3 ······
(1) (2) (3)
P {Nii = n} fii fii fii ······
248 • Probability and Random Processes for Engineers
Therefore, having started from a state i, the average number of transitions (steps)
made by the chain before returning to the same state i for the first time, say µii =
E {N ii }, is known as the mean recurrent time of state i and is calculated by
∞
∑ n fii
(n)
µii = E {Nii } = (9.14)
n=1
Here, state i is said to be periodic with period d(i) if d(i) > 1 and is said to be
aperiodic if d(i) = 1.
This results in the fact that for an irreducible, aperiodic, recurrent Markov
chain, the steady-state distribution is unique and can be given as
1
πi = (9.16)
µii
Clearly, if µii < ∞, that is non-null persistent, we have πi > 0 and if µii = ∞, that
is null persistent, we have πi = 0 and vice-versa.
SOLVED PROBLEMS
Problem 1. In tossing of a fair coin, let {Xn } , n = 1, 2, 3, · · · denote the outcome
of nth toss, then show that the process of getting total number of heads in the first n
trials is a Markov process. Or using a suitable example show that Bernoulli process
is a Markov process.
S OLUTION :
Let us suppose that we observe a sequence of random variables {Xn} , n=1, 2, 3, · · ·
with probability 1/2 when head turns (say 1) and with probability 1/2 when tail
turns (say 0). That is,
1
P (Sn+1 = x + 1/Sn = x) =
2
1
P (Sn+1 = x/Sn = x) =
2
Therefore, the outcome at (n + 1)th trial is affected only by the outcome of the
nth trial but not by the outcome up to (n − 1)th trial. Hence, it is concluded that
in tossing of a fair coin, the process of getting total number of heads in the first n
trials is a Markov process. Therefore, Bernoulli process is a Markov process.
250 • Probability and Random Processes for Engineers
S OLUTION :
Let {X(t)} be a Poisson process defined in the interval (0, t). Then {X(t) = n}
or n(0,t) = n represents that there are n Poisson points in the time interval (0, t).
That is,
eλ t (λ t)n
P {X(t) = n} = , n = 0, 1, 2, · · ·
n!
or
n3
n2
n1
0 t1 t2 t3 t
n2 − n1 n3 − n2
This means that the number of Poisson points occurred at time point t3 depends
only on the number of Poisson points occurred at the most recent time point t2 .
Now,
= P {n(t2 ,t3 ) = n3 − n2 }
1/3
1/3
1/3 2
1
1/4
3/4
1/3 3
2/3
S OLUTION :
The state space of the given Markov chain has three states 1, 2 and 3. Therefore,
from the transition diagram, the one-step transition probability matrix can be given
as follows:
(1) (1) (1)
P11 P12 P13
1/3 1/3 1/3
P(1) = P21
(1) (1) (1)
P22 P23 = 1/4 0 3/4
(1) (1) (1) 1/3 0 2/3
P31 P32 P33
S OLUTION :
∞ (n)
In order to show that state 1 is recurrent we have to show that ∑ f11 = 1.
n=1
Consider
∞
∑ f11
(n) (1) (2) (3)
= f11 + f11 + f11 + · · · · · ·
n=1
(1) (1)
Here, fi j = Pi j , ∀ i, j
From the given transition probability matrix, we have
∑ P1k
(2) (1) (1) (1) (1)
⇒ f11 = fk1 = P12 f21 = (0) (0.25) = 0
j=2
∑ P2k
(2) (1) (1) (1) (1)
f22 = fk2 = P21 f12 = (0.25) (0) = 0
j=1
Now,
∑ P1k
(2) (1) (1) (1) (1)
f12 = fk2 = P11 f12 = (1) (0) = 0
j=1
∑ P2k
(2) (1) (1) (1) (1)
f21 = fk1 = P22 f21 = (0.75) (0.25) = 0.1875
j=2
Now, consider
∑ P1k
(3) (1) (2) (1) (2)
f11 = fk1 = P12 f21 = (0) (0.1875) = 0
j=2
∑ P1k
(3) (1) (2) (1) (2)
f22 = fk1 = P21 f12 = (0.25) (0) = 0
j=1
(n) (n)
Note that f11 = 0 and f22 = 0 for n ≥ 2. Therefore, we have
∞
∑ f11
(n) (1) (2) (3)
= f11 + f11 + f11 + · · · · · · = 1 + 0 + 0 + · · · · · · = 1
n=1
S OLUTION :
(0)
It is given that πi = P {X0 = i} = 41 , i = 1, 2, 3, 4
1 1 1 1
(0) (0) (0) (0)
⇒ π (0)
= π1 , π2 , π3 , π4 = , , ,
4 4 4 4
(1)
where Pi j = P {Xn = j/Xn−1 = i} , i, j = 1, 2, 3, 4, n = 1, 2, 3, · · · · · ·
According to Chapman-Kolmogorov theorem, the two-step transition probability
matrix P(2) can be obtained as
(2) (2) (2) (2)
P11 P12 P13 P14
0 1 0 0 0 1 0 0
(2) (2) (2) (2)
P
21 P22 P23 P24 0.3 0 0.7 0 0.3 0 0.7 0
(2)
P = =
P(2) P(2) P(2) P(2) 0 0.3 0 0.7 0 0.3 0 0.7
31 32 33 34
(2) (2) (2) (2) 0 0 1 0 0 0 1 0
P41 P42 P43 P44
254 • Probability and Random Processes for Engineers
(2)
where Pi j = P {Xn = j/Xn−2 = i} , i, j = 1, 2, 3, 4, n = 2, 3, · · · · · ·
0.30 0.00 0.70 0.00
0.00 0.51 0.00 0.49
⇒ P(2) =
0.09 0.00 0.91 0.00
0 0.30 1.00 0.70
(i) Consider
(1)
P ( X1 = 2/ X0 = 1) = P12 = 1
(1)
P ( X2 = 2/ X1 = 1) = P12 = 1
(1)
P ( X2 = 1/ X1 = 2) = P21 = 0.3
(2)
P ( X2 = 2/ X0 = 2) = P22 = 0.51
4 (1) (1) (1) (1) (1) (1) (1) (1) (1) (1)
(ii) Consider ∑ P2k Pk2 = P21 P12 + P22 P22 + P23 P32 + P24 P42
k=1
4
∑ P2k Pk2
(1) (1)
= (0.3)(1) + (0)(0) + (0.7)(0.3) + (0)(0) = 0.51 (1)
k=1
(2)
P22 = 0.51 (2)
P (B), we have
P (X2 = 2, X1 = 3, X0 = 2) = P (X2 = 2, X1 = 3/X0 = 2) P (X0 = 2)
(0)
= (0.21) π2
1
= (0.21) = 0.0525
4
(v) Consider P (X3 = 4, X2 = 2, X1 = 3, X0 = 2). Using the formula P (A B) =
T
S OLUTION :
(0) (0) (0)
It is given that π (0) = π1 , π2 , π3 = (0.7, 0.2, 0.1)
(0) (0) (0)
⇒ P {X0 = 1} = π1 = 0.7, P {X0 = 2} = π2 = 0.2, P {X0 = 3} = π3 = 0.1
The one-step transition probability matrix P(1) is given as
(1) (1) (1)
P11 P12 P13
0.1 0.5 0.4
P(1) =
(1) (1) (1)
P21 P22 P23 = 0.6 0.2 0.2
(1) (1) (1) 0.3 0.4 0.3
P P
31 32P 33
256 • Probability and Random Processes for Engineers
(1)
where Pi j = P {Xn = j/Xn−1 = i} , i, j = 1, 2, 3, n = 1, 2, 3, · · · · · ·
According to Chapman-Kolmogorov theorem, the two-step transition probability
matrix P(2) can be obtained as
(2) (2) (2)
P11 P12 P13
0.1 0.5 0.4 0.1 0.5 0.4
P(2) = (2) (2) (2)
P21 P22 P23 = 0.6 0.2 0.2 0.6 0.2 0.2
(2) (2) (2) 0.3 0.4 0.3 0.3 0.4 0.3
P P P
31 32 33
0.43 0.31 0.26
= 0.24 0.42 0.34
(2)
where Pi j = P {Xn = j/Xn−2 = i} , i, j = 1, 2, 3, n = 2, 3, · · · · · ·
(i) Consider
3
∑ πk
(2) (0) (2)
P ( X2 = 3) = π3 = Pk3
k=1
(0) (2) (0) (2) (0) (2)
= π1 P13 + π2 P23 + π3 P33
= (0.7)(0.26) + (0.2)(0.34) + (0.1)(0.29) = 0.279
P (X3 = 2, X2 = 3, X1 = 3, X0 = 2)
= P (X3 = 2/X2 = 3, X1 = 3, X0 = 2) P (X2 = 3, X1 = 3, X0 = 2)
= P (X3 = 2/X2 = 3) P (X2 = 3, /X1 = 3, X0 = 2) P(X1 = 3, X0 = 2)
= P (X3 = 2/X2 = 3) P (X2 = 3, /X1 = 3) P(X1 = 3/X0 = 2)P(X0 = 2)
(1) (1) (1) (0)
= P32 P33 P23 π2
= (0.4)(0.3)(0.2)(0.2) = 0.0048
S OLUTION :
It is given that the Markov chain starts on day one (today). Then the probability
distribution that there is rain or no rain today gives the initial probability
distribution as
(0) (0) (0)
π (0) = π1 , π2 = (0.4, 0.6) ⇒ P {X0 = 1} = π1 = 0.4,
(0)
P {X0 = 2} = π2 = 0.6
where X0 = 1 stands for there is rain today and X0 = 2 for no rain today.
The one-step transition probability matrix P(1) is given as
(1) (1) !
(1)
P11 P12 0.6 0.4
P = =
(1) (1) 0.2 0.8
P21 P22
(1)
where Pi j = P {Xn = j/Xn−1 = i} , i, j = 1, 2 n = 1, 2, 3, · · · · · ·
(i) The probability that it will rain for three days from today assuming that it is
raining today can be obtained using the probabilities of transitions given by
P(rain for 3 days from today)
2
∑ πk
(3) (0) (3)
P ( X3 = 1) = π1 = Pk3
k=1
Consider
(2) (2) ! ! ! !
(2)
P11 P12 0.6 0.4 0.6 0.4 0.44 0.56
P = = =
(2) (2) 0.2 0.8 0.2 0.8 0.28 0.72
P21 P22
(3) (3) ! ! ! !
P11 P12 0.44 0.56 0.6 0.4 0.376 0.624
P(3) = = =
(3) (3) 0.28 0.72 0.2 0.8 0.312 0.688
P21 P22
∴ P ( X3 = 1) = (0.4)(0.376) + (0.6)(0.312) = 0.3376
(iii) Since X0 = 2, X1 = 2, X2 = 2, X3 = 2 represent respectively no rain on first
day, second day, third day and fourth day, the probability that there will be
no rain after three days (that is fourth day) can be obtained as
2
∑ πk
(3) (0) (3)
P ( X3 = 2) = π2 = Pk2
k=1
(0) (3) (0) (3)
= π1 P12 + π2 P22
∴ P ( X3 = 2) = (0.4)(0.624) + (0.6)(0.688) = 0.6624
Note:
(3) (3)
Alternatively, we can obtain P ( X3 = 1) = π1 and P ( X3 = 2) = π2 using the
relationship,
(0)
π (3) = π1(3) , π2(3) = π1 P(3)
0.376 0.624
= 0.4 0.6
0.312 0.688
(3)
⇒ P ( X3 = 1) = π1 = (0.4)(0.376) + (0.6)(0.312) = 0.3376
(3)
P ( X3 = 2) = π2 = (0.4)(0.624) + (0.6)(0.688) = 0.6624
S OLUTION :
(i) We know that if the one-step
transition probability matrix is of the form
1 − p p
p(1) = , 0 ≤ p, q ≤ 1 then the n-step TPM can be given
q 1−q
as (Refer to (9.9))
Markov Process and Markov Chain • 259
(n) 1 q p n p −p
P = + (1 − p − q)
p+q q p −q q
0.9 0.1
We have P(1) = ⇒ p = 0.1, q = 0.2
0.2 0.8
(" # " #)
(n) 1 0.2 0.1 n
0.1 −0.1
∴ P = + (1 − 0.1 − 0.2)
0.1 + 0.2 0.2 0.1 −0.2 0.2
" # " #
1 0.2 0.1 (0.7)n 0.1 −0.1
= +
0.3 0.2 0.1 0.3 −0.2 0.2
2 + (0.7)n 1 − (0.7)n
3 3
= 2 − 2(0.7)n 1 + 2(0.7)n
3 3
2 + (0.7)n 1 − (0.7)n
2 1
(ii) lim P(n) = lim
3 3 = 3
3 , ∵ lim (0.7)n = 0
n→∞ 2 − 2(0.7)n 1 + 2(0.7)n 2 1
n→∞ n→∞
3 3 3 3
(iii) We know that steady-state probability distribution π (n) after n steps, if the
initial state probability distribution is π (0) , can be given as (refer (9.6))
2 1
lim π (n)
=
n→∞ 3 3
2 1
⇒ P(X∞ = 1) = , P(X∞ = 2) =
3 3
Problem 9. A man either drives a car or goes by train to office each day. He never
goes two days in a row by train. If he drives one day then he is just as likely to
drive again the next day as he is to travel by train. Now suppose that on the first
day of the week, the man tosses a fair die and then drives to work if and only if a 6
appears. Under this circumstance, (i) verify whether the process of going to office
is Markovian, (ii) obtain initial probability distribution and the one-step transition
probability matrix, (iii) what is the probability that he will go by train on the third
day and (iv) obtain the probability that he will drive to work in the long run.
S OLUTION :
(i) Since the decision process of either going by car or train on a particular day
depends on the mode of transport used just on the previous day, we conclude
that the process is Markovian.
(ii) Since the state space is discrete, the man’s travel pattern is clearly a Markov
chain {Xn } , n = 1, 2, 3, · · · · · · · · · with two states 1 and 2, where 1 stands for
travel by train and 2 stands for drive by car. Since the first day of the week
is the starting point the initial state distribution (the probability of going by
train or car on first day) can be given as
(0)
P {X0 = 1} = π1 = P {Going by train}
5
= P {Getting no six in the toss of a die} =
6
(0)
P {X0 = 2} = π2 = P {Going by car}
1
= P {Getting a six in the toss of a die} =
6
5 1
(0) (0)
⇒ π (0) = π1 , π2 = ,
6 6
Markov Process and Markov Chain • 261
nth day
train(1) car(2)
(1) (1) !
P11 P12
(1) train(1)
th 0 1
P = (n − 1) day =
car(2) 0.5 0.5 (1)
P21 P22
(1)
(1) (1) !
P11 P12
(1) 0 1
⇒ P = =
(1) (1)
P21 P22 0.5 0.5
(iii) Here, if the starting day is the first day of the week, then n = 1 means the
second day (after one step), n = 2 means the third day (after two steps), and
so on. That is, the second day is reached after one step, third day is reached
in two steps, and so on. Therefore, on nth day (that is after n − 1 steps), we
have
1 If the man goes by train
Xn−1 =
2 If the man goes by car
Hence, the probability that the man will go by train or car on the third day
(2)
can be computed by finding the state probabilities P {X2 = 1} = π1 and
(2)
P {X2 = 2} = π2 respectively as follows:
(2) (2)
π (2) = π1 , π2 = π (0) P(2)
Consider
(2) (2) ! ! ! !
(2)
P11 P12 0 1 0 1 0.5 0.5
P = = =
(2) (2) 0.5 0.5 0.5 0.5 0.25 0.75
P21 P22
!
0.5 0.5
5 1 11 13
∴ π =
(2)
=
6 6 0.25 0.75 24 24
Therefore, the probability that the person will travel by train on third day is
(2) 11
P {X2 = 1} = π1 =
24
(iv) The probability that the man will drive car in the long run can be obtained
by finding the steady-state probability distribution called the limiting distri-
bution π = (π1 , π2 ) , where π1 is the probability that the man will travel
by train in the long run and π2 is the probability that the man will travel
262 • Probability and Random Processes for Engineers
by car in the long run. We know that if the initial state probability distribu-
tion π (0) and one-step transition probability matrix P(1) are known, then the
steady-state probability distribution, can be obtained by using the equations
k
π P(1) = π and ∑ πi = 1
i=1
!
0.5 0.5
⇒ (π1 π2 ) = (π1 π2 ) and π1 + π2 = 1
0.25 0.75
Therefore, the probability that the man will travel by car in the long run
2
is .
3
Problem 10. Three persons A, B and C are throwing a ball to each other. A always
throws the ball to B and B always throws the ball to C, but C is just as likely to
throw the ball to B as to A. Show that the process is Markovian. Find the transition
matrix and classify the states.
S OLUTION :
Since the state space is discrete, and the state of ball being with A or B or C depends
on who was having the ball in the immediate past, the pattern of receiving the ball
is clearly a Markov chain {Xn } , n = 1, 2, 3, · · · · · · · · · with three states 1, 2 and 3
where 1 stands for the ball being with A, 2 stands for the ball being with B and 3
stands for the ball being with C. Therefore, the process is Markovian.
The one-step transition probability matrix becomes
nth step
A(1) B(2) C(3)
(1)
(1) (1)
A(1) P11 P12 P13
0 1 0
(1) (1) (1)
(n − 1)th step P21 P22 P23 = 0 0 1
B(2)
A 1 2 B
1/2
1
1/2 3
C
Classification of states
(i) Irreducibility
From the state transition diagram, we get
This implies state 1 is accessible from state 2 (in two steps) and from state
3 (in one step). The state 2 is accessible from both the states 1 and 3 (in one
step). The state 3 is accessible from state 1 (in two steps) and from state 2
(in one step). Since every state is accessible from every other state in some
step, the chain is irreducible.
(ii) Periodicity
From the state transition diagram, we get
(n)
P11 > 0 for n = 3, 5, 6 etc..
⇒ d(1) = GCD {3, 5, 6, · · · } = 1
(n)
P22 > 0 for n = 2, 3, 4, 5, 6 etc..
⇒ d(2) = GCD {2, 3, 4, 5, 6, · · · } = 1
(n)
P33 > 0 for n = 2, 3, 4, 5, 6 etc..
⇒ d(3) = GCD {2, 3, 4, 5, 6, · · · } = 1
0 1 0
π1 π2 π3 0 0 1 = π1 π2 π3 and π1 + π2 + π3 = 1
0.5 0.5 0
⇒ 0.5π3 = π1 , π1 + 0.5π3 = π2 , π2 = π3
⇒ 0.5π3 + π3 + π3 = 1 ⇒ 2.5π3 = 1 ⇒ π3 = 0.4
∴ π1 = 0.2, π2 = 0.4, π3 = 0.4
Now using the result given in (9.16), the mean recurrent times for the three
states can be obtained as
1 1
π1 = ⇒ 0.2 = ⇒ µ1 = 5 < ∞
µ1 µ1
1 1
π2 = ⇒ 0.4 = ⇒ µ2 = 2.5 < ∞
µ2 µ2
1 1
π3 = ⇒ 0.4 = ⇒ µ3 = 2.5 < ∞
µ3 µ3
This shows that all the three states are non-null persistent.
(iv) Ergodicity
Since all the three states are aperiodic and non-null persistent, the given
Markov chain is ergodic.
Problem 11. A fair dice is tossed repeatedly. If Xn denotes the maximum of the
numbers occurring in the first n tosses, then find the transition probability matrix of
the Markov chain. Also find the two-step transition probability matrix and hence
find the probability that the maximum of the numbers occurring in the first two
tosses, that is P {X2 = 6}.
S OLUTION :
We know that in case of the process of tossing a fair dice, the state space (possible
outcomes) is given as {1, 2, 3, 4, 5, 6}. It may be noted that in the initial toss one
of these six numbers can happen and it will be the maximum number. Therefore,
we have the initial state probability distribution as
(0) (0) (0) (0) (0) (0)
1 1 1 1 1 1
π (0)
= π1 , π2 , π3 , π4 , π5 , π6 = , , , , ,
6 6 6 6 6 6
(0)
That is, πi = P {X0 = i} = 61 , i = 1, 2, 3, 4, 5, 6
Markov Process and Markov Chain • 265
Let Xn = i denote the maximum of the numbers occurring in the first n tosses and
let Xn+1 = j be the maximum of the numbers occurring after (n + 1)th toss. Then
different possibilities are obtained as
If i = 1, then j = 1, 2, 3, 4, 5 and 6
1
⇒ P {Xn+1 = j/Xn = 1} = , j = 1, 2, 3, 4, 5, 6
6
If i = 2, then j = 2, 3, 4, 5 and 6
2 1
⇒ P {Xn+1 = 2/Xn = 2} = and P {Xn+1 = j/Xn = 2} = , j = 3, 4, 5, 6
6 6
If i = 3, then j = 3, 4, 5 and 6
3 1
⇒ P {Xn+1 = 3/Xn = 3} = and P {Xn+1 = j/Xn = 3} = , j = 4, 5, 6
6 6
If i = 4, then j = 4, 5 and 6
4 1
⇒ P {Xn+1 = 4/Xn = 4} = and P {Xn+1 = j/Xn = 4} = , j = 5, 6
6 6
If i = 5, then j = 5 and 6
5 1
⇒ P {Xn+1 = 5/Xn = 5} = , P {Xn+1 = 6/Xn = 5} =
6 6
If i = 6, then j=6
6
⇒ P {Xn+1 = 6/Xn = 6} = 1 =
6
Therefore, the probability that the maximum of the numbers occurring in the first
two tosses can be obtained as
6
∑ πi
(0) (2)
P {X2 = 6} = Pi6 .
i=1
(0) (2) (0) (2) (0) (2) (0) (2) (0) (2) (0) (2)
P {X2 = 6} = π1 P16 + π2 P26 + π3 P36 + π4 P46 + π5 P56 + π6 P66
1 11 1 11 1 11 1 11
P {X2 = 6} = + + +
6 36 6 36 6 36 6 36
1 11 1 36
+ +
6 36 6 36
1 1 91
P {X2 = 6} = (11 + 11 + 11 + 11 + 11 + 36) = = 0.4123
6 36 216
EXERCISE PROBLEMS
1. Let {Xn } be a Markov chain with state space {1, 2, 3} with initial probabil-
ity distribution π (0) = (1/4, 1/2, 1/4). If the one-step transition probability
matrix is given by
1/4 3/4 0
P = 1/3 1/3 1/3
0 1/4 3/4
(i) Draw the transition diagram (ii) Show that state 1 is periodic with period 2.
6. An air-conditioner is in one of the three states: off (state 1), low (state 2) or
high (state 3). If it is in off position, the probability that it will be turned to
low is 1/3. If it is in low position, then it will be turned either to off or high
with equal probabilities 1/4. If it is in high position, then the probability that
it will be turned to low is 1/3 or to off is 1/6. (i) Draw the transition diagram,
(ii) obtain the transition probability matrix and (iii) obtain the steady-state
probabilities.
7. Obtain the transition diagram and classify
the states of the Markov chain
0 1 0
with the transition probability matrix 1/2 0 1/2
0 1 0
268 • Probability and Random Processes for Engineers
Find (a) the transition diagram, (b) the steady-state probability distribution
and (c) classify the states.
9. A professor has three pet questions, one of which occurs in every test he
gives. The students know his habit well. He never gives a question twice in
a row. If he had given question one last time, he tosses a coin and gives ques-
tion two if head comes up. If he had given question two last time, he tosses
two coins and switches to question three, if both heads come up. If he had
given question three last time, he tosses three coins and switches to question
one if all three heads come up. (i) Show that the process is Markovian. (ii)
Write the transition probability matrix of the corresponding Markov chain.
(iii) Obtain the probabilities of the pet questions being given in the long run.
10. A man is at an integral part of x-axis between the origin and the point 3. He
takes unit steps to the right with probability 1/3 or to the left with probability
2/3 unless he is at the origin. If he is at the origin he takes a step to the right
to reach the point 1 or if he is at the point 3 he takes a step to the left to
reach the point 2. (i) Obtain the transition probability matrix, (ii) what is
the probability that he is at the point 1 after 2 walk if the initial probability
vector is (1/4, 1/4, 1/4, 1/4) and (iii) what is the probability that he is at
position 1 after a long run.
A PPENDIX A
SOME IMPORTANT RESULTS RELATED
TO RANDOM PROCESSES
Result A.1.1:
Cauchy Criterion: A random process {X(t)} is said to be continuous in mean
square sense if
n o
lim E [X(t + τ ) − X(t)] 2 = 0
τ →0
Result A.1.2:
A random process {X(t)} is said to be continuous if its autocorrelation function
Rxx (t1 , t2 ) is continuous.
− 2E {X(t + τ )X(t)}
= Rxx (t + τ ,t + τ ) + Rxx (t,t) − 2Rxx (t + τ ,t)
If the autocorrelation Rxx (t1 , t2 ) of the random process {X(t)} is continuous, then
we have
n o
lim E [X(t + τ ) − X(t)] 2 = lim {Rxx (t + τ , t + τ ) + Rxx (t, t) − 2Rxx (t + τ , t)} = 0
τ →0 τ →0
270 • Probability and Random Processes for Engineers
Similarly, it can be shown that if {X(t)} is continuous in mean square sense, then
Rxx (τ ) is continuous at τ = 0. Hence the proof.
Result A.1.4:
If {X(t)} is continuous in mean square sense, then its mean is continuous. That is,
lim µx (t + τ ) = µx (t)
τ →0
= {[µx (t + τ ) − µx (t)]}2
∴ lim µx (t + τ ) = µx (t)
ε →0
X(t + τ ) − X(t) ′
lim = X (t)
τ →0 τ
Result A.2.2:
′
A random process {X(t)} is said to have a derivative denoted as X (t) in mean
square sense if
X(t + τ ) − X(t)
2
′
lim E − X (t) =0
τ →0 τ
Result A.2.3:
A random process {X(t)} with autocorrelation function Rxx (t1 , t2 ) has a derivative
∂ 2 Rxx (t1 , t2 )
in mean square sense if exists at t = t1 = t2 .
∂ t1 ∂ t2
272 • Probability and Random Processes for Engineers
X(t + τ ) − X(t)
Proof. Let Y (t, τ ) = and with an assumption that t = t1 = t2 , let
τ
X(t1 + τ1 ) − X(t1 ) X(t2 + τ2 ) − X(t2 )
Y (t1 , τ1 ) = and Y (t2 , τ2 ) =
τ1 τ2
By Cauchy criterion, the mean square derivative of {X(t)} exists, if
n o
lim E [Y (t2 , τ2 ) −Y (t1 , τ1 )]2 = 0
τ1 ,τ2 →0
Consider
n o n o n o
E [Y (t2 , τ2 ) −Y (t1 , τ1 )]2 = E Y 2 (t2 , τ2 ) + E Y 2 (t1 , τ1 )
− 2E {Y (t2 , τ2 )Y (t1 , τ1 )} (A1)
Now, consider
Expanding we get
1
E {Y (t2 , τ2 )Y (t1 , τ1 )} = {Rxx (t2 + τ2 , t1 + τ1 ) − Rxx (t2 + τ2 , t1 )
τ1 τ2
− Rxx (t2 , t1 + τ1 ) + Rxx (t2 , t1 )}
Rxx (t2 + τ2 , t1 + τ1 ) − Rxx (t2 + τ2 , t1 )
1
lim E {Y (t2 , τ2 )Y (t1 , τ1 )} = lim
τ1 →0 τ2 τ1 →0 τ1
Rxx (t2 , t1 + τ1 ) − Rxx (t2 , t1 )
−
τ1
1 ∂ Rxx (t2 + τ2 ,t1 ) ∂ Rxx (t2 ,t1 )
lim lim E {Y (t2 , τ2 )Y (t1 , τ1 )} = lim −
τ2 →0 τ1 →0 τ2 →0 τ2 ∂ t1 ∂ t1
∂ 2 R(t2 ,t1 )
=
∂ t1 ∂ t2
∂ 2 R(t2 ,t1 )
∴ lim E {Y (t2 , τ2 )Y (t1 , τ1 } = (A2)
τ1 →0, τ2 →0 ∂ t1 ∂ t2
Appendix A • 273
∂ 2 R(t1 ,t2 )
This is true provided exists.
∂ t1 ∂ t2
Setting τ1 = τ2 with t = t1 = t2 , in (A2), we have
n o n o ∂ 2 R(t ,t )
1 2
lim E Y 2 (t1 , τ1 ) = lim E Y 2 (t2 , τ2 ) = (A3)
τ1 →0 τ2 →0 ∂ t1 ∂ t2
Substituting (A2) and A(3) in (A1), we have
n o ∂ 2 R(t ,t ) ∂ 2 R(t ,t ) ∂ 2 R(t1 ,t2 )
1 2 1 2
lim E [Y (t2 , τ2 ) −Y (t1 , τ1 )]2 = + −2 =0
τ1 ,τ2 →0 ∂ t1 ∂ t2 ∂ t1 ∂ t2 ∂ t1 ∂ t2
Therefore, we conclude that a random process {X(t)} with autocorrelation func-
∂ 2 Rxx (t1 ,t2 )
tion Rxx (t1 ,t2 ) has a derivative in mean square sense if exists at
∂ t1 ∂ t2
t = t1 = t2 . Further if {X(t)} is a stationary random process, then this is equiv-
∂ 2 Rxx (τ )
alent to the existence of at τ = 0. It may be noted that τ = |t1 − t2 |.
∂ t1 ∂ t2
A.3 INTEGRALS RELATED TO RANDOM PROCESSES
Result A.3.1:
A random process {X(t)} is said to be integrable if
Z b
S=
a
X(t) dt = lim
∆ti →0 i
∑ X(ti )∆ti
where a < t0 < t1 < t2 < · · · < ti < · · ·tn < · · · < b and ∆ti = ti+1 − ti .
Result A.3.2:
The random process {X(t)} is integrable in the mean square sense if the following
integral
Z bZ b
Rxx (ti ,t j )dti dt j
a a
exists for any ti and t j .
Proof. We know that the mean square integral of the random process {X(t)} is
given by
Z b
S=
a
X(t) dt = lim
∆ti →0 i
∑ X(ti )∆ti
Let
Z b
S=
a
X(t) dt = lim
∆ti →0 i
∑ X(ti )∆ti = ∆tlim
j →0
∑ X(t j )∆t j (A4)
j
where a < t0 < t1 < t2 < · · · < ti < · · ·tn < · · · < b, ∆ti = ti+1 −ti and ∆t j = t j+1 −t j .
274 • Probability and Random Processes for Engineers
!2 !2 ( )
= E ∑ X(ti )∆ti +E ∑ X(t j )∆t j − 2E ∑ X(ti )∆ti ∑ X(t j )∆t j (A5)
i j i j
! !
= E ∑ X(ti )∆ti ∑ X(ti )∆ti +E ∑ X(t j )∆t j ∑ X(t j )∆t j
i i j j
( )
−2E ∑ ∑ X(ti )X(t j )∆ti ∆t j
i j
2
By (A4), we have
2
∑∑E ∑∑E
= lim lim X(ti )∆ti X(t j )∆t j + lim lim X(ti )∆ti X(t j )∆t j
∆ti →0 ∆t j →0 i j ∆ti →0 ∆t j →0 i j
Appendix A • 275
∑∑E
− 2 lim lim X(ti )∆ti X(t j )∆t j
∆ti →0 ∆t j →0 i j
= lim
∆ti →0,∆t j →0 i
∑ ∑ Rxx (ti , t j )∆ti ∆t j + ∆ti →0,∆t
lim
j →0
∑ ∑ Rxx (ti , t j )∆ti ∆t j
j i j
−2 lim
∆ti →0,∆t j →0 i
∑ ∑ Rxx (ti , t j )∆ti ∆t j
j
Z bZ b Z bZ b
= Rxx (ti ,t j ) dti dt j + Rxx (ti ,t j ) dti dt j (A6)
a a a a
Z bZ b
−2 Rxx (ti ,t j ) dti dt j = 0
a a
2
Therefore, the random process {X(t)} is integrable in the mean square sense if the
integral ab ab Rxx (ti ,t j ) dti dt j exists. Letting ti = t1 ,t j = t2 , the integral ab ab Rxx (ti ,t j )
R R R R
Result A.3.3:
If {X(t)} is a random process with autocorrelation function Rxx (t1 , t2 ) and if S is
a random variable such that S = ab X(t) dt, then
R
Z b
2 Z bZ b
E X(t)dt = Rxx (t1 ,t2 ) dt1 dt2
a a a
!2 !2 ( )
E ∑ X(ti )∆ti +E ∑ X(t j ) ∆t j − 2E ∑ X(ti ) ∆ti ∑ X(t j ) ∆t j =0
i j i j
!2 !2
⇒ E lim
∆ti →0 i
∑ X(ti ) ∆ti +E lim
∆t j →0 j
∑ X(t j ) ∆t j
( )
− 2E lim lim
∆ti →0 ∆τ j →0 i
∑ ∑ X(ti )X(t j ) ∆ti ∆t j =0
j
276 • Probability and Random Processes for Engineers
By (A4), we have
!2 ( )
2E lim ∑ X(ti )∆ti
∆ti →0 i
− 2E lim lim
∆ti →0 ∆τ j →0 i
∑ ∑ X(ti )X(t j )∆ti ∆t j =0
j
Or
!2 ( )
2E lim ∑ X(t j )∆t j
∆t j →0 j
− 2E lim lim
∆ti →0 ∆τ j →0 i
∑ ∑ X(ti )X(t j )∆ti ∆t j =0
j
Letting ti = t1 ,t j = t2 , we have
Z b
2 Z bZ b
2E X(t)dt −2 Rxx (t1 , t2 )dt1 dt2 = 0
a a a
Z b
2 Z bZ b
⇒ E X(t) dt = Rxx (t1 ,t2 ) dt1 dt2 (A7)
a a a
Z b
S =
a
X(t) dt = lim ∑ X(ti )∆ti
∆ti →0 i
Z b
2 Z bZ b
E(S2 ) = E X(t)dt = Rxx (t1 ,t2 ) dt1 dt2
a a a
Result A.4.1:
If {X(t)} is a random process with autocovariance function C(t1 ,t2 ) then
Z T Z T Z 2T
C(t1 ,t2 )dt1 dt2 = C(τ ) (2T − |τ |) d τ
−T −T −2T
where τ = t1 − t2 orτ = t2 − t1 .
Proof. The contour region for C(t1 , t2 ) is shown in Figure A.1. In order to con-
vert the double integral into a single integral, let us consider the following new
Appendix A • 277
variables defined as
t1 − t2 = τ and t1 + t2 = υ
τ +υ υ −τ
⇒ t1 = and t2 =
2 2
t2
+T
−T 0 +T
t1
−T
t2
+ 2T
υ
+T
− τ + υ = 2T τ + υ = 2T
− 2T −T 0 +T + 2T t1
− τ + υ = − 2T
τ + υ = − 2T
−T
τ
−2T
Figure A.2. The contour region for the transferred function of C(t1 , t2 )
Therefore,
for −2T ≤ τ ≤ 0, we have −2T − τ ≤ υ ≤ 2T + τ and
for 0 ≤ τ ≤ 2T , we have−2T + τ ≤ υ ≤ 2T − τ
Z T Z T Z 0 Z 2T +τ Z 2T Z 2T −τ
1 1
C(t1 ,t2 ) dt1 dt2 = C(τ ) d υ d τ + C(τ ) d υ d τ
−T −T −2T −2T −τ 2 0 −2T +τ 2
Z 0 Z 2T
= C(τ ) (2T + τ ) d τ + C(τ ) (2T − τ ) d τ
−2T 0
Z 2T
= C(τ ) (2T − |τ |) d τ
−2T
The autocovariance function C(τ ) and the function 2T −|τ | for an arbitrary process
are depicted in Figure A.3 for better understanding.
Appendix A • 279
C (τ)
2T − τ
− 2T −a 0 a 2T τ
Z ∞
1
A.5 EVALUATION OF THE INTEGRAL cos τω dω
−∞ (1 + ω 2 )
2
Result A.5.1:
R∞ 1
The integral −∞ cos τω d ω can be evaluated by the contour integration
(1 + ω 2 )
2
R eiaz
technique. Consider C 2
dz, where C is the closed contour consisting of
(1 + z 2 )
the real axis from −R to +R and the upper half of the circle is | z | = R. Refer to
the Figure A.4 given below. The only singular of the integrand lying within C is
the double pole z = i.
i
x
−R +R
We know that
eiτω
Z ∞
(Z )
1 ∞
cos τω d ω = real part of dω
(1 + ω 2 )
2
(1 + ω 2 )
2
−∞ −∞
280 • Probability and Random Processes for Engineers
Now consider
Z ∞
!
eiaz eiaz
2
dz = 2π i i.e. f (z) = 2
at z = i
−∞ (1 + z2 ) (1 + z2 )
d n o
= 2π i lim (z − i)2 f (z)
z→i dz
!
d eiaz
= 2π i lim
z→i dz (i + z)2
( )
(i + z2 )2 eiaz ia − eiaz 2(z + i)
= 2π i lim
z→i (i + z)4
( )
(i + z)iaeiaz − 2eiaz
= 2π i lim
z→i (i + z)3
2 2
( )
2i2 aei a − 2ei a π
= 2π i = (1 + a)e−a
−8i 2
Z ∞
1 π
∴ cos τω d ω = (1 + τ )e−τ
(1 + ω 2 )
2 2
−∞
Result A.6.1:
Let {X(t)} be a stationary random process such that X(t) = A sin (ω t + θ ) where
A is the amplitude, and ω = 2π f is the angular frequency that are assumed con-
stants (A = a and ω = ω0 ) the phase θ is a random variable uniformly distributed
in (0, 2π ). Here the amplitude A, and phase θ are assumed as independent to
each other. However, since amplitude is random, an arbitrary distribution can be
assumed. Now, the power spectral density function of {X(t)} can be obtained as
follows:
Since each realization of the process {X(t)} is a sinusoid at frequency, say ω0 , the
expected power in this process should be located at ω = ω0 and ω = −ω0 . For
given −T ≤ t ≤ T , mathematically this truncated sinusoid implies,
t
XT (t) = a sin (ω0t + θ ) rect
2T
t
where rect is a square impulse of unit height and unit width and centered at
2T
t = 0. Now, the truncated Fourier transform of this truncated sinusoid becomes
XT (ω ) = −iTaei θ sin c[2(ω − ω0 )t] + iTae−i θ sin c[2(ω + ω0 )t]
Appendix A • 281
sin π x
where sinc function is given by sin c (x) = .
πx
n o n o
E |XT (ω |2 = a2 T 2 sin c 2 [2(ω − ω0 )T ] + sin c 2 [2(ω + ω0 )T ]
1 n o 1 2 2 n
Sxx (ω ) = lim E |XT (ω |2 = lim E a T sin c 2 [2(ω − ω0 )T ]
T →∞ 2T T →∞ 2T
o
+ sin c 2 [2(ω + ω 0 )T ]
1 2 n
= lim E a T sin c 2 [2(ω − ω 0 )T ]
T →∞ 2
o
+ sin c 2 [2(ω + ω0 )T ]
Letting v = 2ω T , we have
Z ∞ Z ∞
1 1
g(ω ) d ω = sin c 2 v dv =
−∞ 2 −∞ 2
1
This implies lim T sin c 2 2ω T = δ (ω ) and hence the power spectral density
T →∞ 2
function becomes
E A2 δ (ω − ω 0 ) + δ (ω + ω 0 )
Sxx (ω ) =
2 2
It may be noted that, the average power in a sinusoid process with amplitude A = a
E A2 a2
is = . In fact, this power is evenly split between two frequency points
2 2
ω = ω 0 and ω = −ω 0 .
A PPENDIX B
SERIES AND TRIGONOMETRIC
FORMULAS
2. (1 − x)−1 = 1 + x + x2 + x3 + · · · · · ·
8. sin π = 0, cos π = −1
1 − cos 2A 1 + cos 2A
9. sin2 A = , cos2 A =
2 2
10. sin(a ± b) = sin A cos B ± cos A sin B
1
12. sin A sin B = {cos(A − B) − cos(A + B)}
2
1
13. cos A cos B = {cos(A − B) + cos(A + B)}
2
1
14. sin A cos B = {sin(A + B) + sin(A − B)}
2
Appendix B • 283
1
15. cos A sin B = {sin(A + B) − sin(A − B)}
2
16. sin 2A = 2 sin A cos A
Z π π
sin(A + 2x) 1
25. cos(A+2x)dx = = [sin(A + 2π ) − sin(A − 2π )] = 0
−π 2 −π 2
[Refer (6)]
Z π
eax
26. eax cos bxdx = [a cos bx + b sin bx]
−π a2 + b2
Z π
eax
27. eax sin bxdx = [a sin bx − b cos bx]
−π a2 + b2
Z Z
28. udv = uv − vdu
f ((z)
−∞ 0 z ∞
z −0.09 −0.08 −0.07 −0.06 −0.05 −0.04 −0.03 −0.02 −0.01 −0.00
−3.9 0.0000 0.0000 0.0000 0.0000 0.0000 0.0000 0.0000 0.0000 0.0000 0.0000
−3.8 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001
−3.7 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001
−3.6 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001 0.0001 0.0002 0.0002
−3.5 0.0002 0.0002 0.0002 0.0002 0.0002 0.0002 0.0002 0.0002 0.0002 0.0002
−3.4 0.0002 0.0003 0.0003 0.0003 0.0003 0.0003 0.0003 0.0003 0.0003 0.0003
−3.3 0.0004 0.0004 0.0004 0.0004 0.0004 0.0004 0.0004 0.0005 0.0005 0.0005
−3.2 0.0005 0.0005 0.0005 0.0006 0.0006 0.0006 0.0006 0.0006 0.0007 0.0007
−3.1 0.0007 0.0007 0.0008 0.0008 0.0008 0.0008 0.0009 0.0009 0.0009 0.0010
−3.0 0.0010 0.0010 0.0011 0.0011 0.0011 0.0012 0.0012 0.0013 0.0013 0.0014
−2.9 0.0014 0.0014 0.0015 0.0015 0.0016 0.0016 0.0017 0.0018 0.0018 0.0019
−2.8 0.0019 0.0020 0.0021 0.0021 0.0022 0.0023 0.0023 0.0024 0.0025 0.0026
−2.7 0.0026 0.0027 0.0028 0.0029 0.0030 0.0031 0.0032 0.0033 0.0034 0.0035
−2.6 0.0036 0.0037 0.0038 0.0039 0.0040 0.0041 0.0043 0.0044 0.0045 0.0047
−2.5 0.0048 0.0050 0.0051 0.0052 0.0054 0.0055 0.0057 0.0059 0.0060 0.0062
−2.4 0.0064 0.0066 0.0068 0.0069 0.0071 0.0073 0.0075 0.0078 0.0080 0.0082
−2.3 0.0084 0.0087 0.0089 0.0091 0.0094 0.0096 0.0099 0.0102 0.0104 0.0107
−2.2 0.0110 0.0113 0.0116 0.0119 0.0122 0.0125 0.0129 0.0132 0.0136 0.0139
−2.1 0.0143 0.0146 0.0150 0.0154 0.0158 0.0162 0.0166 0.0170 0.0174 0.0179
−2.0 0.0183 0.0188 0.0192 0.0197 0.0202 0.0207 0.0212 0.0217 0.0222 0.0228
−1.9 0.0233 0.0239 0.0244 0.0250 0.0256 0.0262 0.0268 0.0274 0.0281 0.0287
−1.8 0.0294 0.0301 0.0307 0.0314 0.0322 0.0329 0.0336 0.0344 0.0351 0.0359
−1.7 0.0367 0.0375 0.0384 0.0392 0.0401 0.0409 0.0418 0.0427 0.0436 0.0446
−1.6 0.0455 0.0465 0.0475 0.0485 0.0495 0.0505 0.0516 0.0526 0.0537 0.0548
−1.5 0.0559 0.0571 0.0582 0.0594 0.0606 0.0618 0.0630 0.0643 0.0655 0.0668
−1.4 0.0681 0.0694 0.0708 0.0721 0.0735 0.0749 0.0764 0.0778 0.0793 0.0808
−1.3 0.0823 0.0838 0.0853 0.0869 0.0885 0.0901 0.0918 0.0934 0.0951 0.0968
−1.2 0.0985 0.1003 0.1020 0.1038 0.1057 0.1075 0.1093 0.1112 0.1131 0.1151
−1.1 0.1170 0.1190 0.1210 0.1230 0.1251 0.1271 0.1292 0.1314 0.1335 0.1357
−1.0 0.1379 0.1401 0.1423 0.1446 0.1469 0.1492 0.1515 0.1539 0.1562 0.1587
−0.9 0.1611 0.1635 0.1660 0.1685 0.1711 0.1736 0.1762 0.1788 0.1814 0.1841
−0.8 0.1867 0.1894 0.1922 0.1949 0.1977 0.2005 0.2033 0.2061 0.2090 0.2119
−0.7 0.2148 0.2177 0.2207 0.2236 0.2266 0.2297 0.2327 0.2358 0.2389 0.2420
−0.6 0.2451 0.2483 0.2514 0.2546 0.2578 0.2611 0.2643 0.2676 0.2709 0.2743
−0.5 0.2776 0.2810 0.2843 0.2877 0.2912 0.2946 0.2981 0.3015 0.3050 0.3085
−0.4 0.3121 0.3156 0.3192 0.3228 0.3264 0.3300 0.3336 0.3372 0.3409 0.3446
−0.3 0.3483 0.3520 0.3557 0.3594 0.3632 0.3669 0.3707 0.3745 0.3783 0.3821
−0.2 0.3859 0.3897 0.3936 0.3974 0.4013 0.4052 0.4090 0.4129 0.4168 0.4207
−0.1 0.4247 0.4286 0.4325 0.4364 0.4404 0.4443 0.4483 0.4522 0.4562 0.4602
0.0 0.4641 0.4681 0.4721 0.4761 0.4801 0.4840 0.4880 0.4920 0.4960 0.5000
286 • Probability and Random Processes for Engineers
z 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09
0.0 0.5000 0.5040 0.5080 0.5120 0.5160 0.5199 0.5329 0.5279 0.5319 0.5359
0.1 0.5398 0.5438 0.5478 0.5517 0.5558 0.5596 0.5636 0.5675 0.5714 0.5753
0.2 0.5793 0.5832 0.5871 0.5910 0.5948 0.5987 0.6026 0.6064 0.6103 0.6141
0.3 0.6179 0.6217 0.6255 0.6293 0.6331 0.6368 0.6406 0.6443 0.6480 0.6517
0.4 0.6554 0.6591 0.6628 0.6664 0.6700 0.6736 0.6772 0.6808 0.6844 0.6879
0.5 0.6915 0.6950 0.6985 0.7019 0.7054 0.7088 0.7123 0.7157 0.7190 0.7224
0.6 0.7257 0.7291 0.7324 0.7357 0.7389 0.7422 0.7454 0.7486 0.7517 0.7549
0.7 0.7580 0.7611 0.7642 0.7673 0.7704 0.7734 0.7764 0.7794 0.7823 0.7852
0.8 0.7881 0.7910 0.7939 0.7967 0.7995 0.8023 0.8051 0.8079 0.8106 0.8133
0.9 0.8159 0.8186 0.8212 0.8238 0.8264 0.8289 0.8315 0.8340 0.8365 0.8389
1.0 0.8413 0.8438 0.8461 0.8485 0.8508 0.8531 0.8554 0.8577 0.8599 0.8621
1.1 0.8643 0.8665 0.8686 0.8708 0.8729 0.8749 0.8770 0.8790 0.8810 0.8830
1.2 0.8849 0.8869 0.8888 0.8907 0.8925 0.8944 0.8962 0.8980 0.8997 0.9015
1.3 0.9032 0.9049 0.9066 0.9082 0.9099 0.9115 0.9131 0.9147 0.9162 0.9177
1.4 0.9192 0.9207 0.9222 0.9236 0.9251 0.9265 0.9279 0.9292 0.9306 0.9319
1.5 0.9332 0.9345 0.9357 0.9370 0.9382 0.9394 0.9406 0.9418 0.9429 0.9441
1.6 0.9452 0.9463 0.9474 0.9484 0.9495 0.9505 0.9515 0.9525 0.9535 0.9545
1.7 0.9554 0.9564 0.9573 0.9582 0.9591 0.9599 0.9608 0.9616 0.9625 0.9633
1.8 0.9641 0.9649 0.9656 0.9664 0.9671 0.9678 0.9686 0.9693 0.9699 0.9706
1.9 0.9713 0.9719 0.9726 0.9732 0.9738 0.9744 0.9750 0.9756 0.9761 0.9767
2.0 0.9773 0.9778 0.9783 0.9788 0.9793 0.9798 0.9803 0.9808 0.9812 0.9817
2.1 0.9821 0.9826 0.9830 0.9834 0.9838 0.9842 0.9846 0.9850 0.9854 0.9857
2.2 0.9861 0.9864 0.9868 0.9871 0.9875 0.9878 0.9881 0.9884 0.9887 0.9890
2.3 0.9893 0.9896 0.9898 0.9901 0.9904 0.9906 0.9909 0.9911 0.9913 0.9916
2.4 0.9918 0.9920 0.9922 0.9925 0.9927 0.9929 0.9931 0.9932 0.9934 0.9936
2.5 0.9938 0.9940 0.9941 0.9943 0.9945 0.9946 0.9948 0.9949 0.9950 0.9952
2.6 0.9953 0.9955 0.9956 0.9957 0.9959 0.9960 0.9961 0.9962 0.9963 0.9964
2.7 0.9965 0.9966 0.9967 0.9968 0.9969 0.9970 0.9971 0.9972 0.9973 0.9974
2.8 0.9974 0.9975 0.9976 0.9977 0.9977 0.9978 0.9979 0.9979 0.9980 0.9981
2.9 0.9981 0.9982 0.9983 0.9983 0.9984 0.9984 0.9985 0.9985 0.9986 0.9986
3.0 0.9987 0.9987 0.9987 0.9988 0.9988 0.9989 0.9989 0.9989 0.9990 0.9990
3.1 0.9990 0.9991 0.9991 0.9991 0.9992 0.9992 0.9992 0.9992 0.9993 0.9993
3.2 0.9993 0.9993 0.9994 0.9994 0.9994 0.9994 0.9994 0.9995 0.9995 0.9995
3.3 0.9995 0.9995 0.9996 0.9996 0.9996 0.9996 0.9996 0.9996 0.9996 0.9997
3.4 0.9997 0.9997 0.9997 0.9997 0.9997 0.9997 0.9997 0.9997 0.9997 0.9998
3.5 0.9998 0.9998 0.9998 0.9998 0.9998 0.9998 0.9998 0.9998 0.9998 0.9998
3.6 0.9998 0.9998 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999
3.7 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999
3.8 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 0.9999 1.0000
3.9 1.0000 1.0000 1.0000 1.0000 1.0000 1.0000 1.0000 1.0000 1.0000 1.0000
A NSWERS
ANSWERS TO EXERCISE PROBLEMS
Chapter 1
1. 0.72
2. 0.929258
1 1
4. (a) , (b) ,
72 18
X =x 1 2 3 4 5 6 7 8
(c) 1 4 9 16 27 40 55 72
F(x) =1
72 72 72 72 72 72 72 72
8
5.
9
5
1 1 − e−4
6. (i) k = 2
, (ii) 2
1 − e−a 1 − e−a
7. 1.5 inch3
θ θ2
8. Mean = , Variance=
2 12
9. Mean = 1, Variance= 1
10. 0.04
11. 0.9394
288 • Probability and Random Processes for Engineers
x (1 + 3y2 )
12. (i) f (x) = , 0 < x < 2, f (y) = , 0 < y < 1, f (x/y) = f (x)
2 2
(X and Y are independent)
5 4 5 5 3
(ii) , (iii) E(X)E(Y ) = = = E(XY ), (iv)
8 3 8 6 64
y−b 2
√1 − 12
, −∞ < y < ∞
a
e
13. h(y) =
2π a
0, otherwise
r2
1 , 0 < θ < 2π
r −
14. f (r) = e 2σ2 , 0 < r < ∞ , f (θ ) = 2π
σ2 ∴
0, otherwise
0, otherwise
1 (−1/3.75) x12 −0.50x1 x2 +x22
15. f (x1 , x2 ) = √ e , −∞ < x1 , x2 < ∞
4π 0.9375
Chapter 2
7. E {X(t)} = 0
1
V {X(t)} = cos2 (ω t + θ )
3
1−e−2(t1+t2 )
8. C(t1 , t2 )=E{X(t1 )X(t2 )}−E {X(t1 )} E{X(t2 )}= cos ω (t1 −t2 )
t1 +t2
1
9. E {X(t)} = 0,V {X(t)} =
2
−5 3
10. E {X(t)} = (cost + sint) ,V {X(t)} =
4 16
Chapter 3
1 − e−2(t1 +t2 )
1. E {X(t)} = 0, R(t1 ,t2 ) = cos ω (t1 −t2 ), {X(t)} is not station-
t1 + t2
ary.
1 1
2. fX(t0 ) (x) = F ′ (x) = √ , − sin ω0t0 ≤ x ≤ + sin ω0t0 , {X(t)}is
2π 1 − x2
SSS process.
1
3. E {X(t)} = 0, Rxx (t1 , t2 ) = e−|τ |/2 , {X(t)} is WSS process.
2
1
4. (i) E {X(t)} = 0,V {X(t)} = ,
2
(ii) {X(t)} is SSS process,
A2
(iii) R (t1 ,t2 ) = cos ωτ ,
2
(iv) {X(t)} is WSS process,
(v) Plot for X(t) = (+1) cos (2t + π ) and X (1) = (−1) cos (2t + π ) taking t =
(0, 10)
Rxx (τ )
5. E {Y (t)} = 0, Ryy (t1 ,t2 ) = cos ωτ , {Y (t)} is WSS process.
2
6. Plot for X(t) = cos(2t + π ) taking t = (0, 10) and for R (τ ) = (0.5) cos 2τ
taking τ = (−10, 10)
8. Rww (t, t + τ ) = Rxx (τ )Ryy (τ ) , {W (t)} is WSS process and Rxw (t, t + τ ) =
Rxx (τ )µy , {X(t)} and {W (t)} are JWSS processes.
290 • Probability and Random Processes for Engineers
1 1
10. E {X(t)} = 0, E X 2 (t) = ,V {X(t)} =
2 2
Chapter 4
1. (i) and (iii) are valid autocorrelation functions, (ii) and (iv) are not valid
autocorrelation functions
2. 2A 1 − e−3α
6. 0.43235
A4
1
7. Ryy (τ ) = 1 + cos 2ωτ
4 2
1
8. (i) E {X(t)} = cos ω t,
2
1
(ii) Rxx (t1 ,t2 ) = cos ω t1 cos ω t2 ,
3
1
(iii) Cxx (t1 ,t2 ) = cos ω t1 cos ω t2
12
1
9. E {Y (t)} = 0, Ryy (τ ) = 2 {2Rxx (τ ) − Rxx (0) − Rxx (2τ )}, {Y (t)} is WSS
process
τ
10. Ryy (τ ) = Rxx (τ )+ Rxn (τ )+Rnx (τ )+Rnn (τ ), Ryx (τ ) = Rxx (τ )+ Rnx (τ ), Ryn (τ )
= Rxn (τ ) + Rnn (τ )
Answers to Exercise Problems • 291
Chapter 5
a + 2b a2
1. E {Yn } = , V {Yn } =
2 4
e−5t (5t)n
2. (i) P {Y (t) = n} = , n = 0, 1, 2, · · · · · · ,
n!
(ii) P {Y (2) = 5} = 0.0378,
(iii) E {Y (2)} = V {Y (2)} = 10,
(iv) λ1t = 4 , λ2t = 6 , (λ1 + λ2 )t = 10
t0
5. P {T1 ≤ t0 /X(t) ≤ 1} =
t
2
6. P {T1 ≤ 4/X(10) ≤ 1} =
5
7. P {Y (2) = 3} = 0.00038
21
9. (i) E {X (7)} = V {X (7)} = ,
4
(ii) P {X (3) ≤ 3/X (1) ≤ 1} = 0.3466
Chapter 6
r
1 1 1
7. E {Z(t)} = ,V {Z(t)} = 1−
2π 2 π
m 1 + (n − 1)(2p − 1)2 ,
if m<n
9. Rxx (m, n) =
n 1 + (m − 1)(2p − 1)2 , if n<m
1
10. E(X n ) = 0,V (X n ) = (2n − 1)
n2
Chapter 7
1 1 1
4. Rxx (τ ) = cos τ , E(ZT ) = cos τ , lim Z T = cos τ = Rxx (τ ), {X(t)}
2 2 T →∞ 2
is correlation ergodic.
Answers to Exercise Problems • 293
Z2T
τ 1 − e−2α T
1 q
5. lim c(τ ) 1 − d τ = lim 1− = 0, {X(t)} is
T →∞ T 2T T →∞ α T 2α T
0
mean ergodic.
Z2T
1 τ q
6. lim c(τ ) 1 − d τ = lim = 0, {X(t)} is mean ergodic.
T →∞ 2T 2T T →∞ 2T
−2T
Z2T
1 τ
7. lim c(τ ) 1 − d τ = lim σA2 = σA2 6= 0, {X(t)} is not mean ergodic.
T →∞ T 2T T →∞
0
1 − e4λ T
!
1
8. E X T = 0, V X T = 1− , lim V (X T ) = 0, {X(t)}
4λ T 2λ T T →∞
is mean ergodic.
10 1 − e−4T
9. lim V (X T ) = lim 1− = 0, {X(t)} is mean ergodic.
T →∞ T →∞ T 2T
1 Z2T τ
φ (τ ) 1 − d τ − {R(τ )}2 ,
10. lim V Z T = lim
T →∞ T →∞ 2T 2T
0
Z2T
1 τ
φ (τ ) 1− d τ → {R(τ )}2
∴ lim V Z T → 0 if and only if lim
T →∞ T →∞ 2T 2T
0
T →∞
where φ (τ ) = E {X(t1 + τ )X(t1 )X(t2 + τ )X(t2 ) }
Chapter 8
12.5
1. Rxx (0) = Sq. units
π
√ 2
2. Sxx (ω ) = π e−ω /4
k
3. Rxxx (τ ) = sin aτ
πτ
1 − cos ω 2 sin2 ω /2 sin ω /2 2
4. Sxx (ω ) = 2 = =
ω2 ω 2 /2 ω /2
2
5. Rxx (τ ) = τ 2 sin τ + τ cos τ − sin τ
πτ 3
294 • Probability and Random Processes for Engineers
4a2 b
6. Sxx (ω ) =
4b2 + ω 2
2
7. Sxx (ω ) =
1 + (ω /10)2
1
8. Rxy (τ ) = 2 [aτ − b) sin τ + bτ cos τ ]
πτ
1 1
9. Sxx (ω ) = Aa 2 +
a + (ω + b)2 a2 + (ω − b)2
1
10. Rxx (τ ) = 3 τ 2 sin τ + 2τ cos τ − 2 sin τ
πτ
Chapter 9
1. (i) P {X2 = 1} = 0.1962,
(ii) P (X0 = 1, X1 = 2, X2 = 2) = 0.0625,
(2)
(iii) P (X2 = 2, X1 = 2/X0 = 1) = 0.25, (iv) P12 = 0.4375
2. ∑∞
(n) (1) (2) (3)
n=1 f 11 = f 11 + f 11 + f 11 + · · · · · · = 1 + 0 + 0 + · · · · · · = 1 (state 1 is
recurrent)
∑∞
(n) (1) (2) (3)
n=1 f 22 = f 22 + f 22 + f 22 + · · · · · · = 0.5 + 0 + 0 + · · · · · · = 0.5 < 1 (state
2 is transient)
11 4 21
36 36 36
4 1 7
3. (ii) P(2) =
12 12 12
3 1 5
9 9 9
2 2 1
6. π1 = , π2 = , π3 =
5 5 5
7. States are irreducible, Periodic with period 2, Non-null persistent, Not ergodic.
1
8. (b) π1 = π2 = π3 = , (c) States are irreducible, Periodic with period 3,
3
Non-null persistent, Not ergodic.
Answers to Exercise Problems • 295
0 1/2 1/2
9. (ii) P(1) =
3/4 0 1/4
,
1/8 7/8 0
1 2 4
(iii) π1 = , π2 = , π3 =
3 5 15
0 1 2 3
0 2/3 1/3 0 0
1 2/3 0 1/3 0 ,
10. (i) P(1) =
2 0 2/3 0 1/3
3 0 0 2/3 1/3
(2) 5
(ii) π1 = ,
18
4
(iii) π1 =
15
Index
J R
joint probability density fonction, 21 random fonction, 52
joint probability mass fonction, 21 random process, 52
jointly strict sense stationary, 85 interpretation, 61
jointly wide sense stationary, 85 classification, 62
random sequence, 62, 66
L
continuity related to, 269
limiting case, 13
derivatives related to, 271
limiting distribution, 245
integrals related to, 273
limiting probability, 245
Bernoulli process, 139
M binomial process, 128
Markov chain, 235 normal process, 153
Markov process, 234 standard normal process, 156
Markovian property, 235 random variable, 3
mean ergodic theorem, 188 continuous, 6
mean recurrent time, 247 discrete, 4
moments, 10, 83 one dimensional, 3
moment generating fonction, 10 two dimensional, 21
central moments, l 0 Bernoulli, 12
raw moments, 10 binomial, 11
multivariate normal distribution, 24 normal, 16
N Poisson, 13
narrow band filter, 212 uniform, 15
normal distribution, 16 random walk, 129
normal process (also see Gaussian process), simple random walk, 165
153 random walk process, 153
0 S
semi random process, 145
One-dimensional random variable, 3
sine fonction, 69, 281
P sinusoidal wave, 106
point process, 130 square-law detector process, 158
Poisson points, 130 standard deviation, 10
Poisson process, 130, 128, 132 statistical averages, 63, 65
power spectrum, 206 states, 233
power spectral density, 206 absorbing states, 246
cross power spectral density, 208 communicating states, 246
properties, 209 persistent states, 247
Probability, 1 recurrent states, 247
axioms, 2 transient state 247
probability mass fonction, 5, 63 ergodic state, 248
probability density fonction, 6, 63 stationarity, 82
Index 299
steady state probabilities, 245 U
stochastic inputs, 215 unbiased estimator, 187
strict sense stationary process, 82 uniform distribution, 15
uniform random variable, 15
T
telegraphic signal, 145 V
time average, 182 variance, 9
transformation of random variables, 23 variance covariance matrix, 25, 154
discrete case, 23 W
continuous case, 23 Wiener-Khinchin theorem, 213
transformation of integrals, 276 Wiener process, 167, 153
transition probabilities, 235 white noise process, 67, 114
transition probability matrix, 236 white Gaussian noise, 153
wide sense stationary process, 83
Probability & Random Probability & TM
978-93-89976-41-0
Distributed by:
9 789389 976410
TM