Gaussian Random Vectors
Gaussian Random Vectors
The standard normal random vector. Recall that Z is a standard normal (or
Gaussian) random variable if its pdf is given by
1 z2
f (x) = √ e− 2 , −∞ < z < ∞.
2π
For briefness, we denoted this distribution N (0, 1). All other normal random vari-
ables can be obtained as linear transformations of Z. Thus, if X = µ + σZ, then
X ∼ N (µ, σ 2 ). We will use the same idea to introduce multivariate normal distri-
butions.
Let Z1 , ..., Zn be independent standard normal random variables. We define stan-
dard normal random vector as
Z1
Z2
Z=
... .
Zn
where 0 = 0n×1 is the column vector of 0’s, and I = In×n is the identity matrix.
The distribution of the standard normal random vector Z is denoted N (0, I). Due
to independence of its components, the joint pdf of the standard normal vector Z
is given by
Yn
1 zi2 1 1 ∑n 1 1 ′ 1
√ e− 2 = e− 2 i=1 zi = e− 2 z z = e− 2 ∥z∥ .
2 1 2
fZ (z) = n/2 n/2 n/2
i=1
2π (2π) (2π) (2π)
Note that here we treated vector variable z as a column vector. Then z′ z coincides
with the dot product z · z and defines the squared Euclidean norm ∥z∥2 .
Defining normal random vectors via linear transformations. Now let us
consider a linear transformation
1
where µ is an arbitrary n × 1 column-vector and A is a square n × n matrix (with
real entries). In a more detailed form,
X
n
Y i = µi + Aij Zj .
j=1
We say that random vector Y has an n-dimensional normal distribution, with the
mean vector µ and variance-covariance matrix V, and denote this as
Y ∼ N (µ, V).
Multivariate normal pdf. We will show next that the random vector Y has a
pdf if and only if the matrix A is invertible, and derive its pdf in the latter case.
So, let first the matrix A be invertible. Hence, V = AA′ is also invertible. In this
case, one can invert the relation
y = µ + Az
as follows
X
n
−1
z = A (y − µ), or zi = A−1 ij
(yj − µj ).
j=1
2
1 −1 )′ A−1 (y−µ) 1 ′ −1 A−1 (y−µ)
e− 2 (y−µ)(A e− 2 (y−µ)(A )
1 1
= =
(2π)n/2 |V|1/2 (2π)n/2 |V|1/2
1 ′ −1 (y−µ) 1 −1 (y−µ)
e− 2 (y−µ)(AA ) e− 2 (y−µ)V
1 1
= =
(2π)n/2 |V|1/2 (2π)n/2 |V|1/2
1 −1
e− 2 (y−µ)V (y−µ) .
1
(1)
|2πV|1/2
α1 y1 + α2 y2 ∈ L.
which is impossible.
Remark.
RR To make a better sense of (2), think of a two dimensional integral
G
f (x, y) dxdy, where the region G is a straight line. Such integral necessarily
vanishes, since the volume of any straight line equals 0.
From the all-important formula (1), one can derive a number of corollaries.
Corollary 1. Let Y1 , ..., Yn be jointly Gaussian, with σi2 = VarYi > 0. Then Yi
are independent if and only if they are uncorrelated. Indeed, independent r.v. are
always uncorrelated. Now, if Yi are uncorrelated and jointly Gaussian, then
σ12 0 ··· 0 σ1−2 0 ··· 0
0 σ22 ··· 0 0 σ2−2 ··· 0
V = CovY =
... .. .. , so that V−1 =
... .. .. .
. .
.. ..
. . . .
0 0 · · · σn2 0 0 · · · σn−2
3
Since
X
n
X
n
X
n
(yi − µi )2
−1 −1 −1
(y−µ)V (y−µ) = (yi −µi ) V (yj −µj ) = V (yi −µi ) =
2
,
i,j=1
ij
i=1
ii
i=1
σi2
and
Y
n
|V| = σi2 ,
i=1
one gets
1 − 12
∑n (yi −µi )2 Y
n
1 (y −µ )
− i 2i Y2 n
fY (y) = Q e
i=1 σ2
i = e 2σ
i = fYi (yi ),
(2π)n/2 ( ni=1 σi2 )1/2 i=1
(2πσi2 )1/2 i=1
which shows that Yi are independent random variables Yi ∼ N (µi , σi2 ), by one of
our equivalent definitions of independence.
One can farther generalize this important property. Suppose a Gaussian random
n-vector X can be partitioned as follows
X1 µ1
Xn×1 = ∼ N ,V ,
X2 µ2
where
and
X
n
′ −1
(x − µ) V (x − µ) = (xi − µi )(V−1 )ij (xj − µj ) =
i,j=1
X
p
X
n
−1
(xi − µi )(V )ij (xj − µj ) + (xi − µi )(V−1 )ij (xj − µj ) =
i,j=1 i,j=p+1
4
Thus,
1 − 12 (x1 −µ1 )′ V1−1 (x1 −µ1 )− 12 (x2 −µ2 )′ V2−1 (x2 −µ2 )
fX (x) = e =
(2π)(p+q)/2 (|V1 | · |V2 |)1/2
1 ′ −1 1 ′ −1
e− 2 (x1 −µ1 ) V1 (x1 −µ1 )
e− 2 (x2 −µ2 ) V2 (x2 −µ2 )
1 1
· = fX1 (x1 )fX2 (x2 ).
(2π)p/2 (|V 1 |)1/2 (2π)q/2 (|V 2 |)1/2
Here the equality to 0 on the right-hand side is excluded, because it would imply
that
X
n
u · (X − EX) = ui (Xi − EXi ) = 0
i=1
Y = µ + AZ,
5
Corollary 2. a) Any normal vector Y ∼ N (µ, V) having a pdf is a linear function
of a standard normal random vector Z.
b) Any linear function U = Cn×1 + Bn×n Yn×1 with an invertible matrix B is also
normal (since, ultimately, it is also a linear function of a standard normal random
vector Z).
Orthogonal transformations. A square matrix An×n is called orthogonal, if
its row-vectors Ai· are orthonormal, i.e. mutually orthogonal and of length one:
1 , i=j
Ai· · Aj· =
0 , i ̸= j.
AA′ = I, or A−1 = A′ ,
A′ A = I.
which shows that the column-vectors of an orthogonal matrix are also orthonor-
mal.
Orthogonal matrices have some special properties. First, such matrices preserve
lengths of vectors: if we consider the linear transformation