GMM
GMM
IIT Indore
https://chandreshiit.github.io
1
Slides credit goes to Yi, Yung
November 17, 2024 1 / 26
Warm-Up
https://www.youtube.com/watch?v=q71Niz856KE
K
X K
X
p(x) = πk pk (x), 0 ≤ πk ≤ 1, πk = 1
k=1 k=1
K
X K
X
p(x|θ) = N (x|µk , Σk ), 0 ≤ πk ≤ 1, πk = 1,
k=1 k=1
where the parameters θ := {µk , Σk , πk : k = 1, . . . , K }
• Example. p(x|θ) = 0.5N (x| − 2, 1/2) + 0.2N (x|1, 2) + 0.3N (x|4, 1)
• How much is each component k responsible, if the data xn is sampled from the
current mixture model?
PK
• rn = (rnk : k = 1, . . . , K ) is a probability distribution, so k=1 rnk = 1
• Soft assignment of xn to the K mixture components
and go to S2.
- The update equation in M-step is still mysterious, which will be covered later.
PN
rnk xn
µnew
k = Pn=1
N
, k = 1, . . . , K
n=1 rnk
N
1 X
Σnew
k = rnk (xn − µk )(xn − µk )T , k = 1, . . . , K
Nk
n=1
PN
n=1 rnk
πknew = , k = 1, . . . , K
N
2
In probabilistic PCA, z was continuous, so we integrated them out.
L11(3) November 17, 2024 20 / 26
Joint Distribution, Likelihood, and Posterior (2)
• Now, for all data samples X , each data xn has zn = [zn1 , . . . , znK ]T , but with the
same prior π.
p(znk = 1)p(xn |znk = 1) πk N (xn |µk , Σk )
p(znk = 1|xn ) = PK = PK = rnk
j=1 p(znj = 1)p(xn |znj = 1) j=1 πj N (xn |µj , Σj )
1)