0% found this document useful (0 votes)
81 views2 pages

Lec 33

This document summarizes the Gaussian mixture model (GMM) using the expectation-maximization (EM) algorithm. It discusses: 1. A 2-component GMM where all 5 parameters are unknown and estimated using EM. 2. An M-component GMM derived similarly using EM, with the maximization step estimating the parameters. 3. Segmenting an MRI brain image into 2 components using a 2-component GMM, and into 3 components using a 3-component GMM, showing better results.

Uploaded by

todkarvarsha
Copyright
© Attribution Non-Commercial (BY-NC)
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
81 views2 pages

Lec 33

This document summarizes the Gaussian mixture model (GMM) using the expectation-maximization (EM) algorithm. It discusses: 1. A 2-component GMM where all 5 parameters are unknown and estimated using EM. 2. An M-component GMM derived similarly using EM, with the maximization step estimating the parameters. 3. Segmenting an MRI brain image into 2 components using a 2-component GMM, and into 3 components using a 3-component GMM, showing better results.

Uploaded by

todkarvarsha
Copyright
© Attribution Non-Commercial (BY-NC)
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd

Gaussian Mixture Model by Using EM

Shubing Wang∗
April 17, 2004

1. 2-component Gaussian Mixture Model. We consider the


The MRI image of bone slice The histgram of the reduced intensity list
500
250

450

case that all the 5 parameters 50


200 400

T 350

θ = (p, µ1 , σ1 , µ2 , σ2 ) 100
150 300

are unknown. By using EM algorithm:


250

150
100 200

• Estimation Step: define p˜i = P (yi ∼ f1 |Θ), then


150

200
50 100

pf1 (yi |Θ) 50

p̃ =
250

. 50 100 150 200 250


0 0
50 100 150 200 250

pf1 (yi |Θ) + (1 − p)f2 (yi |Θ)


So Figure 1: Left: mid sagittal brain image. Right: histogram of
0
X
0
√ image intensity.
Q(Θ, Θ ) = {[p˜i (log p − log 2π
i 500

(Yi − µ01 )2
− log σ10 − )] 450
The Gaussian mix−
2σ12 ture model
√ 400
+[(1 − p˜i )(log(1 − p0 ) − log 2π
(Yi − µ02 )2 350
− log σ20 − )]}
2σ22 300

• Maximization Step: 250


P
i p˜i
p0 = 200
n
150
P
p˜ Y
0
µ1 = Pi i i
i p˜i 100
sP
(Yi − µ01 )2
i p˜iP
50
σ10 =
i p˜i 0
P 50 100 150 200 250
(1 − p˜i )Yi
µ02 = Pi
i (1 − p˜i ) Figure 2: The comparation of the original data and the fitted
sP Gaussian Mixture Model
i (1P− p˜i )(Yi − µ02 )2
σ20 =
i (1 − p˜i )
3. Segmentation by Using Gaussian Mixture Model. We
2. M -component Gaussian Mixture Model. It can derived start with data as shown in Figure 1. We always should
similarly by using EM with the Maximization step: get the histgram as well, since the initial guess is cru-
cial when one is trying to do EM algorithm. A good ini-
P
0 i p˜ ji
pj = , 1≤j≤M tial guess will absolutely improve the preformance of the
n
programming and histgrams provide the clue for goog
P
p˜ Y
µ0j = Pi ji i , if 1 ≤ j < M initial guess. One can get the 2-component Gaussian
i p˜ ji
sP Mixture Model as shown in Figure 2. The result of seg-
0 2
i p˜ji (Yi − µj ) mentation as shown in Figure 3.
σj0 = P , if 1 ≤ j < M
i p˜ji Intuitively, 3-component Gaussian Mixture Model will
P PM −1 give a better result. Figure 4 and Figure 5 show the re-
i (1 − j=1 p˜ji )Yi
0
µM = P PM −1 sult of segmentation by using the 3-component Gaussian
i (1 − j=1 p˜ji ) Mixture Model.
v
uP PM −1
u i (1 − j=1 p˜ji )(Yi − µ0M )2
0
σM = t P PM −1
i (1 − j=1 p˜ji )
∗ Email: wang6@[Link]
1 1

0.9 0.9

50 0.8 50 0.8

0.7 0.7

100 0.6 100 0.6

0.5 0.5

150 150
0.4 0.4

0.3 0.3

200 200
0.2 0.2

0.1 0.1

250 250
0 0
50 100 150 200 250 50 100 150 200 250

Figure 3: The two components of the 2-component Gaussian


Mixture Model

The Gaussian Mixture Model and the Original Data


500

450

400

350

300

250

200

150

100

50

0
50 100 150 200 250

Figure 4: The fitting 3-component Gaussian Mixture Model

1 1

0.9 0.9

50 0.8 50 0.8

0.7 0.7

100 0.6 100 0.6

0.5 0.5

150 150
0.4 0.4

0.3 0.3

200 200
0.2 0.2

0.1 0.1

250 250
0 0
50 100 150 200 250 50 100 150 200 250
1

0.9

50 0.8

0.7

100 0.6

0.5

150
0.4

0.3

200
0.2

0.1

250
0
50 100 150 200 250

Figure 5: The 3 segments of the 3-component Gaussian Mix-


ture Model

You might also like