Pattern Recognition and Machine Learning

(Jeff_L) #1
456 9. MIXTURE MODELS AND EM

9.3 ( ) www Consider a Gaussian mixture model in which the marginal distribution
p(z)for the latent variable is given by (9.10), and the conditional distributionp(x|z)
for the observed variable is given by (9.11). Show that the marginal distribution
p(x), obtained by summingp(z)p(x|z)over all possible values ofz, is a Gaussian
mixture of the form (9.7).

9.4 ( ) Suppose we wish to use the EM algorithm to maximize the posterior distri-
bution over parametersp(θ|X)for a model containing latent variables, whereXis
the observed data set. Show that the E step remains the same as in the maximum
likelihood case, whereas in the M step the quantity to be maximized is given by
Q(θ,θold)+lnp(θ)whereQ(θ,θold)is defined by (9.30).

9.5 ( ) Consider the directed graph for a Gaussian mixture model shown in Figure 9.6.
By making use of the d-separation criterion discussed in Section 8.2, show that the
posterior distribution of the latent variables factorizes with respect to the different
data points so that

p(Z|X,μ,Σ,π)=

∏N

n=1

p(zn|xn,μ,Σ,π). (9.80)

9.6 ( ) Consider a special case of a Gaussian mixture model in which the covari-
ance matricesΣkof the components are all constrained to have a common value
Σ. Derive the EM equations for maximizing the likelihood function under such a
model.

9.7 ( ) www Verify that maximization of the complete-data log likelihood (9.36) for
a Gaussian mixture model leads to the result that the means and covariances of each
component are fitted independently to the corresponding group of data points, and
the mixing coefficients are given by the fractions of points in each group.

9.8 ( ) www Show that if we maximize (9.40) with respect toμkwhile keeping the
responsibilitiesγ(znk)fixed, we obtain the closed form solution given by (9.17).

9.9 ( ) Show that if we maximize (9.40) with respect toΣkandπkwhile keeping the
responsibilitiesγ(znk)fixed, we obtain the closed form solutions given by (9.19)
and (9.22).

9.10 ( ) Consider a density model given by a mixture distribution

p(x)=

∑K

k=1

πkp(x|k) (9.81)

and suppose that we partition the vectorxinto two parts so thatx =(xa,xb).
Show that the conditional densityp(xb|xa)is itself a mixture distribution and find
expressions for the mixing coefficients and for the component densities.
Free download pdf