Assume we choose the same prior distribution for a GMM as in Example 14.3.1. Use the EM
Question:
Assume we choose the same prior distribution for a GMM as in Example 14.3.1. Use the EM algorithm to derive the MAP estimation for GMMs:
a. Give the formulae to update all GMM parameters MAP iteratively.
b. If we approximate the true posterior distribution of a GMM p¹jxº by another approximate distribution q˜¹º as q˜¹º / p¹ºQ¹jMAPº, where Q¹º is the auxiliary function in the EM algorithm, derive this approximate posterior distribution q˜¹º, and compare it with the variational distribution q¹º in Example 14.3.1.
Fantastic news! We've Found the answer you've been seeking!
Step by Step Answer:
Related Book For
Machine Learning Fundamentals A Concise Introduction
ISBN: 9781108940023
1st Edition
Authors: Hui Jiang
Question Posted: