In Bayesian inference, the probability density function of a vector
is typically estimated from a training set of such vectors
as shown in Equation 5.1 [5].
By integrating over ,
we are essentially integrating over all
the pdf models possible. This involves varying the families of pdfs
and all their parameters. However, often, this is impossible and
instead a family is selected and only its parametrization
is
varied. Each
is a parametrization of the pdf of
and
is weighted by its likelihood given the training set. However,
computing the integral 5.1 is not always straightforward and Bayesian inference is
approximated via maximum a posteriori (MAP) or maximum likelihood (ML)
estimation as in Equation 5.2. The EM algorithm is
frequently utilized to perform these maximizations.