Estimation of Finite Mixture Distributions Through Bayesian Sampling

Abstract
A formal Bayesian analysis of a mixture model usually leads to intractable calculations, since the posterior distribution takes into account all the partitions of the sample. We present approximation methods which evaluate the posterior distribution and Bayes estimators by Gibbs sampling, relying on the missing data structure of the mixture model. The data augmentation method is shown to converge geometrically, since a duality principle transfers properties from the discrete missing data chain to the parameters. The fully conditional Gibbs alternative is shown to be ergodic and geometric convergence is established in the normal case. We also consider non‐informative approximations associated with improper priors, assuming that the sample corresponds exactly to a k‐component mixture.

This publication has 11 references indexed in Scilit: