Estimation of finite mixture distributions through Bayesian sampling. (English) Zbl 0796.62028
Summary: A formal Bayesian analysis of a mixture model usually leads to intractable calculations, since the posterior distribution takes into account all the partitions of the sample. We present approximation methods which evaluate the posterior distribution and Bayes estimators by Gibbs sampling, relying on the missing data structure of the mixture model. The data augmentation method is shown to converge geometrically, since a duality principle transfers properties from the discrete missing data chain to the parameters. The fully conditional Gibbs alternative is shown to be ergodic and geometric convergence is established in the normal case. We also consider non-informative approximations associated with improper priors, assuming that the sample corresponds exactly to a \(k\)-component mixture.
MSC:
62F15 | Bayesian inference |
65C05 | Monte Carlo methods |
62F10 | Point estimation |
65C99 | Probabilistic methods, stochastic differential equations |