Improved Initialization of the EM Algorithm for Mixture Model Parameter Estimation

A commonly used tool for estimating the parameters of a mixture model is the Expectation−Maximization (EM) algorithm, which is an iterative procedure that can serve as a maximum-likelihood estimator. The EM algorithm has well-documented drawbacks, such as the need for good initial values a...

Full description

Bibliographic Details
Main Authors: Branislav Panić, Jernej Klemenc, Marko Nagode
Format: Article
Language:English
Published: MDPI AG 2020-03-01
Series:Mathematics
Subjects:
Online Access:https://www.mdpi.com/2227-7390/8/3/373
Description
Summary:A commonly used tool for estimating the parameters of a mixture model is the Expectation&#8722;Maximization (EM) algorithm, which is an iterative procedure that can serve as a maximum-likelihood estimator. The EM algorithm has well-documented drawbacks, such as the need for good initial values and the possibility of being trapped in local optima. Nevertheless, because of its appealing properties, EM plays an important role in estimating the parameters of mixture models. To overcome these initialization problems with EM, in this paper, we propose the Rough-Enhanced-Bayes mixture estimation (REBMIX) algorithm as a more effective initialization algorithm. Three different strategies are derived for dealing with the unknown number of components in the mixture model. These strategies are thoroughly tested on artificial datasets, density&#8722;estimation datasets and image&#8722;segmentation problems and compared with state-of-the-art initialization methods for the EM. Our proposal shows promising results in terms of clustering and density-estimation performance as well as in terms of computational efficiency. All the improvements are implemented in the <b>rebmix</b> R package.
ISSN:2227-7390