Robust Estimation for Mixture of Probability Tables based on beta-likelihood

Modeling of a large joint probability table is problematic when its variables have a large number of categories. In such a case, a mixture of simpler probability tables could be a good model. And the estimation of such a large probability table frequently has another problem of data sparseness. When constructing mixture models with sparse data, EM estimators based on the β-likelihood are expected more appropriate than those based on the log likelihood. Experimental results show that a mixture model estimated by the βlikelihood approximates a large joint probability table with sparse data more appropriately than EM estimators.

[1]  G. McLachlan,et al.  The EM algorithm and extensions , 1996 .

[2]  Takafumi Kanamori,et al.  Information Geometry of U-Boost and Bregman Divergence , 2004, Neural Computation.

[3]  Thomas Hofmann,et al.  Probabilistic Latent Semantic Analysis , 1999, UAI.

[4]  Shinto Eguchi,et al.  Robust estimation in the normal mixture model , 2006 .

[5]  James S. Albus,et al.  Brains, behavior, and robotics , 1981 .

[6]  A. Agresti,et al.  Categorical Data Analysis , 1991, International Encyclopedia of Statistical Science.