Adaptive density estimation for clustering with gaussian mixtures

C. Maugis-Rabusseau; B. Michel

ESAIM: Probability and Statistics (2013)

  • Volume: 17, page 698-724
  • ISSN: 1292-8100

Abstract

top
Gaussian mixture models are widely used to study clustering problems. These model-based clustering methods require an accurate estimation of the unknown data density by Gaussian mixtures. In Maugis and Michel (2009), a penalized maximum likelihood estimator is proposed for automatically selecting the number of mixture components. In the present paper, a collection of univariate densities whose logarithm is locally β-Hölder with moment and tail conditions are considered. We show that this penalized estimator is minimax adaptive to the β regularity of such densities in the Hellinger sense.

How to cite

top

Maugis-Rabusseau, C., and Michel, B.. "Adaptive density estimation for clustering with gaussian mixtures." ESAIM: Probability and Statistics 17 (2013): 698-724. <http://eudml.org/doc/273632>.

@article{Maugis2013,
abstract = {Gaussian mixture models are widely used to study clustering problems. These model-based clustering methods require an accurate estimation of the unknown data density by Gaussian mixtures. In Maugis and Michel (2009), a penalized maximum likelihood estimator is proposed for automatically selecting the number of mixture components. In the present paper, a collection of univariate densities whose logarithm is locally β-Hölder with moment and tail conditions are considered. We show that this penalized estimator is minimax adaptive to the β regularity of such densities in the Hellinger sense.},
author = {Maugis-Rabusseau, C., Michel, B.},
journal = {ESAIM: Probability and Statistics},
keywords = {rate adaptive density estimation; gaussian mixture clustering; hellinger risk; non asymptotic model selection; Gaussian mixture clustering; Hellinger risk},
language = {eng},
pages = {698-724},
publisher = {EDP-Sciences},
title = {Adaptive density estimation for clustering with gaussian mixtures},
url = {http://eudml.org/doc/273632},
volume = {17},
year = {2013},
}

TY - JOUR
AU - Maugis-Rabusseau, C.
AU - Michel, B.
TI - Adaptive density estimation for clustering with gaussian mixtures
JO - ESAIM: Probability and Statistics
PY - 2013
PB - EDP-Sciences
VL - 17
SP - 698
EP - 724
AB - Gaussian mixture models are widely used to study clustering problems. These model-based clustering methods require an accurate estimation of the unknown data density by Gaussian mixtures. In Maugis and Michel (2009), a penalized maximum likelihood estimator is proposed for automatically selecting the number of mixture components. In the present paper, a collection of univariate densities whose logarithm is locally β-Hölder with moment and tail conditions are considered. We show that this penalized estimator is minimax adaptive to the β regularity of such densities in the Hellinger sense.
LA - eng
KW - rate adaptive density estimation; gaussian mixture clustering; hellinger risk; non asymptotic model selection; Gaussian mixture clustering; Hellinger risk
UR - http://eudml.org/doc/273632
ER -

References

top
  1. [1] J.-P. Baudry, C. Maugis and B. Michel, Slope heuristics: overview and implementation. Stat. Comput.22 (2011) 455–470. Zbl1322.62007MR2865029
  2. [2] L. Birgé, A new lower bound for multiple hypothesis testing. IEEE Trans. Inform. Theory.51 (2005) 1611–1615. Zbl1283.62030MR2241522
  3. [3] W. Cheney and W. Light, A course in approximation theory, Graduate Studies in Mathematics, vol. 101 of Amer. Math. Soc. Providence, RI (2009). Zbl1167.41001MR2474372
  4. [4] S. Ghosal, J.K. Ghosh and R.V. Ramamoorthi, Posterior consistency of Dirichlet mixtures in density estimation. Ann. Stat.27 (1999) 143–158. Zbl0932.62043MR1701105
  5. [5] S. Ghosal and A. van der Vaart, Entropy and rates of convergence for maximum likelihood and Bayes estimation for mixtures of normal densities. Ann. Stat. 29 (2001) 1233–1263,. Zbl1043.62025MR1873329
  6. [6] S. Ghosal and A. van der Vaart, Posterior convergence rates of Dirichlet mixtures at smooth densities. Ann. Stat.35 (2007) 697–723. Zbl1117.62046MR2336864
  7. [7] U. Grenander, Abstract inference. John Wiley and Sons Inc., New York (1981). Zbl0505.62069MR599175
  8. [8] T. Hangelbroek and A. Ron, Nonlinear approximation using Gaussian kernels. J. Functional Anal.259 (2010) 203–219. Zbl1203.41015MR2610384
  9. [9] J.A. Hartigan, Clustering algorithms, Probab. Math. Stat. John Wiley and Sons, New York-London-Sydney (1975). Zbl0372.62040MR405726
  10. [10] T. Hastie, R. Tibshirani and J. Friedman, The elements of statistical learning, Data mining, inference, and prediction. Statistics. Springer, New York, 2nd edition (2009). Zbl1273.62005MR2722294
  11. [11] W. Kruijer, J. Rousseau and A van der Vaart, Adaptive Bayesian Density Estimation with Location-Scale Mixtures. Electron. J. Statist. 4 (2010) 1225–1257. Zbl1329.62188MR2735885
  12. [12] B. Lindsay, Mixtures Models: Theory, Geometry and Applications. IMS, Hayward, CA (1995). Zbl1163.62326
  13. [13] P. Massart, Concentration Inequalities and Model Selection. École d’été de Probabilités de Saint-Flour, 2003. Lect. Notes Math. Springer (2007). Zbl1170.60006MR2319879
  14. [14] C. Maugis and B. Michel, Adaptive density estimation for clustering with Gaussian mixtures (2011). arXiv:1103.4253v2. Zbl06282493
  15. [15] C. Maugis and B. Michel, Data-driven penalty calibration: a case study for Gaussian mixture model selection. ESAIM: PS 15 (2011) 320–339. Zbl06157520MR2870518
  16. [16] C. Maugis and B. Michel, A non asymptotic penalized criterion for Gaussian mixture model selection. ESAIM: PS 15 (2011) 41–68. Zbl06157507MR2870505
  17. [17] G. McLachlan and D. Peel, Finite Mixture Models. Wiley (2000). Zbl0963.62061MR1789474
  18. [18] A.B. Tsybakov, Introduction to nonparametric estimation. Statistics. Springer, New York (2009). Zbl1029.62034MR2724359
  19. [19] J. Wolfowitz, Minimax estimation of the mean of a normal distribution with known variance. Ann. Math. Stat.21 (1950) 218–230. Zbl0038.09801MR35950

NotesEmbed ?

top

You must be logged in to post comments.