Adaptive density estimation for clustering with gaussian mixtures

C. Maugis-Rabusseau; B. Michel

ESAIM: Probability and Statistics (2013)

  • Volume: 17, page 698-724
  • ISSN: 1292-8100

Abstract

top
Gaussian mixture models are widely used to study clustering problems. These model-based clustering methods require an accurate estimation of the unknown data density by Gaussian mixtures. In Maugis and Michel (2009), a penalized maximum likelihood estimator is proposed for automatically selecting the number of mixture components. In the present paper, a collection of univariate densities whose logarithm is locally β-Hölder with moment and tail conditions are considered. We show that this penalized estimator is minimax adaptive to the β regularity of such densities in the Hellinger sense.

How to cite

top

Maugis-Rabusseau, C., and Michel, B.. "Adaptive density estimation for clustering with gaussian mixtures." ESAIM: Probability and Statistics 17 (2013): 698-724. <http://eudml.org/doc/273632>.

@article{Maugis2013,
abstract = {Gaussian mixture models are widely used to study clustering problems. These model-based clustering methods require an accurate estimation of the unknown data density by Gaussian mixtures. In Maugis and Michel (2009), a penalized maximum likelihood estimator is proposed for automatically selecting the number of mixture components. In the present paper, a collection of univariate densities whose logarithm is locally β-Hölder with moment and tail conditions are considered. We show that this penalized estimator is minimax adaptive to the β regularity of such densities in the Hellinger sense.},
author = {Maugis-Rabusseau, C., Michel, B.},
journal = {ESAIM: Probability and Statistics},
keywords = {rate adaptive density estimation; gaussian mixture clustering; hellinger risk; non asymptotic model selection; Gaussian mixture clustering; Hellinger risk},
language = {eng},
pages = {698-724},
publisher = {EDP-Sciences},
title = {Adaptive density estimation for clustering with gaussian mixtures},
url = {http://eudml.org/doc/273632},
volume = {17},
year = {2013},
}

TY - JOUR
AU - Maugis-Rabusseau, C.
AU - Michel, B.
TI - Adaptive density estimation for clustering with gaussian mixtures
JO - ESAIM: Probability and Statistics
PY - 2013
PB - EDP-Sciences
VL - 17
SP - 698
EP - 724
AB - Gaussian mixture models are widely used to study clustering problems. These model-based clustering methods require an accurate estimation of the unknown data density by Gaussian mixtures. In Maugis and Michel (2009), a penalized maximum likelihood estimator is proposed for automatically selecting the number of mixture components. In the present paper, a collection of univariate densities whose logarithm is locally β-Hölder with moment and tail conditions are considered. We show that this penalized estimator is minimax adaptive to the β regularity of such densities in the Hellinger sense.
LA - eng
KW - rate adaptive density estimation; gaussian mixture clustering; hellinger risk; non asymptotic model selection; Gaussian mixture clustering; Hellinger risk
UR - http://eudml.org/doc/273632
ER -

References

top
  1. [1] J.-P. Baudry, C. Maugis and B. Michel, Slope heuristics: overview and implementation. Stat. Comput.22 (2011) 455–470. Zbl1322.62007MR2865029
  2. [2] L. Birgé, A new lower bound for multiple hypothesis testing. IEEE Trans. Inform. Theory.51 (2005) 1611–1615. Zbl1283.62030MR2241522
  3. [3] W. Cheney and W. Light, A course in approximation theory, Graduate Studies in Mathematics, vol. 101 of Amer. Math. Soc. Providence, RI (2009). Zbl1167.41001MR2474372
  4. [4] S. Ghosal, J.K. Ghosh and R.V. Ramamoorthi, Posterior consistency of Dirichlet mixtures in density estimation. Ann. Stat.27 (1999) 143–158. Zbl0932.62043MR1701105
  5. [5] S. Ghosal and A. van der Vaart, Entropy and rates of convergence for maximum likelihood and Bayes estimation for mixtures of normal densities. Ann. Stat. 29 (2001) 1233–1263,. Zbl1043.62025MR1873329
  6. [6] S. Ghosal and A. van der Vaart, Posterior convergence rates of Dirichlet mixtures at smooth densities. Ann. Stat.35 (2007) 697–723. Zbl1117.62046MR2336864
  7. [7] U. Grenander, Abstract inference. John Wiley and Sons Inc., New York (1981). Zbl0505.62069MR599175
  8. [8] T. Hangelbroek and A. Ron, Nonlinear approximation using Gaussian kernels. J. Functional Anal.259 (2010) 203–219. Zbl1203.41015MR2610384
  9. [9] J.A. Hartigan, Clustering algorithms, Probab. Math. Stat. John Wiley and Sons, New York-London-Sydney (1975). Zbl0372.62040MR405726
  10. [10] T. Hastie, R. Tibshirani and J. Friedman, The elements of statistical learning, Data mining, inference, and prediction. Statistics. Springer, New York, 2nd edition (2009). Zbl1273.62005MR2722294
  11. [11] W. Kruijer, J. Rousseau and A van der Vaart, Adaptive Bayesian Density Estimation with Location-Scale Mixtures. Electron. J. Statist. 4 (2010) 1225–1257. Zbl1329.62188MR2735885
  12. [12] B. Lindsay, Mixtures Models: Theory, Geometry and Applications. IMS, Hayward, CA (1995). Zbl1163.62326
  13. [13] P. Massart, Concentration Inequalities and Model Selection. École d’été de Probabilités de Saint-Flour, 2003. Lect. Notes Math. Springer (2007). Zbl1170.60006MR2319879
  14. [14] C. Maugis and B. Michel, Adaptive density estimation for clustering with Gaussian mixtures (2011). arXiv:1103.4253v2. Zbl06282493
  15. [15] C. Maugis and B. Michel, Data-driven penalty calibration: a case study for Gaussian mixture model selection. ESAIM: PS 15 (2011) 320–339. Zbl06157520MR2870518
  16. [16] C. Maugis and B. Michel, A non asymptotic penalized criterion for Gaussian mixture model selection. ESAIM: PS 15 (2011) 41–68. Zbl06157507MR2870505
  17. [17] G. McLachlan and D. Peel, Finite Mixture Models. Wiley (2000). Zbl0963.62061MR1789474
  18. [18] A.B. Tsybakov, Introduction to nonparametric estimation. Statistics. Springer, New York (2009). Zbl1029.62034MR2724359
  19. [19] J. Wolfowitz, Minimax estimation of the mean of a normal distribution with known variance. Ann. Math. Stat.21 (1950) 218–230. Zbl0038.09801MR35950

NotesEmbed ?

top

You must be logged in to post comments.

To embed these notes on your page include the following JavaScript code on your page where you want the notes to appear.

Only the controls for the widget will be shown in your chosen language. Notes will be shown in their authored language.

Tells the widget how many notes to show per page. You can cycle through additional notes using the next and previous controls.

    
                

Note: Best practice suggests putting the JavaScript code just before the closing </body> tag.