Random thresholds for linear model selection

Marc Lavielle; Carenne Ludeña

ESAIM: Probability and Statistics (2008)

  • Volume: 12, page 173-195
  • ISSN: 1292-8100

Abstract

top
A method is introduced to select the significant or non null mean terms among a collection of independent random variables. As an application we consider the problem of recovering the significant coefficients in non ordered model selection. The method is based on a convenient random centering of the partial sums of the ordered observations. Based on L-statistics methods we show consistency of the proposed estimator. An extension to unknown parametric distributions is considered. Simulated examples are included to show the accuracy of the estimator. An example of signal denoising with wavelet thresholding is also discussed.

How to cite

top

Lavielle, Marc, and Ludeña, Carenne. "Random thresholds for linear model selection." ESAIM: Probability and Statistics 12 (2008): 173-195. <http://eudml.org/doc/250386>.

@article{Lavielle2008,
abstract = { A method is introduced to select the significant or non null mean terms among a collection of independent random variables. As an application we consider the problem of recovering the significant coefficients in non ordered model selection. The method is based on a convenient random centering of the partial sums of the ordered observations. Based on L-statistics methods we show consistency of the proposed estimator. An extension to unknown parametric distributions is considered. Simulated examples are included to show the accuracy of the estimator. An example of signal denoising with wavelet thresholding is also discussed. },
author = {Lavielle, Marc, Ludeña, Carenne},
journal = {ESAIM: Probability and Statistics},
keywords = {Adaptive estimation; linear model selection; hard thresholding; random thresholding; L-statistics; adaptive estimation; linear model selection; -statistics},
language = {eng},
month = {1},
pages = {173-195},
publisher = {EDP Sciences},
title = {Random thresholds for linear model selection},
url = {http://eudml.org/doc/250386},
volume = {12},
year = {2008},
}

TY - JOUR
AU - Lavielle, Marc
AU - Ludeña, Carenne
TI - Random thresholds for linear model selection
JO - ESAIM: Probability and Statistics
DA - 2008/1//
PB - EDP Sciences
VL - 12
SP - 173
EP - 195
AB - A method is introduced to select the significant or non null mean terms among a collection of independent random variables. As an application we consider the problem of recovering the significant coefficients in non ordered model selection. The method is based on a convenient random centering of the partial sums of the ordered observations. Based on L-statistics methods we show consistency of the proposed estimator. An extension to unknown parametric distributions is considered. Simulated examples are included to show the accuracy of the estimator. An example of signal denoising with wavelet thresholding is also discussed.
LA - eng
KW - Adaptive estimation; linear model selection; hard thresholding; random thresholding; L-statistics; adaptive estimation; linear model selection; -statistics
UR - http://eudml.org/doc/250386
ER -

References

top
  1. F. Abramovich and Y. Benjamini, Adaptive thresholding of wavelet coefficients. CSDA4 (1996) 351–361.  
  2. A. Antoniadis, J. Bigot and T. Sapatinas, Wavelet estimators in nonparametric regression: a comparative simulation study. J. Statist. Software6 (2001) 1–83.  
  3. B. Arnold, N. Balakrishnan and H. Nagaraja, A first course in order statistics. Wiley series in probability (1993).  Zbl1172.62017
  4. A. Barron, L. Birgé and P. Massart, Risk bounds for model selection via penalization. Probab. Theory Related Fields113 (1999) 301–413.  Zbl0946.62036
  5. L. Birgé and P. Massart, Minimal penalties for Gaussian model selection. Probab. Theor. Rel. Fields138 (2007) 33–73.  Zbl1112.62082
  6. O. Bousquet, Concentration inequalities for sub-additive functions using the entropy method, in Stochastic inequalities and applications, Progr. Probab. Birkhäuser, Basel 56 (2003) 213–247.  Zbl1037.60015
  7. T. Cai, Adaptive wavelet estimation: a block thresholding and oracle inequality approach. Ann. Stat.3 (1999) 898–924.  Zbl0954.62047
  8. L. de Haan, On regular variation and its application to the weak convergence of sample extremes. 3rd ed., Mathematical Centre Tracts 32 Amsterdam (1975).  
  9. D. Donoho and J. Jin, Higher criticism for detecting sparse heterogeneous mixtures. Ann. Stat.32 (2004) 962–994.  Zbl1092.62051
  10. D. Donoho and I. Johnstone, Ideal spatial adaptation by wavelet shrinkage. Biometrika81 (1994) 425–455.  Zbl0815.62019
  11. M. Falk, A note on uniform asymptotic normality of intermidiate order statistics. Ann. Inst. Statist. Math41 (1989) 19–29.  Zbl0693.62019
  12. J. Fan and R. Li, Variable selection via nonconcave penalized likelihood and its oracle properties. J. Amer. Statist. Assoc.96 (2001) 1348–1360.  Zbl1073.62547
  13. T. Hastie, R. Tibshirani and J. Friedman, The elements of statistical learning. Springer, Series in statistics (2001).  Zbl0973.62007
  14. N. Meinhausen and J. Rice, Estimating the proportion of false null hypothesis among a large number of independently tested hypothesis. Ann. Stat.34 (2006) 373–393.  Zbl1091.62059
  15. M.S. Pinsker, Optimal filtration of square-integrable signals in Gaussian noise. Probl. Peredachi Inform.2 (1980) 52–68.  
  16. G. Shorak and J. Wellner, Empirical processes with Applications to Statistics. Wiley (1986).  
  17. R. Tibshirani, Regression shrinkage and selection via the lasso. J. Roy. Statist. Soc. B58 (1996) 267–288.  Zbl0850.62538

NotesEmbed ?

top

You must be logged in to post comments.

To embed these notes on your page include the following JavaScript code on your page where you want the notes to appear.

Only the controls for the widget will be shown in your chosen language. Notes will be shown in their authored language.

Tells the widget how many notes to show per page. You can cycle through additional notes using the next and previous controls.

    
                

Note: Best practice suggests putting the JavaScript code just before the closing </body> tag.