Une procédure de pénalisation en sélection de modèle repose sur la construction d’une forme pour la pénalité ainsi que sur le choix d’une constante de calibration. Dans cet article, nous étudions, pour le problème d’estimation de la densité, les pénalités obtenues par rééchantillonnage de pénalités idéales. Nous montrons l’efficacité de ces procédures pour l’estimation de la forme des pénalités en prouvant, pour les estimateurs sélectionnés, des inégalités d’oracle fines sans termes résiduelles; les résultats sont valides sous des hypothèses faibles à la fois sur la densité inconnue et sur les collections de modèles. Ces pénalités sont de plus faciles à calibrer puisque la constante asymptotiquement optimale peut être calculée en fonction des poids de rééchantillonnage. En pratique, le nombre de données est toujours fini, nous étudions donc également l’heuristique de pente et justifions l’algorithme de pente qui permet de calibrer la constante de calibration à partir des données.
In order to calibrate a penalization procedure for model selection, the statistician has to choose a shape for the penalty and a leading constant. In this paper, we study, for the marginal density estimation problem, the resampling penalties as general estimators of the shape of an ideal penalty. We prove that the selected estimator satisfies sharp oracle inequalities without remainder terms under a few assumptions on the marginal density and the collection of models. We also study the slope heuristic, which yields a data-driven choice of the leading constant in front of the penalty when the complexity of the models is well-chosen.
@article{AIHPB_2012__48_3_884_0, author = {Lerasle, Matthieu}, title = {Optimal model selection in density estimation}, journal = {Annales de l'I.H.P. Probabilit\'es et statistiques}, volume = {48}, year = {2012}, pages = {884-908}, doi = {10.1214/11-AIHP425}, mrnumber = {2976568}, zbl = {1244.62052}, language = {en}, url = {http://dml.mathdoc.fr/item/AIHPB_2012__48_3_884_0} }
Lerasle, Matthieu. Optimal model selection in density estimation. Annales de l'I.H.P. Probabilités et statistiques, Tome 48 (2012) pp. 884-908. doi : 10.1214/11-AIHP425. http://gdmltest.u-ga.fr/item/AIHPB_2012__48_3_884_0/
[1] Statistical predictor identification. Ann. Inst. Statist. Math. 22 (1970) 203-217. | MR 286233 | Zbl 0259.62076
.[2] Information theory and an extension of the maximum likelihood principle. In Second International Symposium on Information Theory (Tsahkadsor, 1971) 267-281. Akadémiai Kiadó, Budapest, 1973. | MR 483125 | Zbl 0283.62006
.[3] Resampling and model selection. Ph.D. thesis, Université Paris-Sud 11, 2007.
.[4] Model selection by resampling penalization. Electron. J. Stat. 3 (2009) 557-624. | MR 2519533 | Zbl pre06166454
.[5] Data-driven calibration of penalties for least-squares regression. J. Mach. Learn. Res. 10 (2009) 245-279.
and .[6] Risk bounds for model selection via penalization. Probab. Theory Related Fields 113(3) (1999) 301-413. | MR 1679028 | Zbl 0946.62036
, and .[7] Slope heuristics: Overview and implementation. Report INRIA, 2010. Available at http://hal.archives-ouvertes.fr/hal-00461639/fr/.
, and .[8] Model selection for density estimation with -loss. Preprint, 2008.
.[9] From model selection to adaptive estimation. In Festschrift for Lucien Le Cam 55-87. Springer, New York, 1997. | MR 1462939 | Zbl 0920.62042
and .[10] Minimal penalties for Gaussian model selection. Probab. Theory Related Fields 138(1-2) (2007) 33-73. | MR 2288064 | Zbl 1112.62082
and .[11] A Bennett concentration inequality and its application to suprema of empirical processes. C. R. Math. Acad. Sci. Paris 334(6) (2002) 495-500. | MR 1890640 | Zbl 1001.60021
.[12] Sparse density estimation with penalties. In Learning Theory 530-543. Lecture Notes in Comput. Sci. 4539. Springer, Berlin, 2007. | MR 2397610 | Zbl 1203.62053
, and .[13] Density estimation via cross validation: Model selection point of view. Preprint, 2008. Available at arXiv.org:08110802.
.[14] Density estimation by wavelet thresholding. Ann. Statist. 24(2) (1996) 508-539. | MR 1394974 | Zbl 0860.62032
, , and .[15] Bootstrap methods: Another look at the jackknife. Ann. Statist. 7(1) (1979) 1-26. | MR 515681 | Zbl 0406.62024
.[16] Adaptive density estimation under dependence. ESAIM Probab. Stat. 14 (2010) 151-172. | Numdam | MR 2654551 | Zbl 1209.62056
and .[17] Exponential inequalities, with constants, for U-statistics of order two. In Stochastic Inequalities and Applications 55-69. Progr. Probab. 56. Birkhäuser, Basel, 2003. | MR 2073426 | Zbl 1036.60015
and .[18] Concentration around the mean for maxima of empirical processes. Ann. Probab. 33(3) (2005) 1060-1077. | MR 2135312 | Zbl 1066.60023
and .[19] Some comments on . Technometrics 15 (1973) 661-675. | Zbl 0269.62061
.[20] Concentration Inequalities and Model Selection. Lecture Notes in Mathematics 1896. Springer, Berlin, 2007. Lectures from the 33rd Summer School on Probability Theory held in Saint-Flour, July 6-23, 2003, With a foreword by Jean Picard. | MR 2319879 | Zbl 1170.60006
.[21] Adaptive density estimation using the blockwise Stein method. Bernoulli 12(2) (2006) 351-370. | MR 2218559 | Zbl 1098.62040
.[22] Linear and convex aggregation of density estimators. Math. Methods Statist. 16(3) (2007) 260-280. | MR 2356821 | Zbl 1231.62057
and .[23] Empirical choice of histograms and kernel density estimators. Scand. J. Stat. 9(2) (1982) 65-78. | MR 668683 | Zbl 0501.62028
.[24] Estimating the dimension of a model. Ann. Statist. 6 (1978) 461-464. | MR 468014 | Zbl 0379.62005
.[25] Cross-validatory choice and assessment of statistical predictions. J. R. Stat. Soc. Ser. B Stat. Methodol. 36 (1974) 111-147. With discussion by G. A. Barnard, A. C. Atkinson, L. K. Chan, A. P. Dawid, F. Downton, J. Dickey, A. G. Baker, O. Barndorff-Nielsen, D. R. Cox, S. Giesser, D. Hinkley, R. R. Hocking and A. S. Young and with a reply by the authors. | MR 356377 | Zbl 0308.62063
.[26] New concentration inequalities in product spaces. Invent. Math. 126(3) (1996) 505-563. | MR 1419006 | Zbl 0893.60001
.