Iterative feature selection in least square regression estimation
Alquier, Pierre
Ann. Inst. H. Poincaré Probab. Statist., Tome 44 (2008) no. 2, p. 47-88 / Harvested from Project Euclid
This paper presents a new algorithm to perform regression estimation, in both the inductive and transductive setting. The estimator is defined as a linear combination of functions in a given dictionary. Coefficients of the combinations are computed sequentially using projection on some simple sets. These sets are defined as confidence regions provided by a deviation (PAC) inequality on an estimator in one-dimensional models. We prove that every projection the algorithm actually improves the performance of the estimator. We give all the estimators and results at first in the inductive case, where the algorithm requires the knowledge of the distribution of the design, and then in the transductive case, which seems a more natural application for this algorithm as we do not need particular information on the distribution of the design in this case. We finally show a connection with oracle inequalities, making us able to prove that the estimator reaches minimax rates of convergence in Sobolev and Besov spaces.
Publié le : 2008-02-15
Classification:  Regression estimation,  Statistical learning,  Confidence regions,  Thresholding methods,  Support vector machines,  62G08,  62G15,  68T05
@article{1203969868,
     author = {Alquier, Pierre},
     title = {Iterative feature selection in least square regression estimation},
     journal = {Ann. Inst. H. Poincar\'e Probab. Statist.},
     volume = {44},
     number = {2},
     year = {2008},
     pages = { 47-88},
     language = {en},
     url = {http://dml.mathdoc.fr/item/1203969868}
}
Alquier, Pierre. Iterative feature selection in least square regression estimation. Ann. Inst. H. Poincaré Probab. Statist., Tome 44 (2008) no. 2, pp.  47-88. http://gdmltest.u-ga.fr/item/1203969868/