We propose a feature selection method for density estimation with quadratic loss. This method relies on the study of unidimensional approximation models and on the definition of confidence regions for the density thanks to these models. It is quite general and includes cases of interest like detection of relevant wavelets coefficients or selection of support vectors in SVM. In the general case, we prove that every selected feature actually improves the performance of the estimator. In the case where features are defined by wavelets, we prove that this method is adaptative near minimax (up to a log term) in some Besov spaces. We end the paper by simulations indicating that it must be possible to extend the adaptation result to other features.

Keywords: density estimation, support vector machines, kernel algorithms, thresholding methods, wavelets

@article{PS_2008__12__438_0, author = {Alquier, Pierre}, title = {Density estimation with quadratic loss : a confidence intervals method}, journal = {ESAIM: Probability and Statistics}, pages = {438--463}, publisher = {EDP-Sciences}, volume = {12}, year = {2008}, doi = {10.1051/ps:2007050}, mrnumber = {2437718}, language = {en}, url = {http://www.numdam.org/articles/10.1051/ps:2007050/} }

TY - JOUR AU - Alquier, Pierre TI - Density estimation with quadratic loss : a confidence intervals method JO - ESAIM: Probability and Statistics PY - 2008 SP - 438 EP - 463 VL - 12 PB - EDP-Sciences UR - http://www.numdam.org/articles/10.1051/ps:2007050/ DO - 10.1051/ps:2007050 LA - en ID - PS_2008__12__438_0 ER -

Alquier, Pierre. Density estimation with quadratic loss : a confidence intervals method. ESAIM: Probability and Statistics, Volume 12 (2008), pp. 438-463. doi : 10.1051/ps:2007050. http://www.numdam.org/articles/10.1051/ps:2007050/

[1] A new look at the statistical model identification. IEEE Trans. Autom. Control 19 (1974) 716-723. | MR | Zbl

,[2] Iterative Feature Selection In Least Square Regression Estimation. Ann. Inst. H. Poincaré B: Probab. Statist. 44 (2008) 47-88. | Numdam | MR

,[3] Adaptative Approximation and Learning by Greedy Algorithms, preprint (2006). | MR | Zbl

, , and ,[4] Kernel Projection Machine: A New Tool for Pattern Recognition. Proceedings of NIPS (2004).

, , and ,[5] A training algorithm for optimal margin classifiers, in Proceedings of the 5th Annual ACM Workshop on Computational Learning Theory, D. Haussler (ed.), ACM Press (1992) 144-152.

, and ,[6] Wavelet Estimation for Samples with Random Uniform Design. Stat. Probab. Lett. 42 (1999) 313-321. | MR | Zbl

and ,[7] Statistical learning theory and stochastic optimization, Lecture Notes, Saint-Flour Summer School on Probability Theory (2001), Springer. | MR | Zbl

,[8] PAC-Bayesian Inductive and Transductive Learning, manuscript (2006).

,[9] A PAC-Bayesian approach to adaptative classification, preprint Laboratoire de Probabilités et Modèles Aléatoires (2003).

,[10] Wavelet methods in numerical analysis, in Handbook of numerical analysis, Vol. VII, North-Holland, Amsterdam (2000) 417-711. | MR | Zbl

,[11] Ten Lectures on Wavelets. SIAM, Philadelphia (1992). | MR | Zbl

,[12] Ideal Spatial Adaptation by Wavelets. Biometrika 81 (1994) 425-455. | MR | Zbl

and ,[13] Density Estimation by Wavelet Thresholding. Ann. Statist. 24 (1996) 508-539. | MR | Zbl

, , and ,[14] Nonparametric roughness penalties for probability densities. Biometrika 58 (1971) 255-277. | MR | Zbl

and ,[15] Wavelets, Approximations and Statistical Applications. Lecture Notes in Statistics, Springer (1998). | MR | Zbl

, , and ,[16] Exact Mean Integrated Square Error. Ann. Statist. 20 (1992) 712-736. | MR | Zbl

and ,[17] Symmetrization Approach to Concentration Inequalities for Empirical Processes. Ann. Probab. 31 (2003) 2068-2081. | MR | Zbl

,[18] R Development Core Team, R: A Language And Environment For Statistical Computing, R Foundation For Statistical Computing, Vienna, Austria, 2004. URL http://www.R-project.org.

[19] Large Scale Multiple Kernel Learning. J. Machine Learning Research 7 (2006) 1531-1565. | MR

, , and ,[20] Modeling by shortest data description. Automatica 14 (1978) 465-471. | Zbl

,[21] PAC-Bayesian Generalization Error Bounds for Gaussian Process Classification. J. Machine Learning Res. 3 (2002) 233-269. | MR | Zbl

,[22] The Relevance Vector Machine, in Advances in Neural Information Processing Systems, San Mateo, CA (2000). Morgan Kaufmann.

,[23] Introduction à l'estimation non-paramétrique. Mathématiques et Applications, Springer (2004). | MR | Zbl

,[24] The nature of statistical learning theory. Springer Verlag (1998). | MR | Zbl

,[25] Chen-xi Zhang and Ya-zhu Chen, SVM for density estimation and application to medical image segmentation. J. Zhejiang Univ. Sci. B 7 (2006).

, ,*Cited by Sources: *