Nous nous intéressons à l'estimation de l'espérance conditionnelle d'une variable gaussienne. Ce problème est courant lorsque l'on veut estimer le graphe ou la distribution d'un modèle graphique gaussien. Dans cet article, nous introduisons une procédure de sélection de modèle basée sur la minimisation d'un critère des moindres carrés pénalisés. Cette méthode générale permet de traiter un grand nombre de problèmes comme la sélection ordonnée ou la sélection complête de variables. De plus, elle reste valable dans un cadre de « grande dimension »: lorsque le nombre de covariables est bien plus élevé que le nombre d'observations. L'estimateur obtenue vérifie une inégalité oracle non-asymptotique et ce quelque soit la corrélation entre les covariables. Nous calculons également des vitesses minimax d'estimation dans ce cadre et montrons que notre procédure vérifie diverses propriétés d'adaptation.
We consider the problem of estimating the conditional mean of a real gaussian variable Y=∑i=1pθiXi+ɛ where the vector of the covariates (Xi)1≤i≤p follows a joint gaussian distribution. This issue often occurs when one aims at estimating the graph or the distribution of a gaussian graphical model. We introduce a general model selection procedure which is based on the minimization of a penalized least squares type criterion. It handles a variety of problems such as ordered and complete variable selection, allows to incorporate some prior knowledge on the model and applies when the number of covariates p is larger than the number of observations n. Moreover, it is shown to achieve a non-asymptotic oracle inequality independently of the correlation structure of the covariates. We also exhibit various minimax rates of estimation in the considered framework and hence derive adaptivity properties of our procedure.
Mots-clés : model selection, linear regression, oracle inequalities, gaussian graphical models, minimax rates of estimation
@article{AIHPB_2010__46_2_480_0, author = {Verzelen, Nicolas}, title = {High-dimensional gaussian model selection on a gaussian design}, journal = {Annales de l'I.H.P. Probabilit\'es et statistiques}, pages = {480--524}, publisher = {Gauthier-Villars}, volume = {46}, number = {2}, year = {2010}, doi = {10.1214/09-AIHP321}, mrnumber = {2667707}, zbl = {1191.62076}, language = {en}, url = {http://www.numdam.org/articles/10.1214/09-AIHP321/} }
TY - JOUR AU - Verzelen, Nicolas TI - High-dimensional gaussian model selection on a gaussian design JO - Annales de l'I.H.P. Probabilités et statistiques PY - 2010 SP - 480 EP - 524 VL - 46 IS - 2 PB - Gauthier-Villars UR - http://www.numdam.org/articles/10.1214/09-AIHP321/ DO - 10.1214/09-AIHP321 LA - en ID - AIHPB_2010__46_2_480_0 ER -
%0 Journal Article %A Verzelen, Nicolas %T High-dimensional gaussian model selection on a gaussian design %J Annales de l'I.H.P. Probabilités et statistiques %D 2010 %P 480-524 %V 46 %N 2 %I Gauthier-Villars %U http://www.numdam.org/articles/10.1214/09-AIHP321/ %R 10.1214/09-AIHP321 %G en %F AIHPB_2010__46_2_480_0
Verzelen, Nicolas. High-dimensional gaussian model selection on a gaussian design. Annales de l'I.H.P. Probabilités et statistiques, Tome 46 (2010) no. 2, pp. 480-524. doi : 10.1214/09-AIHP321. http://www.numdam.org/articles/10.1214/09-AIHP321/
[1] Statistical predictor identification. Ann. Inst. Statist. Math. 22 (1970) 203-217. | MR | Zbl
.[2] A new look at the statistical model identification. IEEE Trans. Automat. Control 19 (1974) 716-723. | MR | Zbl
.[3] Model selection by resampling penalization. Electron. J. Stat. 3 (2009) 557-624. | MR
.[4] Gaussian model selection with an unknown variance. Ann. Statist. 37 (2009) 630-672. | MR | Zbl
, and .[5] Simultaneous analysis of Lasso and Dantzig selector. Ann. Statist. 37 (2009) 1705-1732. | MR | Zbl
, and .[6] A new lower bound for multiple hypothesis testing. IEEE Trans. Inform. Theory 51 (2005) 1611-1615. | MR | Zbl
.[7] Minimum contrast estimators on sieves: Exponential bounds and rates of convergence. Bernoulli 4 (1998) 329-375. | MR | Zbl
and .[8] Gaussian model selection. J. Eur. Math. Soc. (JEMS) 3 (2001) 203-268. | EuDML | MR | Zbl
and .[9] Minimal penalties for Gaussian model selection. Probab. Theory Related Fields 138 (2007) 33-73. | MR | Zbl
and .[10] Aggregation for Gaussian regression. Ann. Statist. 35 (2007) 1674-1697. | MR | Zbl
, and .[11] Sparsity oracle inequalities for the Lasso. Electron. J. Stat. 1 (2007) 169-194 (electronic). | MR | Zbl
, and .[12] Decoding by linear programming. IEEE Trans. Inform. Theory 51 (2005) 4203-4215. | MR
and .[13] The Dantzig selector: Statistical estimation when p is much larger than n. Ann. Statist. 35 (2007) 2313-2351. | MR | Zbl
and . and Y. Plan. Near-ideal model selection by l1 minimization. Ann. Statist. To appear, 2009. |[15] Probabilistic Networks and Expert Systems. Statistics for Engineering and Information Science. Springer, New York, 1999. | MR | Zbl
, , and .[16] Statistics for Spatial Data. Wiley Series in Probability and Mathematical Statistics: Applied Probability and Statistics. Wiley, New York, 1993. (Revised reprint of the 1991 edition, Wiley.) | MR | Zbl
.[17] Local operator theory, random matrices and Banach spaces. In Handbook of the Geometry of Banach Spaces, Vol. I 317-366. North-Holland, Amsterdam, 2001. | MR | Zbl
and .[18] Estimation of Gaussian graphs by model selection. Electron. J. Stat. 2 (2008) 542-563. | MR
.[19] Power-law correlations, related models for long-range dependence and their simulation. J. Appl. Probab. 37 (2000) 1104-1109. | MR | Zbl
.[20] Estimating high-dimensional directed acyclic graphs with the PC-algorithm. J. Mach. Learn. Res. 8 (2007) 613-636.
and .[21] Adaptive estimation of a quadratic functional by model selection. Ann. Statist. 28 (2000) 1302-1338. | MR | Zbl
and .[22] Graphical Models. Oxford Statistical Science Series 17. The Clarendon Press, Oxford University Press, New York, 1996. | MR | Zbl
.[23] Some comments on Cp. Technometrics 15 (1973) 661-675. | Zbl
.[24] Concentration Inequalities and Model Selection. Lecture Notes in Mathematics 1896. Springer, Berlin, 2007. (Lectures from the 33rd Summer School on Probability Theory held in Saint-Flour, July 6-23, 2003, with a foreword by Jean Picard.) | MR | Zbl
.[25] High-dimensional graphs and variable selection with the Lasso. Ann. Statist. 34 (2006) 1436-1462. | MR | Zbl
and .[26] Decoupling. Probability and Its Applications. Springer, New York, 1999. (From dependence to independence, randomly stopped processes. U-statistics and processes. Martingales and beyond.) | MR | Zbl
and .[27] Moments for the inverted Wishart distribution. Scand. J. Statist. 15 (1988) 97-109. | MR | Zbl
.[28] Gaussian Markov Random Fields: Theory and Applications. Monographs on Statistics and Applied Probability 104. Chapman & Hall/CRC, London, 2005. | MR | Zbl
and .[29] Causal protein-signaling networks derived from multiparameter single-cell data. Science 308 (2005) 523-529.
, , , and .[30] An empirical Bayes approach to inferring large-scale gene association network. Bioinformatics 21 (2005) 754-764.
and .[31] Estimating the dimension of a model. Ann. Statist. 6 (1978) 461-464. | MR | Zbl
.[32] An optimal selection of regression variables. Biometrika 68 (1981) 45-54. | MR | Zbl
.[33] An asymptotically optimal histogram selection rule. In Proceedings of the Berkeley Conference in Honor of Jerzy Neyman and Jack Kiefer, Vol. II (Berkeley, Calif., 1983) 513-520. Wadsworth Statist./Probab. Ser. Wadsworth, Belmont, CA, 1985. | MR
.[34] Regression shrinkage and selection via the lasso. J. Roy. Statist. Soc. Ser. B 58 (1996) 267-288. | MR | Zbl
.[35] Optimal rates of aggregation. In 16th Annual Conference on Learning Theory 2777 303-313. Springer, Heidelberg, 2003. | Zbl
.[36] Goodness-of-fit tests for high-dimensional Gaussian linear models. Ann. Statist. To appear, 2009. | MR | Zbl
and .[37] Information-theoretic limits on sparsity recovery in the high-dimensional and noisy setting. Technical Report 725, Department of Statistics, UC Berkeley, 2007.
.[38] Sparse graphical Gaussian modelling of the isoprenoid gene network in arabidopsis thaliana. Genome Biology 5 (2004), no. R92.
, , , , , , , , , , , and .[39] MR
and B. Yu. On model selection consistency of Lasso. J. Mach. Learn. Res. 7 (2006) 2541-2563. |[40] The adaptive Lasso and its oracle properties. J. Amer. Statist. Assoc. 101 (2006) 1418-1429. | MR | Zbl
.Cité par Sources :