Statistica Sinica

doi:10.5705/ss.2010.081

Ching-Kang Ing and Tze Leung Lai

Abstract:We introduce a fast stepwise regression method, called the orthogonal greedy algorithm (OGA), that selects input variables to enter a -dimensional linear regression model (with , the sample size) sequentially so that the selected variable at each step minimizes the residual sum squares. We derive the convergence rate of OGA and develop a consistent model selection procedure along the OGA path that can adjust for potential spuriousness of the greedily chosen regressors among a large number of candidate variables. The resultant regression estimate is shown to have the oracle property of being equivalent to least squares regression on an asymptotically minimal set of relevant regressors under a strong sparsity condition.

Key words and phrases:Componentwise linear regression, greedy algorithm, high-dimensional information criterion, Lasso, oracle property and inequalities, sparsity.