Back To Index Previous Article Next Article Full Text

Statistica Sinica 25 (2015), 1185-1206

A SMALL-SAMPLE CHOICE OF THE TUNING PARAMETER
IN RIDGE REGRESSION
Philip S. Boonstra, Bhramar Mukherjee and Jeremy M. G. Taylor
University of Michigan

Abstract: We propose new approaches for choosing the shrinkage parameter in ridge regression, a penalized likelihood method for regularizing linear regression coefficients, when the number of observations is small relative to the number of parameters. Existing methods may lead to extreme choices of this parameter, either shrinking the coefficients insufficiently or by too much. Within this “small-n, large-p” context, we suggest a correction to the common generalized cross-validation (GCV) method that preserves the asymptotic optimality of the original GCV. We also introduce the notion of a “hyperpenalty”, which shrinks the shrinkage parameter itself, and make a specific recommendation regarding the choice of hyperpenalty that empirically works well in a broad range of scenarios. A simple algorithm jointly estimates the shrinkage parameter and regression coefficients in the hyperpenalized likelihood. In a comprehensive simulation study of small-sample scenarios and in the analysis of a gene-expression dataset, our proposed approaches offer superior prediction over nine other existing methods.

Key words and phrases: Akaike’s information criterion, cross-validation, generalized cross-validation, hyperpenalty, marginal likelihood, penalized likelihood.

Back To Index Previous Article Next Article Full Text