Back To Index Previous Article Next Article Full Text

Statistica Sinica 36 (2026), 259-278

PENALIZED REGRESSION WITH MULTIPLE
LOSS FUNCTIONS AND VARIABLE
SELECTION BY VOTING

Guorong Dai*1, Ursula U. Müller2 and Raymond J. Carroll2

1Fudan University and 2Texas A&M University

Abstract: We consider a sparse linear model with a fixed design matrix in a high dimensional scenario. We introduce a new variable selection procedure called "voting", which combines the results from multiple regression models with different penalized loss functions to select the relevant predictors. A predictor is included in the final model if it receives enough votes, i.e. is selected by most of the individual models. By employing multiple different loss functions our method takes various properties of the error distribution into account. This is in contrast to the standard penalized regression approach, which typically relies on just one criterion. When that single criterion is not met the standard approach is likely to fail, whereas our method is still able to identify the underlying sparse model. Working with the voting procedure reduces the number of predictors that are incorrectly selected, which simplifies the structure and improves the interpretability of the fitted model. We prove model selection consistency and illustrate the advantages of our method numerically using simulated and real data sets.

Key words and phrases: High dimensional data, linear model, model selection consistency, sparse estimators.


Back To Index Previous Article Next Article Full Text