Abstract
High-dimensional functional data have become increasingly prevalent in
modern applications such as high-frequency financial data and neuroimaging data
analysis. We investigate a class of high-dimensional linear regression models, where
each predictor is a random element in an infinite-dimensional function space, and
the number of functional predictors p can potentially be ultra-high. Assuming that
each of the unknown coefficient functions belongs to some reproducing kernel Hilbert
space (RKHS), we regularize the fitting of the model by imposing a group elastic-net
type of penalty on the RKHS norms of the coefficient functions. We show that our
loss function is Gateaux sub-differentiable, and our functional elastic-net estimator
exists uniquely in the product RKHS. Under suitable sparsity assumptions and a
functional version of the irrepresentable condition, we derive a non-asymptotic tail
bound for variable selection consistency of our method.
Allowing the number of
true functional predictors q to diverge with the sample size, we also show a postselection refined estimator can achieve the oracle minimax optimal prediction rate.
The proposed methods are illustrated through simulation studies and a real-data
application from the Human Connectome Project.
Information
| Preprint No. | SS-2025-0151 |
|---|---|
| Manuscript ID | SS-2025-0151 |
| Complete Authors | Xingche Guo, Yehua Li, Tailen Hsing |
| Corresponding Authors | Yehua Li |
| Emails | yehuali@ucr.edu |
References
- Cai, T. T. and Hall, P. (2006). Prediction in functional linear regression. The Annals of Statistics, 34:2159–2179.
- Cai, T. T. and Yuan, M. (2012). Minimax and adaptive prediction for functional linear regression. Journal of the American Statistical Association, 107:1201–1216.
- Crambes, C., Kneip, A., and Sarda, P. (2009). Smoothing spline estimators for functional linear regression. The Annals of Statistics, 37:35–72.
- Duncan, J., Seitz, R. J., Kolodny, J., Bor, D., Herzog, H., Ahmed, A., Newell,
- F. N., and Emslie, H. (2000). A neural basis for general intelligence. Science, 289(5478):457–460.
- Fan, J. and Li, R. (2001). Variable selection via nonconcave penalized likelihood and its oracle properties. Journal of the American Statistical Association, 96:1348–1360.
- Fan, Y., James, G. M., and Radchenko, P. (2015). Functional additive regression. The Annals of Statistics, 43:2296–2325.
- Finn, E. S., Shen, X., Scheinost, D., Rosenberg, M. D., Huang, J., Chun,
- M. M., Papademetris, X., and Constable, R. T. (2015). Functional connectome fingerprinting: identifying individuals using patterns of brain connectivity. Nature Neuroscience, 18(11):1664–1671.
- Friedman, J., Hastie, T., H¨ofling, H., and Tibshirani, R. (2007). Pathwise coordinate optimization. The Annals of Applied Statistics, 1(2):302–332.
- Greene, A. S., Gao, S., Scheinost, D., and Constable, R. T. (2018). Taskinduced brain state manipulation improves prediction of individual traits. Nature Communications, 9(1):2807.
- Hsing, T. and Eubank, R. (2015). Theoretical foundations of functional data analysis, with an introduction to linear operators, volume 997. John Wiley & Sons.
- James, G. (2002). Generalized linear models with functional predictor variables. Journal of the Royal Statistical Society, Series B, 64:411–432.
- Jia, J. and Yu, B. (2010). On model selection consistency of the elastic net when p ≫n. Statistica Sinica, 20:595–611.
- Jung, R. E. and Haier, R. J. (2007). The parieto-frontal integration theory (p-fit) of intelligence: converging neuroimaging evidence. Behavioral and Brain Sciences, 30(2):135–154.
- Lee, K.-Y., Ji, D., Li, L., Constable, T., and Zhao, H. (2023). Conditional functional graphical models. Journal of the American Statistical Association, 118(541):257–271.
- Lei, J. (2014). Adaptive global testing for functional linear models. Journal of the American Statistical Association, 109:624–634.
- Liu, Y., Li, Y., Carroll, R. J., and Wang, N. (2022). Predictive functional linear models with diverging number of semiparametric single-index interactions. Journal of Econometrics, 230(2):221–239.
- Ma, P., Huang, J. Z., and Zhang, N. (2015). Efficient computation of smoothing splines via adaptive basis sampling. Biometrika, 102(3):631–645.
- M¨uller, H. G. and Stadtm¨uller, U. (2005). Generalized functional linear models. The Annals of Statistics, 33:774–805.
- Qiao, X., Guo, S., and James, G. M. (2019). Functional graphical models. Journal of the American Statistical Association, 114(525):211–222.
- Ramsay, J. O. and Silverman, B. W. (2005). Functional Data Analysis.
- Springer, New York, 2nd edition.
- Ravikumar, P., Lafferty, J., Liu, H., and Wasserman, L. (2009). Sparse additive models. Journal of the Royal Statistical Society: Series B, 71(5):1009–1030.
- Reiss, P. T. and Ogden, R. T. (2007). Functional principal component regression and functional partial least squares. Journal of the American Statistical Association, 102:984–996.
- Ruppert, D., Wand, M. P., and Carroll, R. J. (2003). Semiparametric Regression. Cambridge university press.
- Shang, Z. and Cheng, G. (2015). Nonparametric inference in generalized functional linear models. The Annals of Statistics, 43:1742–1773.
- Sun, X., Du, P., Wang, X., and Ma, P. (2018). Optimal penalized functionon-function regression under a reproducing kernel hilbert space. Journal of the American Statistical Association, 113:1601–1611.
- Tibshirani, R. J. (1996). Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society, Series B, 58:267–288. Van Essen, D. C., Smith, S. M., Barch, D. M., Behrens, T. E., Yacoub, E.,
- Ugurbil, K., Consortium, W.-M. H., et al. (2013). The wu-minn human connectome project: an overview. Neuroimage, 80:62–79.
- Wahba, G. (1990). Spline Models for Observational Data. SIAM, Philadelphia.
- Wainwright, M. J. (2009). Sharp thresholds for high-dimensional and noisy sparsity recovery using ℓ1-constrained quadratic programming (lasso). IEEE Transactions on Information Theory, 55(5):2183–2202.
- Wang, X., Zhu, H., and Initiative, A. D. N. (2017). Generalized scalar-onimage regression models via total variation. Journal of the American Statistical Association, 112(519):1156–1168.
- Xu, D. and Wang, Y. (2021). Low-rank approximation for smoothing spline via eigensystem truncation. Stat, 10(1):e355.
- Xue, K. and Yao, F. (2021). Hypothesis testing in large-scale functional linear regression. Statistica Sinica, 31:1101 – 1123.
- Yuan, M. and Lin, Y. (2006). Model selection and estimation in regression with grouped variables. Journal of the Royal Statistical Society: Series B, 68(1):49–67.
- Zapata, J., Oh, S. Y., and Petersen, A. (2021). Partial separability and functional graphical models for multivariate Gaussian processes. Biometrika, 109(3):665–681.
- Zhang, C.-H. (2010). Nearly unbiased variable selection under minimax concave penalty. The Annals of Statistics, 38:894–942.
- Zhao, P. and Yu, B. (2006). On model selection consistency of lasso. Journal of Machine Learning Research, 7:2541–2563.
- Zhou, H., Yao, F., and Zhang, H. (2023). Functional linear regression for discretely observed data: from ideal to reality. Biometrika, 110(2):381–393.
- Zou, H. and Hastie, T. (2005). Regularization and variable selection via the elastic net. Journal of the Royal Statistical Society: Series B, 67(2):301–320.
- Zou, H. and Zhang, H. H. (2009). On the adaptive elastic-net with a diverging number of parameters. The Annals of Statistics, 37:1733 – 1751.
Acknowledgments
The authors thank the editor, the associate editor, and two anonymous referees
for their many helpful and constructive comments, which led to significant
improvements to our paper.