Back To Index Previous Article Next Article Full Text

Statistica Sinica 35 (2025), 919-937

REGULARIZED ADAPTIVE HUBER MATRIX
REGRESSION AND DISTRIBUTED LEARNING

Yue Wang1, Wenqi Lu2, Lei Wang2, Zhongyi Zhu3,
Hongmei Lin*4 and Heng Lian1,5

1City University of Hong Kong, 2Nankai University, 3Fudan University,
4Shanghai University of International Business and Economics
and 5City University of Hong Kong Shenzhen Research Institute

Abstract: Matrix regression provides a powerful technique for analyzing matrix-type data, as exemplified by many contemporary applications. Despite the rapid advance, distributed learning for robust matrix regression to deal with heavy-tailed noises in the big data regime still remains untouched. In this paper, we first consider adaptive Huber matrix regression with a nuclear norm penalty, which enjoys insensitivity to heavy-tailed noises without losing the statistical accuracy. To further enhance the scalability in massive data applications, we employ the communication-efficient surrogate likelihood framework to develop distributed robust matrix regression, which can be efficiently implemented through the ADMM algorithms. Under only bounded (1+δ)-th moment on the noise for some δ ∈ (0, 1], we provide upper bounds for the estimation error of the central estimator and the distributed estimator, and prove they can achieve the same rate as established with sub-Gaussian tails when only the second moment of noise exists. Numerical studies verify the advantage of the proposed method over existing methods in heavy-tailed noise settings.

Key words and phrases: Big data, communication-efficient, Huber loss, nuclear norm, robust matrix regression.

Back To Index Previous Article Next Article Full Text