TY - JOUR
T1 - A Tuning-free Robust and Efficient Approach to High-dimensional Regression
AU - Wang, Lan
AU - Peng, Bo
AU - Bradic, Jelena
AU - Li, Runze
AU - Wu, Yunan
N1 - Publisher Copyright:
© 2020 American Statistical Association.
PY - 2020
Y1 - 2020
N2 - We introduce a novel approach for high-dimensional regression with theoretical guarantees. The new procedure overcomes the challenge of tuning parameter selection of Lasso and possesses several appealing properties. It uses an easily simulated tuning parameter that automatically adapts to both the unknown random error distribution and the correlation structure of the design matrix. It is robust with substantial efficiency gain for heavy-tailed random errors while maintaining high efficiency for normal random errors. Comparing with other alternative robust regression procedures, it also enjoys the property of being equivariant when the response variable undergoes a scale transformation. Computationally, it can be efficiently solved via linear programming. Theoretically, under weak conditions on the random error distribution, we establish a finite-sample error bound with a near-oracle rate for the new estimator with the simulated tuning parameter. Our results make useful contributions to mending the gap between the practice and theory of Lasso and its variants. We also prove that further improvement in efficiency can be achieved by a second-stage enhancement with some light tuning. Our simulation results demonstrate that the proposed methods often outperform cross-validated Lasso in various settings.
AB - We introduce a novel approach for high-dimensional regression with theoretical guarantees. The new procedure overcomes the challenge of tuning parameter selection of Lasso and possesses several appealing properties. It uses an easily simulated tuning parameter that automatically adapts to both the unknown random error distribution and the correlation structure of the design matrix. It is robust with substantial efficiency gain for heavy-tailed random errors while maintaining high efficiency for normal random errors. Comparing with other alternative robust regression procedures, it also enjoys the property of being equivariant when the response variable undergoes a scale transformation. Computationally, it can be efficiently solved via linear programming. Theoretically, under weak conditions on the random error distribution, we establish a finite-sample error bound with a near-oracle rate for the new estimator with the simulated tuning parameter. Our results make useful contributions to mending the gap between the practice and theory of Lasso and its variants. We also prove that further improvement in efficiency can be achieved by a second-stage enhancement with some light tuning. Our simulation results demonstrate that the proposed methods often outperform cross-validated Lasso in various settings.
UR - http://www.scopus.com/inward/record.url?scp=85097763077&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85097763077&partnerID=8YFLogxK
U2 - 10.1080/01621459.2020.1840989
DO - 10.1080/01621459.2020.1840989
M3 - Article
AN - SCOPUS:85097763077
SN - 0162-1459
VL - 115
SP - 1700
EP - 1714
JO - Journal of the American Statistical Association
JF - Journal of the American Statistical Association
IS - 532
ER -