FUNDAMENTAL BARRIERS TO HIGH-DIMENSIONAL REGRESSION WITH CONVEX PENALTIES
成果类型:
Article
署名作者:
Celentano, Michael; Montanari, Andrea
署名单位:
Stanford University
刊物名称:
ANNALS OF STATISTICS
ISSN/ISSBN:
0090-5364
DOI:
10.1214/21-AOS2100
发表日期:
2022
页码:
170-196
关键词:
message-passing algorithms
statistical estimation
variable selection
phase-transitions
robust regression
state evolution
Lasso
slope
INFORMATION
shrinkage
摘要:
In high-dimensional regression, we attempt to estimate a parameter vector beta(0) is an element of R-P from n less than or similar to p observations {(y(i) , x(i))}(i <= n) , where x(i) is an element of R-P is a vector of predictors and y(i) is a response variable. A well-established approach uses convex regularizers to promote specific structures (e.g., sparsity) of the estimate (beta) over cap while allowing for practical algorithms. Theoretical analysis implies that convex penalization schemes have nearly optimal estimation properties in certain settings. However, in general the gaps between statistically optimal estimation (with unbounded computational resources) and convex methods are poorly understood. We show that when the statistican has very simple structural information about the distribution of the entries of beta(0), a large gap frequently exists between the best performance achieved by any convex regularizer satisfying a mild technical condition and either: (i) the optimal statistical error or (ii) the statistical error achieved by optimal approximate message passing algorithms. Remarkably, a gap occurs at high enough signal-to-noise ratio if and only if the distribution of the coordinates of beta(0) is not log-concave. These conclusions follow from an analysis of standard Gaussian designs. Our lower bounds are expected to be generally tight, and we prove tightness under certain conditions.