Gradient descent with random initialization: fast global convergence for nonconvex phase retrieval
成果类型:
Article; Proceedings Paper
署名作者:
Chen, Yuxin; Chi, Yuejie; Fan, Jianqing; Ma, Cong
署名单位:
Princeton University; Carnegie Mellon University; Princeton University
刊物名称:
MATHEMATICAL PROGRAMMING
ISSN/ISSBN:
0025-5610
DOI:
10.1007/s10107-019-01363-6
发表日期:
2019
页码:
5-37
关键词:
摘要:
This paper considers the problem of solving systems of quadratic equations, namely, recovering an object of interest x?Rn from m quadratic equations/samples yi=(ai?x?)2,1im. This problem, also dubbed as phase retrieval, spans multiple domains including physical sciences and machine learning. We investigate the efficacy of gradient descent (or Wirtinger flow) designed for the nonconvex least squares problem. We prove that under Gaussian designs, gradient descentwhen randomly initializedyields an E-accurate solution in O(logn+log(1/E)) iterations given nearly minimal samples, thus achieving near-optimal computational and sample complexities at once. This provides the first global convergence guarantee concerning vanilla gradient descent for phase retrieval, without the need of (i) carefully-designed initialization, (ii) sample splitting, or (iii) sophisticated saddle-point escaping schemes. All of these are achieved by exploiting the statistical models in analyzing optimization algorithms, via a leave-one-out approach that enables the decoupling of certain statistical dependency between the gradient descent iterates and the data.
来源URL: