RESTRICTED STRONG CONVEXITY IMPLIES WEAK SUBMODULARITY

成果类型:
Article
署名作者:
Elenberg, Ethan R.; Khanna, Rajiv; Dimakis, Alexandros G.; Negahban, Sahand
署名单位:
University of Texas System; University of Texas Austin; Yale University
刊物名称:
ANNALS OF STATISTICS
ISSN/ISSBN:
0090-5364
DOI:
10.1214/17-AOS1679
发表日期:
2018
页码:
3539-3568
关键词:
subset-selection M-ESTIMATORS maximization RECOVERY bounds
摘要:
We connect high-dimensional subset selection and submodular maximization. Our results extend the work of Das and Kempe [In ICML (2011) 1057-1064] from the setting of linear regression to arbitrary objective functions. For greedy feature selection, this connection allows us to obtain strong multiplicative performance bounds on several methods without statistical modeling assumptions. We also derive recovery guarantees of this form under standard assumptions. Our work shows that greedy algorithms perform within a constant factor from the best possible subset-selection solution for a broad class of general objective functions. Our methods allow a direct control over the number of obtained features as opposed to regularization parameters that only implicitly control sparsity. Our proof technique uses the concept of weak submodularity initially defined by Das and Kempe. We draw a connection between convex analysis and submodular set function theory which may be of independent interest for other statistical learning applications that have combinatorial structure.