FAST LEARNING RATE OF MULTIPLE KERNEL LEARNING: TRADE-OFF BETWEEN SPARSITY AND SMOOTHNESS
成果类型:
Article
署名作者:
Suzuki, Taiji; Sugiyama, Masashi
署名单位:
University of Tokyo; Institute of Science Tokyo; Tokyo Institute of Technology
刊物名称:
ANNALS OF STATISTICS
ISSN/ISSBN:
0090-5364
DOI:
10.1214/13-AOS1095
发表日期:
2013
页码:
1381-1405
关键词:
group lasso
摘要:
We investigate the learning rate of multiple kernel learning (MKL) with l(1) and elastic-net regularizations. The elastic-net regularization is a composition of an l(1)-regularizer for inducing the sparsity and an l(2)-regularizer for controlling the smoothness. We focus on a sparse setting where the total number of kernels is large, but the number of nonzero components of the ground truth is relatively small, and show sharper convergence rates than the learning rates have ever shown for both l(1) and elastic-net regularizations. Our analysis reveals some relations between the choice of a regularization function and the performance. If the ground truth is smooth, we show a faster convergence rate for the elastic-net regularization with less conditions than l(1)-regularization; otherwise, a faster convergence rate for the l(1)-regularization is shown.
来源URL: