Universal Online Convex Optimization With Minimax Optimal Second-Order Dynamic Regret

成果类型:
Article
署名作者:
Gokcesu, Hakan; Kozat, Suleyman Serdar
署名单位:
Ihsan Dogramaci Bilkent University; Turkcell Turkey
刊物名称:
IEEE TRANSACTIONS ON AUTOMATIC CONTROL
ISSN/ISSBN:
0018-9286
DOI:
10.1109/TAC.2024.3381866
发表日期:
2024
页码:
3865-3880
关键词:
Heuristic algorithms optimization Convex functions vectors programming Complexity theory Upper bound Convex Optimization dynamic regret Gradient descent minimax optimal online learning universal guarantee
摘要:
We introduce an online convex optimization algorithm which utilizes projected subgradient descent with optimal adaptive learning rates. We provide a second-order minimax-optimal dynamic regret guarantee (i.e., dependent on the sum of squared subgradient norms) for a sequence of general convex functions, which may not have strong-convexity, smoothness, exp-concavity or even proper Lipschitz-continuity. The guarantee holds against any comparator sequence with bounded path variation (i.e., the sum of distances between successive decisions). We generate a lower bound of the worst-case second-order dynamic regret by incorporating actual subgradient norms. We show that this bound matches our regret guarantee within a constant, making our algorithm minimax-optimal. We also derive the extension for learning in each decision coordinate-block individually. We demonstrate how to best preserve our regret guarantee in a truly online manner, when the bound on the path variation of the comparator grows in time or the feedback regarding that arrives partially as time goes on. We then build on our algorithm to eliminate the need of any knowledge on the comparator path variation, and provide minimax-optimal second-order regret guarantees with no a priori information. Our approach can compete against all comparator sequences simultaneously (i.e., universally) in a minimax-optimal manner, i.e., each regret guarantee depends on the respective comparator path variation. We discuss modifications that address complexity reductions for time, computation and memory. We further improve our results by making the regret guarantees also dependent on the comparator sets' minimum-bounding diameters, in addition to the respective path variations.