论文标题
凸和光滑功能的动态遗憾
Dynamic Regret of Convex and Smooth Functions
论文作者
论文摘要
我们在非平稳环境中调查在线凸优化,并选择动态遗憾作为性能度量,定义为在线算法产生的累积损失与任何可行比较器序列的累积损失之间的差异。令$ t $为时间范围,$ p_t $是基本上反映环境非平稳性的路径长度,最新的动态遗憾是$ \ nathcal {o}(\ sqrt {t(t(1+p_t)}))$。尽管事实证明,这种结合是最小的凸函数,但在本文中,我们证明了通过利用平滑度条件进一步增强动态遗憾。具体而言,我们提出了能够利用平滑度的新颖的在线算法,并通过问题依赖性数量来替代对动态遗憾的依赖性:损失函数梯度的变化,比较器序列的累积损失以及前两个项的最小值。这些数量最多是$ \ MATHCAL {O}(T)$,而在良性环境中可能要小得多。因此,我们的结果适应问题的固有难度,因为对于容易问题,界限比现有结果更紧,同时在最坏情况下保证了相同的速率。
We investigate online convex optimization in non-stationary environments and choose the dynamic regret as the performance measure, defined as the difference between cumulative loss incurred by the online algorithm and that of any feasible comparator sequence. Let $T$ be the time horizon and $P_T$ be the path-length that essentially reflects the non-stationarity of environments, the state-of-the-art dynamic regret is $\mathcal{O}(\sqrt{T(1+P_T)})$. Although this bound is proved to be minimax optimal for convex functions, in this paper, we demonstrate that it is possible to further enhance the dynamic regret by exploiting the smoothness condition. Specifically, we propose novel online algorithms that are capable of leveraging smoothness and replace the dependence on $T$ in the dynamic regret by problem-dependent quantities: the variation in gradients of loss functions, the cumulative loss of the comparator sequence, and the minimum of the previous two terms. These quantities are at most $\mathcal{O}(T)$ while could be much smaller in benign environments. Therefore, our results are adaptive to the intrinsic difficulty of the problem, since the bounds are tighter than existing results for easy problems and meanwhile guarantee the same rate in the worst case.