论文标题
随着任意延迟的预测在线学习
Projection-free Online Learning with Arbitrary Delays
论文作者
论文摘要
通过较便宜的计算(例如线性优化(LO))避开了预测操作的无预测在线学习,由于其在处理具有复杂约束的高维问题方面的效率,最近引起了人们的兴趣。但是,先前的研究假定任何查询梯度都会立即揭示,这可能不会在实践中限制并限制其应用。为了解决这一限制,我们将在线Frank-Wolfe(OFW)算法和在线平滑投影(OSPF)算法分别概括为基于最先进的LO预测在线算法,用于非平滑和平滑功能,以延迟设置的延迟设置,可以通过任意巡回赛延迟设置。具体而言,我们广义OFW的主要思想是在收到任何延迟梯度后执行与原始OFW类似的更新,并对每回合做出最新决定。此外,OSPF的基本变化是替换最初在每个更新中使用的查询梯度总和,并使用可用梯度的总和。尽管它们很简单,但我们的小说分析表明,在相对较大的延迟下,概括的OFW和OSPF分别在非删除环境中享有与OFW和OSPF相同的遗憾。
Projection-free online learning, which eschews the projection operation via less expensive computations such as linear optimization (LO), has received much interest recently due to its efficiency in handling high-dimensional problems with complex constraints. However, previous studies assume that any queried gradient is revealed immediately, which may not hold in practice and limits their applications. To address this limitation, we generalize the online Frank-Wolfe (OFW) algorithm and the online smooth projection-free (OSPF) algorithm, which are state-of-the-art LO-based projection-free online algorithms for non-smooth and smooth functions respectively, into a delayed setting where queried gradients can be delayed by arbitrary rounds. Specifically, the main idea of our generalized OFW is to perform an update similar to the original OFW after receiving any delayed gradient, and play the latest decision for each round. Moreover, the essential change on OSPF is to replace the sum of queried gradients, which is originally utilized in each update, with the sum of available gradients. Despite their simplicities, our novel analysis shows that under a relatively large amount of delay, the generalized OFW and OSPF enjoy the same regret bound as OFW and OSPF in the non-delayed setting, respectively.