论文标题
非凸线的快速目标和二元性差距收敛强,concave concave min-max问题与PL状态
Fast Objective & Duality Gap Convergence for Non-Convex Strongly-Concave Min-Max Problems with PL Condition
论文作者
论文摘要
本文着重于解决平滑非凸的随机方法强烈concove min-max问题,由于它们在深度学习中的潜在应用(例如,深度AUC最大化,分布在分布方面优化),由于其潜在的应用而受到了越来越多的关注。但是,大多数现有的算法在实践中都很慢,并且它们的分析围绕着汇聚到几乎固定点。我们考虑利用Polyak-lojasiewicz(PL)条件来设计更快的随机算法,并具有更强的收敛保证。尽管PL条件已用于设计许多随机最小化算法,但它们用于非凸线最小值优化的应用仍然很少见。在本文中,我们提出和分析了基于阶段的近端方法的通用框架,并具有许多可嵌入的随机更新。快速收敛是根据原始客观差距和二元性差距建立的。与现有研究相比,(i)我们的分析基于一个新型的Lyapunov函数,该功能由原始物镜差距和正规函数的二元性差距组成,并且(ii)结果更全面,速率提高了,速率更好地依赖于不同假设下的条件数量。我们还进行了深入而非深度学习实验,以验证我们方法的有效性。
This paper focuses on stochastic methods for solving smooth non-convex strongly-concave min-max problems, which have received increasing attention due to their potential applications in deep learning (e.g., deep AUC maximization, distributionally robust optimization). However, most of the existing algorithms are slow in practice, and their analysis revolves around the convergence to a nearly stationary point.We consider leveraging the Polyak-Lojasiewicz (PL) condition to design faster stochastic algorithms with stronger convergence guarantee. Although PL condition has been utilized for designing many stochastic minimization algorithms, their applications for non-convex min-max optimization remain rare. In this paper, we propose and analyze a generic framework of proximal stage-based method with many well-known stochastic updates embeddable. Fast convergence is established in terms of both the primal objective gap and the duality gap. Compared with existing studies, (i) our analysis is based on a novel Lyapunov function consisting of the primal objective gap and the duality gap of a regularized function, and (ii) the results are more comprehensive with improved rates that have better dependence on the condition number under different assumptions. We also conduct deep and non-deep learning experiments to verify the effectiveness of our methods.