论文标题

基于能量的模型的自我适应噪声对抗性估计

Self-Adapting Noise-Contrastive Estimation for Energy-Based Models

论文作者

Xu, Nathaniel

论文摘要

具有噪声对抗性估计(NCE)的基于能量能量的模型(EBM)在理论上是可行的,但实际上具有挑战性。有效的学习需要噪声分布与目标分布大致相似,尤其是在高维域中。以前的工作探索了将噪声分布建模为单独的生成模型,然后同时将此噪声模型与EBM一起训练。尽管此方法允许更有效的噪声对抗性估计,但它以额外的内存和训练复杂性为代价。取而代之的是,本文提出了一种自我适应性NCE算法,该算法将EBM沿其训练轨迹的静态实例作为噪声分布。在训练过程中,这些静态实例逐渐收敛到目标分布,从而规避同时训练辅助噪声模型的需求。此外,我们在Bregman Diverence的框架中表达了这种自我适应性的NCE算法,并表明它是EBM的最大似然学习的概括。在一系列噪声更新间隔中评估了我们的算法的性能,实验结果表明,较短的更新间隔有利于更高的合成质量。

Training energy-based models (EBMs) with noise-contrastive estimation (NCE) is theoretically feasible but practically challenging. Effective learning requires the noise distribution to be approximately similar to the target distribution, especially in high-dimensional domains. Previous works have explored modelling the noise distribution as a separate generative model, and then concurrently training this noise model with the EBM. While this method allows for more effective noise-contrastive estimation, it comes at the cost of extra memory and training complexity. Instead, this thesis proposes a self-adapting NCE algorithm which uses static instances of the EBM along its training trajectory as the noise distribution. During training, these static instances progressively converge to the target distribution, thereby circumventing the need to simultaneously train an auxiliary noise model. Moreover, we express this self-adapting NCE algorithm in the framework of Bregman divergences and show that it is a generalization of maximum likelihood learning for EBMs. The performance of our algorithm is evaluated across a range of noise update intervals, and experimental results show that shorter update intervals are conducive to higher synthesis quality.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源