论文标题
转向:神经odes的简单时间正则化
STEER: Simple Temporal Regularization For Neural ODEs
论文作者
论文摘要
训练神经普通微分方程(ODE)通常在计算上很昂贵。实际上,计算此类模型的正向通过涉及解决一种在训练过程中可以任意复杂的颂歌。最近的作品表明,将颂歌的动力定于正规,可以部分缓解这种情况。在本文中,我们提出了一种新的正则化技术:在训练过程中随机抽样颂歌的结束时间。提出的正则化易于实施,在开销中可以忽略不计,并且在各种任务中有效。此外,该技术与提出的其他几种od动力学的方法是正交的,因此可以与它们结合使用。我们通过实验对归一化流,时间序列模型和图像识别的实验表明,所提出的正则化可以显着减少训练时间,甚至可以改善基线模型的性能。
Training Neural Ordinary Differential Equations (ODEs) is often computationally expensive. Indeed, computing the forward pass of such models involves solving an ODE which can become arbitrarily complex during training. Recent works have shown that regularizing the dynamics of the ODE can partially alleviate this. In this paper we propose a new regularization technique: randomly sampling the end time of the ODE during training. The proposed regularization is simple to implement, has negligible overhead and is effective across a wide variety of tasks. Further, the technique is orthogonal to several other methods proposed to regularize the dynamics of ODEs and as such can be used in conjunction with them. We show through experiments on normalizing flows, time series models and image recognition that the proposed regularization can significantly decrease training time and even improve performance over baseline models.