论文标题

深度学习培训程序增加

Deep Learning Training Procedure Augmentations

论文作者

Simionescu, Cristian

论文摘要

深度学习的最新进展大大提高了各种任务的绩效,例如对象检测,图像细分,情感分析。直到最近,大多数研究方向的重点一直在击败最先进的结果上。这已经实现了越来越大的模型和技术的利用,这些模型和技术有助于培训程序从给定数据集中提取更多的预测能力。尽管这带来了很好的结果,但其中许多具有现实世界的应用,但深度学习的其他相关方面仍然被忽略和未知。在这项工作中,我们将介绍几种新颖的深度学习训练技术,尽管能够提供显着的性能增长,但他们还揭示了有关收敛速度,优化景观平滑度和对抗性鲁棒性的几个有趣的分析结果。这项工作中介绍的方法如下: $ \ bullet $完美订购近似;广义模型不可知的课程学习方法。结果表明,该技术在改善训练时间方面的有效性,并为深网的培训过程提供了一些新的见解。 $ \ bullet $级联总和增强;通过利用更平稳的优化景观来利用更多数据点进行线性插值的混合延伸。这可以用于计算机视觉任务,以提高预测性能并提高被动模型的鲁棒性。

Recent advances in Deep Learning have greatly improved performance on various tasks such as object detection, image segmentation, sentiment analysis. The focus of most research directions up until very recently has been on beating state-of-the-art results. This has materialized in the utilization of bigger and bigger models and techniques which help the training procedure to extract more predictive power out of a given dataset. While this has lead to great results, many of which with real-world applications, other relevant aspects of deep learning have remained neglected and unknown. In this work, we will present several novel deep learning training techniques which, while capable of offering significant performance gains they also reveal several interesting analysis results regarding convergence speed, optimization landscape smoothness, and adversarial robustness. The methods presented in this work are the following: $\bullet$ Perfect Ordering Approximation; a generalized model agnostic curriculum learning approach. The results show the effectiveness of the technique for improving training time as well as offer some new insight into the training process of deep networks. $\bullet$ Cascading Sum Augmentation; an extension of mixup capable of utilizing more data points for linear interpolation by leveraging a smoother optimization landscape. This can be used for computer vision tasks in order to improve both prediction performance as well as improve passive model robustness.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源