论文标题
对线性最小二乘问题的深网的无特征力组成培训
Eigendecomposition-Free Training of Deep Networks for Linear Least-Square Problems
论文作者
论文摘要
可以通过求解线性最小二乘问题来解决许多经典的计算机视觉问题,例如基本矩阵计算和姿势估计,从3D到2D对应关系,可以通过找到对应于代表线性系统的矩阵的特征值的特征向量来解决。将其纳入深度学习框架将使我们能够明确编码已知的几何概念,而不是让网络隐含地从数据中学习。但是,在网络中执行特征分类需要区分此操作的能力。在理论上可行的同时,这在实践中的优化过程中引入了数值不稳定性。在本文中,我们引入了一种无特征分类的方法来训练深层网络,其损失取决于与网络预测的矩阵的零特征值相对应的特征向量。我们证明,我们的方法比使用两个通用任务的特征组成的明确差异(包括宽基线立体声),包括宽基线立体声,透视n-n-point问题和椭圆拟合,更强大。从经验上讲,我们的方法具有更好的收敛属性,并产生最先进的结果。
Many classical Computer Vision problems, such as essential matrix computation and pose estimation from 3D to 2D correspondences, can be tackled by solving a linear least-square problem, which can be done by finding the eigenvector corresponding to the smallest, or zero, eigenvalue of a matrix representing a linear system. Incorporating this in deep learning frameworks would allow us to explicitly encode known notions of geometry, instead of having the network implicitly learn them from data. However, performing eigendecomposition within a network requires the ability to differentiate this operation. While theoretically doable, this introduces numerical instability in the optimization process in practice. In this paper, we introduce an eigendecomposition-free approach to training a deep network whose loss depends on the eigenvector corresponding to a zero eigenvalue of a matrix predicted by the network. We demonstrate that our approach is much more robust than explicit differentiation of the eigendecomposition using two general tasks, outlier rejection and denoising, with several practical examples including wide-baseline stereo, the perspective-n-point problem, and ellipse fitting. Empirically, our method has better convergence properties and yields state-of-the-art results.