论文标题
无监督的精确暹罗跟踪学习
Unsupervised Learning of Accurate Siamese Tracking
论文作者
论文摘要
无监督的学习在各种计算机视觉任务中很受欢迎,包括视觉对象跟踪。但是,事先无监督的跟踪方法在很大程度上依赖于模板搜索对的空间监督,并且仍然无法在很长的时间内跟踪具有较大差异的对象。由于可以通过及时跟踪视频来获得无限的自我划定信号,因此我们通过向前追踪视频来调查暹罗跟踪器的发展。我们提出了一个新颖的无监督跟踪框架,在该框架中,我们可以在分类分支和回归分支上学习时间对应关系。具体而言,要在正向传播过程中传播可靠的模板特征,以便可以在周期中训练跟踪器,我们首先提出一致性传播转换。然后,我们在向后传播过程中的常规周期训练中确定了一个不适的惩罚问题。因此,提出了一个可区分的区域掩码来选择特征,并隐式惩罚跟踪中间帧的错误。此外,由于嘈杂的标签可能会降低培训,因此我们提出了一种面具引导的损失重新加权策略,以根据伪标签的质量来分配动态权重。在广泛的实验中,我们的跟踪器的表现优于无监督方法的大幅度差距,并以大规模数据集(例如TrackingNet和Lasot)的监督方法进行表现。代码可在https://github.com/florinshum/alsta上获得。
Unsupervised learning has been popular in various computer vision tasks, including visual object tracking. However, prior unsupervised tracking approaches rely heavily on spatial supervision from template-search pairs and are still unable to track objects with strong variation over a long time span. As unlimited self-supervision signals can be obtained by tracking a video along a cycle in time, we investigate evolving a Siamese tracker by tracking videos forward-backward. We present a novel unsupervised tracking framework, in which we can learn temporal correspondence both on the classification branch and regression branch. Specifically, to propagate reliable template feature in the forward propagation process so that the tracker can be trained in the cycle, we first propose a consistency propagation transformation. We then identify an ill-posed penalty problem in conventional cycle training in backward propagation process. Thus, a differentiable region mask is proposed to select features as well as to implicitly penalize tracking errors on intermediate frames. Moreover, since noisy labels may degrade training, we propose a mask-guided loss reweighting strategy to assign dynamic weights based on the quality of pseudo labels. In extensive experiments, our tracker outperforms preceding unsupervised methods by a substantial margin, performing on par with supervised methods on large-scale datasets such as TrackingNet and LaSOT. Code is available at https://github.com/FlorinShum/ULAST.