论文标题
几乎没有图像级弱注释增强的语义分割
Few-Shot Semantic Segmentation Augmented with Image-Level Weak Annotations
论文作者
论文摘要
尽管深层神经网络在语义细分任务中取得了巨大进展,但基于神经网络的传统方法通常会缺少大量像素级注释。几乎没有一个像素级注释的示例仅解决了几个镜头语义细分的最新进展。但是,这些少数几种方法无法轻易应用于多路或弱注释设置。在本文中,我们将少量分段范式推向了一个场景,在该场景中,可以使用图像级注释来帮助一些像素级注释的训练过程。我们的关键思想是通过从图像级标记的数据中融合知识来学习更好的类原型表示。具体来说,我们提出了一个称为PAIA的新框架,以通过集成图像级注释来学习公制空间中的类原型表示。此外,通过考虑伪面罩的不确定性,蒸馏的软遮罩平均合并策略旨在处理图像级注释中的干扰。两个数据集的广泛经验结果表明PAIA的表现出色。
Despite the great progress made by deep neural networks in the semantic segmentation task, traditional neural-networkbased methods typically suffer from a shortage of large amounts of pixel-level annotations. Recent progress in fewshot semantic segmentation tackles the issue by only a few pixel-level annotated examples. However, these few-shot approaches cannot easily be applied to multi-way or weak annotation settings. In this paper, we advance the few-shot segmentation paradigm towards a scenario where image-level annotations are available to help the training process of a few pixel-level annotations. Our key idea is to learn a better prototype representation of the class by fusing the knowledge from the image-level labeled data. Specifically, we propose a new framework, called PAIA, to learn the class prototype representation in a metric space by integrating image-level annotations. Furthermore, by considering the uncertainty of pseudo-masks, a distilled soft masked average pooling strategy is designed to handle distractions in image-level annotations. Extensive empirical results on two datasets show superior performance of PAIA.