论文标题
带有课堂对比度学习的转导剪辑
Transductive CLIP with Class-Conditional Contrastive Learning
论文作者
论文摘要
受视力语言预培训模型的显着零射门概括能力的启发,我们试图利用剪辑模型的监督来减轻数据标记的负担。但是,这种监督不可避免地包含标签噪声,从而大大降低了分类模型的判别能力。在这项工作中,我们提出了TransDuctive剪辑,这是一个学习具有从头开始的嘈杂标签的分类网络的新型框架。首先,提出了一种类似的对比学习机制,以减轻对伪标签的依赖并提高对嘈杂标签的耐受性。其次,合奏标签被用作伪标签更新策略,以稳定具有嘈杂标签的深神经网络的培训。该框架可以通过结合两种技术有效地从夹子模型中降低嘈杂标签的影响。多个基准数据集的实验证明了对其他最新方法的实质性改进。
Inspired by the remarkable zero-shot generalization capacity of vision-language pre-trained model, we seek to leverage the supervision from CLIP model to alleviate the burden of data labeling. However, such supervision inevitably contains the label noise, which significantly degrades the discriminative power of the classification model. In this work, we propose Transductive CLIP, a novel framework for learning a classification network with noisy labels from scratch. Firstly, a class-conditional contrastive learning mechanism is proposed to mitigate the reliance on pseudo labels and boost the tolerance to noisy labels. Secondly, ensemble labels is adopted as a pseudo label updating strategy to stabilize the training of deep neural networks with noisy labels. This framework can reduce the impact of noisy labels from CLIP model effectively by combining both techniques. Experiments on multiple benchmark datasets demonstrate the substantial improvements over other state-of-the-art methods.