论文标题
用硬性样本进行对比学习
Contrastive Learning with Hard Negative Samples
论文作者
论文摘要
您如何为对比度学习采样良好的负面例子?我们认为,与指标学习一样,对形式的对比度学习受益于硬否定样本(即难以与锚点区分开的点)。使用硬否负面的主要挑战是,对比方法必须保持不受监督,这使得采用使用真实相似性信息的现有负面抽样策略是不可避免的。作为回应,我们开发了一种新的无监督抽样方法,用于选择用户可以控制硬度的硬性负面样本。此抽样的限制案例导致一个表示将每个类紧密簇的表示形式,并将不同的类推动到尽可能远的地方。所提出的方法改善了跨多种模式的下游性能,只需要几行代码即可实现,并且没有介绍计算开销。
How can you sample good negative examples for contrastive learning? We argue that, as with metric learning, contrastive learning of representations benefits from hard negative samples (i.e., points that are difficult to distinguish from an anchor point). The key challenge toward using hard negatives is that contrastive methods must remain unsupervised, making it infeasible to adopt existing negative sampling strategies that use true similarity information. In response, we develop a new family of unsupervised sampling methods for selecting hard negative samples where the user can control the hardness. A limiting case of this sampling results in a representation that tightly clusters each class, and pushes different classes as far apart as possible. The proposed method improves downstream performance across multiple modalities, requires only few additional lines of code to implement, and introduces no computational overhead.