论文标题
拥抱歧义:使用上下文同义词知识改善面向相似性的任务
Embracing Ambiguity: Improving Similarity-oriented Tasks with Contextual Synonym Knowledge
论文作者
论文摘要
上下文同义词知识对于那些面向相似性的任务至关重要,其核心挑战在于在其上下文中捕获实体之间的语义相似性,例如实体链接和实体匹配。但是,由于其预训练的目标(例如蒙版语言建模(MLM))的固有局限性,大多数预训练的语言模型(PLM)缺乏同义知识。将同义词注入PLM的现有作品通常遇到两个严重的问题:(i)忽略同义词的歧义,以及(ii)破坏对原始PLM的语义理解,这是由同义词的确切语义相似性与从原始语料库中学到的广泛概念相关性之间的不一致引起的。为了解决这些问题,我们提出了PICSO,这是一个灵活的框架,该框架支持从多个域中通过新颖的实体感知的适配器注入上下文同义词知识,该知识的重点是上下文中实体(同义词)的语义。同时,PICSO将同义知识存储在适配器结构的其他参数中,从而阻止其破坏对原始PLM的语义理解。广泛的实验表明,PICSO在四个不同的面向相似性的任务上可以极大地超过原始PLM和其他知识和同义词注入模型。此外,对胶水进行的实验证明,PICSO还有益于一般的自然语言理解任务。代码和数据将是公开的。
Contextual synonym knowledge is crucial for those similarity-oriented tasks whose core challenge lies in capturing semantic similarity between entities in their contexts, such as entity linking and entity matching. However, most Pre-trained Language Models (PLMs) lack synonym knowledge due to inherent limitations of their pre-training objectives such as masked language modeling (MLM). Existing works which inject synonym knowledge into PLMs often suffer from two severe problems: (i) Neglecting the ambiguity of synonyms, and (ii) Undermining semantic understanding of original PLMs, which is caused by inconsistency between the exact semantic similarity of the synonyms and the broad conceptual relevance learned from the original corpus. To address these issues, we propose PICSO, a flexible framework that supports the injection of contextual synonym knowledge from multiple domains into PLMs via a novel entity-aware Adapter which focuses on the semantics of the entities (synonyms) in the contexts. Meanwhile, PICSO stores the synonym knowledge in additional parameters of the Adapter structure, which prevents it from corrupting the semantic understanding of the original PLM. Extensive experiments demonstrate that PICSO can dramatically outperform the original PLMs and the other knowledge and synonym injection models on four different similarity-oriented tasks. In addition, experiments on GLUE prove that PICSO also benefits general natural language understanding tasks. Codes and data will be public.