论文标题

强化基于学习的N- ARY跨句子关系提取

Reinforcement Learning-based N-ary Cross-Sentence Relation Extraction

论文作者

Yuan, Chenhan, Rossi, Ryan, Katz, Andrew, Eldardiry, Hoda

论文摘要

基于遥远监督的n- ary跨句子关系提取的模型假定,提及n实体的连续句子描述了这些n实体的关系。但是,一方面,此假设引入了嘈杂的标记数据并损害模型的性能。另一方面,一些非连续句子也描述了一个关系,这些句子不能在此假设下标记。在本文中,我们通过较弱的遥远监督假设来解决这一问题,以解决第二个问题,并提出了一个新的句子分布估算器模型来解决第一个问题。该估计器选择正确标记的句子以减轻嘈杂数据的效果是两级代理增强学习模型。此外,提出了一种新型的通用关系提取器,具有混合的注意机制和PCNN方法,以便可以将其部署在任何任务中,包括连续和非连续句子。实验表明,与基线模型相比,提出的模型可以减少嘈杂数据的影响,并在一般N- ARY跨句子关系提取任务上实现更好的性能。

The models of n-ary cross sentence relation extraction based on distant supervision assume that consecutive sentences mentioning n entities describe the relation of these n entities. However, on one hand, this assumption introduces noisy labeled data and harms the models' performance. On the other hand, some non-consecutive sentences also describe one relation and these sentences cannot be labeled under this assumption. In this paper, we relax this strong assumption by a weaker distant supervision assumption to address the second issue and propose a novel sentence distribution estimator model to address the first problem. This estimator selects correctly labeled sentences to alleviate the effect of noisy data is a two-level agent reinforcement learning model. In addition, a novel universal relation extractor with a hybrid approach of attention mechanism and PCNN is proposed such that it can be deployed in any tasks, including consecutive and nonconsecutive sentences. Experiments demonstrate that the proposed model can reduce the impact of noisy data and achieve better performance on general n-ary cross sentence relation extraction task compared to baseline models.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源