论文标题
从未知的偏见中脱颖而出的NLU模型
Towards Debiasing NLU Models from Unknown Biases
论文作者
论文摘要
NLU模型通常会利用偏见来实现高数据集特定的性能,而无需正确学习预期的任务。最近提出的证词方法被证明可以有效缓解这种趋势。但是,这些方法依赖于应知道偏差类型的主要假设,该假设应已知A-Priori,这将其应用限制在许多NLU任务和数据集中。在这项工作中,我们通过引入一个自我欺骗的框架来介绍弥合这一差距的第一步,该框架阻止了模型主要利用偏见而无需事先了解偏见。所提出的框架是一般的,并且与现有的偏见方法互补。我们表明,它允许这些现有方法保留挑战数据集的改进(即,旨在暴露模型对偏见的依赖的示例集),而无需专门针对某些偏见。此外,评估表明,应用框架会改善整体鲁棒性。
NLU models often exploit biases to achieve high dataset-specific performance without properly learning the intended task. Recently proposed debiasing methods are shown to be effective in mitigating this tendency. However, these methods rely on a major assumption that the types of bias should be known a-priori, which limits their application to many NLU tasks and datasets. In this work, we present the first step to bridge this gap by introducing a self-debiasing framework that prevents models from mainly utilizing biases without knowing them in advance. The proposed framework is general and complementary to the existing debiasing methods. We show that it allows these existing methods to retain the improvement on the challenge datasets (i.e., sets of examples designed to expose models' reliance on biases) without specifically targeting certain biases. Furthermore, the evaluation suggests that applying the framework results in improved overall robustness.