论文标题
从信息理论的角度重新审视注意力重量作为解释
Revisiting Attention Weights as Explanations from an Information Theoretic Perspective
论文作者
论文摘要
注意机制最近在一系列NLP任务上表现出了令人印象深刻的表现,并且注意力评分通常被用作模型解释性的代理。但是,关于注意力重量是否可以用于识别模型最重要的输入存在争论。我们从信息理论的角度来解决这个问题,通过测量模型输出与隐藏状态之间的相互信息。从广泛的实验中,我们得出以下结论:(i)在保留隐藏状态和模型输出之间的信息(与缩放点产生相比)之间,加性和深度注意机制可能会更好; (ii)消融研究表明,加性的注意力可以积极学会解释其输入隐藏表示的重要性; (iii)当注意力值几乎相同时,使用gumbel-softmax使用温度低于1的Gumbel-Softmax的等级顺序与相互信息(IV)的等级顺序不一致,往往会产生更偏斜的注意力评分分布,因此与软效果相比,并且对于可解释的设计是更好的选择; (v)一些构建块更好地保留了相互信息和注意力权重订单的有序列表之间的相关性(例如,Bilstm编码器的组合和添加剂的注意力)。我们的发现表明,当注意机制与其他模型元素仔细结合时,注意机制确实有可能充当模型解释的捷径。
Attention mechanisms have recently demonstrated impressive performance on a range of NLP tasks, and attention scores are often used as a proxy for model explainability. However, there is a debate on whether attention weights can, in fact, be used to identify the most important inputs to a model. We approach this question from an information theoretic perspective by measuring the mutual information between the model output and the hidden states. From extensive experiments, we draw the following conclusions: (i) Additive and Deep attention mechanisms are likely to be better at preserving the information between the hidden states and the model output (compared to Scaled Dot-product); (ii) ablation studies indicate that Additive attention can actively learn to explain the importance of its input hidden representations; (iii) when attention values are nearly the same, the rank order of attention values is not consistent with the rank order of the mutual information(iv) Using Gumbel-Softmax with a temperature lower than one, tends to produce a more skewed attention score distribution compared to softmax and hence is a better choice for explainable design; (v) some building blocks are better at preserving the correlation between the ordered list of mutual information and attention weights order (for e.g., the combination of BiLSTM encoder and Additive attention). Our findings indicate that attention mechanisms do have the potential to function as a shortcut to model explanations when they are carefully combined with other model elements.