论文标题
通过随机注意机制可解释且可推广的图形学习
Interpretable and Generalizable Graph Learning via Stochastic Attention Mechanism
论文作者
论文摘要
可以解释的图表学习是需要的,因为许多科学应用都取决于学习模型来从图形结构数据中收集见解。先前的工作主要集中在使用事后方法来解释前训练的模型(尤其是图形神经网络)。他们反对固有的可解释模型,因为这些模型的良好解释性通常是以其预测准确性为代价。但是,这些事后方法通常无法提供稳定的解释,并且可能提取与任务相关的特征。在这项工作中,我们通过提出图形随机关注(GSAT)来解决这些问题。 GSAT从信息瓶颈原理中得出,将随机性注入了注意力权重,以阻止任务 - iRrelevant图组件的信息,同时学习降低随机性的注意力以选择与任务相关的子图以进行解释。事实证明,所选的子图不包含与某些假设下的任务相关的模式。八个数据集上的大量实验表明,GSAT在解释AUC中的最高最高为20%$ \ uparrow $和5%$ \ uparrow $的预测准确性。我们的代码可在https://github.com/graph-com/gsat上找到。
Interpretable graph learning is in need as many scientific applications depend on learning models to collect insights from graph-structured data. Previous works mostly focused on using post-hoc approaches to interpret pre-trained models (graph neural networks in particular). They argue against inherently interpretable models because the good interpretability of these models is often at the cost of their prediction accuracy. However, those post-hoc methods often fail to provide stable interpretation and may extract features that are spuriously correlated with the task. In this work, we address these issues by proposing Graph Stochastic Attention (GSAT). Derived from the information bottleneck principle, GSAT injects stochasticity to the attention weights to block the information from task-irrelevant graph components while learning stochasticity-reduced attention to select task-relevant subgraphs for interpretation. The selected subgraphs provably do not contain patterns that are spuriously correlated with the task under some assumptions. Extensive experiments on eight datasets show that GSAT outperforms the state-of-the-art methods by up to 20%$\uparrow$ in interpretation AUC and 5%$\uparrow$ in prediction accuracy. Our code is available at https://github.com/Graph-COM/GSAT.