论文标题

DropMessage:统一图形神经网络的随机掉落

DropMessage: Unifying Random Dropping for Graph Neural Networks

论文作者

Fang, Taoran, Xiao, Zhiqing, Wang, Chunping, Xu, Jiarong, Yang, Xuan, Yang, Yang

论文摘要

图神经网络(GNN)是图表表示学习的强大工具。尽管它们的发展迅速,但GNN还是面临一些挑战,例如过度拟合,过度锻炼和非舒适性。先前的工作表明,可以通过随机删除方法来缓解这些问题,从而通过随机掩盖输入的一部分将增强数据集成到模型中。但是,在GNN上随机下降的一些开放问题仍有待解决。首先,找到适合所有考虑不同数据集和模型差异的通用方法是一项挑战。其次,引入GNN的增强数据导致参数的不完整覆盖范围和不稳定的培训过程。第三,没有关于随机删除方法在GNN的有效性的理论分析。在本文中,我们提出了一种称为DropMessage的新型随机掉落方法,该方法在消息通讯过程中直接在传播消息上执行删除操作。更重要的是,我们发现DropMessage为大多数现有的随机掉落方法提供了一个统一的框架,基于我们对其有效性进行理论分析。此外,我们详细阐述了DropMessage的优势:它通过减少样本方差来稳定训练过程;从信息理论的角度来看,它使信息多样性保持不变,使其成为其他方法的理论上界。为了评估我们提出的方法,我们进行了实验,旨在在五个公共数据集和两个具有各种骨干模型的工业数据集上进行多个任务。实验结果表明,DropMessage具有有效性和概括的优势,并且可以大大减轻上述问题。

Graph Neural Networks (GNNs) are powerful tools for graph representation learning. Despite their rapid development, GNNs also face some challenges, such as over-fitting, over-smoothing, and non-robustness. Previous works indicate that these problems can be alleviated by random dropping methods, which integrate augmented data into models by randomly masking parts of the input. However, some open problems of random dropping on GNNs remain to be solved. First, it is challenging to find a universal method that are suitable for all cases considering the divergence of different datasets and models. Second, augmented data introduced to GNNs causes the incomplete coverage of parameters and unstable training process. Third, there is no theoretical analysis on the effectiveness of random dropping methods on GNNs. In this paper, we propose a novel random dropping method called DropMessage, which performs dropping operations directly on the propagated messages during the message-passing process. More importantly, we find that DropMessage provides a unified framework for most existing random dropping methods, based on which we give theoretical analysis of their effectiveness. Furthermore, we elaborate the superiority of DropMessage: it stabilizes the training process by reducing sample variance; it keeps information diversity from the perspective of information theory, enabling it become a theoretical upper bound of other methods. To evaluate our proposed method, we conduct experiments that aims for multiple tasks on five public datasets and two industrial datasets with various backbone models. The experimental results show that DropMessage has the advantages of both effectiveness and generalization, and can significantly alleviate the problems mentioned above.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源