论文标题

现实世界神经网络的链图解释

A Chain Graph Interpretation of Real-World Neural Networks

论文作者

Shen, Yuesong, Cremers, Daniel

论文摘要

在过去的十年中,深入学习研究和应用程序中实现了最先进的结果的繁荣。但是,大多数进步都是从经验上确立的,并且他们的理论分析仍然缺乏。一个主要问题是,我们当前对神经网络(NNS)作为函数近似器的解释太通用了,无法支持深入分析。在本文中,我们通过提出一种替代解释来纠正此问题,该解释将NNS识别为链图(CGS)和Feed-Forwward作为近似推理程序。 CG的解释指定了概率图形模型的丰富理论框架内每个NN组件的性质,而同时仍然足够笼统地覆盖具有任意深度,多支化和多样性激活的真实世界NN,以及包括卷积的结构,包括卷积 /相互发达的层,残留的层,残基和辍学。我们用具体的例子证明,CG解释可以为各种NN技术提供新颖的理论支持和见解,并得出了新的深度学习方法,例如部分倒塌的饲料 - 前向推理的概念。因此,这是一个有前途的框架,可以加深我们对神经网络的理解,并为未来的深度学习研究提供了连贯的理论表述。

The last decade has witnessed a boom of deep learning research and applications achieving state-of-the-art results in various domains. However, most advances have been established empirically, and their theoretical analysis remains lacking. One major issue is that our current interpretation of neural networks (NNs) as function approximators is too generic to support in-depth analysis. In this paper, we remedy this by proposing an alternative interpretation that identifies NNs as chain graphs (CGs) and feed-forward as an approximate inference procedure. The CG interpretation specifies the nature of each NN component within the rich theoretical framework of probabilistic graphical models, while at the same time remains general enough to cover real-world NNs with arbitrary depth, multi-branching and varied activations, as well as common structures including convolution / recurrent layers, residual block and dropout. We demonstrate with concrete examples that the CG interpretation can provide novel theoretical support and insights for various NN techniques, as well as derive new deep learning approaches such as the concept of partially collapsed feed-forward inference. It is thus a promising framework that deepens our understanding of neural networks and provides a coherent theoretical formulation for future deep learning research.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源