论文标题

通过执行本地和全球紧凑性来改善对抗性的鲁棒性

Improving Adversarial Robustness by Enforcing Local and Global Compactness

论文作者

Bui, Anh, Le, Trung, Zhao, He, Montague, Paul, deVel, Olivier, Abraham, Tamas, Phung, Dinh

论文摘要

深层神经网络容易受到精心摄动的影响,这一事实严重影响了在某些应用领域中深入学习的使用。在许多针对此类攻击的国防模型中,对抗性训练是最成功的方法,它始终抵抗广泛的攻击。在这项工作中,基于先前研究的观察结果,即干净的数据示例的表示及其对抗性示例在深神经网的较高层中变得更加不同,我们建议在深层神经网络的中间层上实现局部/全球紧凑性和聚类假设。我们进行全面的实验,以了解每个组件的隔离行为(即局部/全球紧凑性和聚类假设),并将我们所提出的模型与最新的对抗性训练方法进行比较。实验结果表明,通过我们提出的组件增强对抗性训练可以进一步提高网络的鲁棒性,从而导致更高的不受干扰和对抗性的预测性能。

The fact that deep neural networks are susceptible to crafted perturbations severely impacts the use of deep learning in certain domains of application. Among many developed defense models against such attacks, adversarial training emerges as the most successful method that consistently resists a wide range of attacks. In this work, based on an observation from a previous study that the representations of a clean data example and its adversarial examples become more divergent in higher layers of a deep neural net, we propose the Adversary Divergence Reduction Network which enforces local/global compactness and the clustering assumption over an intermediate layer of a deep neural network. We conduct comprehensive experiments to understand the isolating behavior of each component (i.e., local/global compactness and the clustering assumption) and compare our proposed model with state-of-the-art adversarial training methods. The experimental results demonstrate that augmenting adversarial training with our proposed components can further improve the robustness of the network, leading to higher unperturbed and adversarial predictive performances.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源