论文标题

深度学习对抗性鲁棒性的机遇和挑战:一项调查

Opportunities and Challenges in Deep Learning Adversarial Robustness: A Survey

论文作者

Silva, Samuel Henrique, Najafirad, Peyman

论文摘要

当我们寻求在虚拟和受控域以外部署机器学习模型时,不仅要分析它在大多数时候起作用的准确性或事实至关重要,而且如果这种模型确实是强大且可靠的,则至关重要。本文研究了实施对手训练的算法的策略,以确保机器学习算法的安全性。我们提供了一种分类学来对对抗性攻击和防御进行分类,在最小 - 最大设置中制定强大的优化问题,并将其分为3个子类别,即:对抗(RE)训练,正则化方法和认证的防御。我们将对抗性示例产生,对抗(RE)训练的防御机制的最新和重要结果作为对扰动的主要防御。我们还调查了添加正规化术语的莫托德人,以改变梯度的行为,从而使攻击者更难实现其目标。另外,我们已经调查了通过精确解决优化问题或使用上限或下限的近似值来正式得出鲁棒性证书的方法。此外,我们讨论了大多数最近提出未来研究观点的算法所面临的挑战。

As we seek to deploy machine learning models beyond virtual and controlled domains, it is critical to analyze not only the accuracy or the fact that it works most of the time, but if such a model is truly robust and reliable. This paper studies strategies to implement adversary robustly trained algorithms towards guaranteeing safety in machine learning algorithms. We provide a taxonomy to classify adversarial attacks and defenses, formulate the Robust Optimization problem in a min-max setting and divide it into 3 subcategories, namely: Adversarial (re)Training, Regularization Approach, and Certified Defenses. We survey the most recent and important results in adversarial example generation, defense mechanisms with adversarial (re)Training as their main defense against perturbations. We also survey mothods that add regularization terms that change the behavior of the gradient, making it harder for attackers to achieve their objective. Alternatively, we've surveyed methods which formally derive certificates of robustness by exactly solving the optimization problem or by approximations using upper or lower bounds. In addition, we discuss the challenges faced by most of the recent algorithms presenting future research perspectives.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源