论文标题
线性回归的对抗培训中的精确折衷
Precise Tradeoffs in Adversarial Training for Linear Regression
论文作者
论文摘要
尽管表现突破,但现代学习模型在其输入中非常容易受到小小的对抗性扰动的影响。尽管各种最近的\ emph {对抗训练}方法已经有效地提高了对扰动输入的鲁棒性(稳健的精度),但通常,这种好处伴随着良性输入(标准精度)的准确性降低,从而导致经常竞争的目标之间的折衷。使事情进一步复杂化,最近的经验证据表明,其他各种其他因素(培训数据的规模和质量,模型规模等)都以令人惊讶的方式影响了这一权衡。在本文中,我们对具有高斯特征的线性回归背景下的对抗性训练的作用提供了精确而全面的理解。特别是,无论训练数据的计算能力或大小如何,我们都表征了任何算法可实现的准确性之间的基本权衡。此外,我们精确地表征了当代迷你最大对抗训练方法在高维度中实现的标准/鲁棒精度和相应的权衡,在高维度中,数据点的数量和模型的参数相互成比例地增长。我们的对抗训练算法的理论还促进了对各种因素(训练数据的大小和质量,模型过度化等)如何影响这两个竞争精确度之间的权衡的严格研究。
Despite breakthrough performance, modern learning models are known to be highly vulnerable to small adversarial perturbations in their inputs. While a wide variety of recent \emph{adversarial training} methods have been effective at improving robustness to perturbed inputs (robust accuracy), often this benefit is accompanied by a decrease in accuracy on benign inputs (standard accuracy), leading to a tradeoff between often competing objectives. Complicating matters further, recent empirical evidence suggest that a variety of other factors (size and quality of training data, model size, etc.) affect this tradeoff in somewhat surprising ways. In this paper we provide a precise and comprehensive understanding of the role of adversarial training in the context of linear regression with Gaussian features. In particular, we characterize the fundamental tradeoff between the accuracies achievable by any algorithm regardless of computational power or size of the training data. Furthermore, we precisely characterize the standard/robust accuracy and the corresponding tradeoff achieved by a contemporary mini-max adversarial training approach in a high-dimensional regime where the number of data points and the parameters of the model grow in proportion to each other. Our theory for adversarial training algorithms also facilitates the rigorous study of how a variety of factors (size and quality of training data, model overparametrization etc.) affect the tradeoff between these two competing accuracies.