论文标题

通过平衡对抗训练提高稳健的公平性

Improving Robust Fairness via Balance Adversarial Training

论文作者

Sun, Chunyu, Xu, Chenye, Yao, Chengyuan, Liang, Siyuan, Wu, Yichao, Liang, Ding, Liu, XiangLong, Liu, Aishan

论文摘要

对抗训练(AT)方法有效地防止对抗性攻击,但它们引入了不同类别之间的准确性和鲁棒性的严重差异(称为鲁棒公平问题)。以前建议的公平健壮的学习(FRL)可以适应不同的类别以提高公平性。但是,表现良好的班级的表现降低了,导致性能下降。在本文中,我们在对抗训练期间观察到了两个不公平现象:在产生每个类别的对抗性示例(源级公平)和产生对抗性示例(目标级公平)时产生对抗性示例的不​​同困难。从观察结果中,我们提出平衡对抗训练(BAT)来解决强大的公平问题。关于来源级的公平性,我们调整了每个班级的攻击强度和困难,以在决策边界附近生成样本,以便更容易,更公平的模型学习;考虑到目标级公平,通过引入统一的分布约束,我们鼓励每个班级的对抗性示例生成过程都有公平的趋势。在多个数据集(CIFAR-10,CIFAR-100和IMAGENETTE)上进行的广泛实验表明,我们的方法可以显着超过其他基准,以减轻强大的公平性问题(在最差的类精确度上+5-10 \%)

Adversarial training (AT) methods are effective against adversarial attacks, yet they introduce severe disparity of accuracy and robustness between different classes, known as the robust fairness problem. Previously proposed Fair Robust Learning (FRL) adaptively reweights different classes to improve fairness. However, the performance of the better-performed classes decreases, leading to a strong performance drop. In this paper, we observed two unfair phenomena during adversarial training: different difficulties in generating adversarial examples from each class (source-class fairness) and disparate target class tendencies when generating adversarial examples (target-class fairness). From the observations, we propose Balance Adversarial Training (BAT) to address the robust fairness problem. Regarding source-class fairness, we adjust the attack strength and difficulties of each class to generate samples near the decision boundary for easier and fairer model learning; considering target-class fairness, by introducing a uniform distribution constraint, we encourage the adversarial example generation process for each class with a fair tendency. Extensive experiments conducted on multiple datasets (CIFAR-10, CIFAR-100, and ImageNette) demonstrate that our method can significantly outperform other baselines in mitigating the robust fairness problem (+5-10\% on the worst class accuracy)

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源