论文标题

不公平的发现和预防几次回归

Unfairness Discovery and Prevention For Few-Shot Regression

论文作者

Zhao, Chen, Chen, Feng

论文摘要

我们研究了对历史数据中对歧视(或偏见)敏感的几乎没有射击的元学习模型的公平性。基于偏见数据训练的机器学习模型往往会对少数群体的用户做出不公平的预测。尽管以前已经研究过此问题,但现有方法主要旨在根据大量培训数据来检测和控制受保护变量(例如种族,性别)对目标预测的依赖性效应。这些方法带有两个主要缺点,(1)缺乏显示所有变量的全球原因可视化; (2)缺乏对看不见的任务的准确性和公平性的概括。在这项工作中,我们首先使用因果贝叶斯知识图发现了与数据的歧视,该图不仅证明了受保护变量对目标的依赖性,而且还表明了所有变量之间的因果效应。接下来,我们根据风险差异开发一种新型算法,以量化图中每个受保护变量的歧视性影响。此外,为了保护预测免受不公平的影响,提出了元学习中快速适应的偏置对照方法,这有效地减轻了每个任务的统计差异,因此它确保了基于偏见和少量数据样本的预测的独立性独立性。与现有的元学习模型不同,通过利用(联合国)受保护组之间的平均差异来解决回归问题,从而有效地降低了任务的群体不公平性。通过对综合和现实世界数据集的广泛实验,我们证明了我们提出的不公平性发现和预防方法有效地检测歧视并减轻模型输出的偏见,并推广出有限量的培训样本的准确性和公平性,以使其准确性和公平性,以免看到不见了的任务。

We study fairness in supervised few-shot meta-learning models that are sensitive to discrimination (or bias) in historical data. A machine learning model trained based on biased data tends to make unfair predictions for users from minority groups. Although this problem has been studied before, existing methods mainly aim to detect and control the dependency effect of the protected variables (e.g. race, gender) on target prediction based on a large amount of training data. These approaches carry two major drawbacks that (1) lacking showing a global cause-effect visualization for all variables; (2) lacking generalization of both accuracy and fairness to unseen tasks. In this work, we first discover discrimination from data using a causal Bayesian knowledge graph which not only demonstrates the dependency of the protected variable on target but also indicates causal effects between all variables. Next, we develop a novel algorithm based on risk difference in order to quantify the discriminatory influence for each protected variable in the graph. Furthermore, to protect prediction from unfairness, a fast-adapted bias-control approach in meta-learning is proposed, which efficiently mitigates statistical disparity for each task and it thus ensures independence of protected attributes on predictions based on biased and few-shot data samples. Distinct from existing meta-learning models, group unfairness of tasks are efficiently reduced by leveraging the mean difference between (un)protected groups for regression problems. Through extensive experiments on both synthetic and real-world data sets, we demonstrate that our proposed unfairness discovery and prevention approaches efficiently detect discrimination and mitigate biases on model output as well as generalize both accuracy and fairness to unseen tasks with a limited amount of training samples.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源