论文标题

limeout:一种合奏的方法来改善过程公平性

LimeOut: An Ensemble Approach To Improve Process Fairness

论文作者

Bhargava, Vaishnavi, Couceiro, Miguel, Napoli, Amedeo

论文摘要

人工智能和机器学习越来越多地存在于人类生活的几个方面,尤其是那些处理决策的人。这些算法中的许多决定都是在没有人类监督的情况下以及不透明的决策过程做出的。这引起了人们对这些过程对某些社会群体的潜在偏见的担忧,这可能需要不公平的结果,并可能侵犯人权行为。处理这种偏见的模型是维持公众信任的主要问题之一。 在本文中,我们解决了过程或程序公平的问题。更确切地说,我们考虑通过减少对敏感特征的依赖,同时提高(或至少保持其准确性),使分类器更加公平。为了实现这两者,我们从基于神经的方法中的“辍学”技术中汲取灵感,并提出一个依靠“功能辍学”来解决过程公平的框架。我们利用“石灰解释”来评估分类器的公平性并确定要删除的敏感功能。这会产生一个分类器池(通过特征辍学),其合奏在经验上被证明较少依赖敏感特征,并且对准确性的改善或没有影响。

Artificial Intelligence and Machine Learning are becoming increasingly present in several aspects of human life, especially, those dealing with decision making. Many of these algorithmic decisions are taken without human supervision and through decision making processes that are not transparent. This raises concerns regarding the potential bias of these processes towards certain groups of society, which may entail unfair results and, possibly, violations of human rights. Dealing with such biased models is one of the major concerns to maintain the public trust. In this paper, we address the question of process or procedural fairness. More precisely, we consider the problem of making classifiers fairer by reducing their dependence on sensitive features while increasing (or, at least, maintaining) their accuracy. To achieve both, we draw inspiration from "dropout" techniques in neural based approaches, and propose a framework that relies on "feature drop-out" to tackle process fairness. We make use of "LIME Explanations" to assess a classifier's fairness and to determine the sensitive features to remove. This produces a pool of classifiers (through feature dropout) whose ensemble is shown empirically to be less dependent on sensitive features, and with improved or no impact on accuracy.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源