论文标题
Fairxgboost:XGBoost中的公平感知分类
FairXGBoost: Fairness-aware Classification in XGBoost
论文作者
论文摘要
高度调节的领域(例如财务)长期以来都赞成使用可扩展,透明,健壮并产生更好性能的机器学习算法。这种算法最突出的例子之一是XGBoost。同时,在这些受监管的领域中建立公平和公正的模型也越来越兴趣,并且已经提出了许多偏见降低算法。但是,这些偏见减少方法中的大多数仅限于特定模型家族,例如逻辑回归或支持向量机模型,因此使建模者艰难地决定从偏见降低算法和可伸缩性,透明度,透明度,诸如XGBOOST等算法的性能之间进行选择。我们的目标是通过提出XGBoost的公平变体来利用两全其美的世界,该变体享有XGBoost的所有优势,同时还匹配了最先进的偏见降低算法的公平水平。此外,提出的解决方案在对原始XGBoost库的更改方面几乎不需要,因此可以轻松采用。我们对公平社区中使用的标准基准数据集进行了对我们提出的方法的经验分析。
Highly regulated domains such as finance have long favoured the use of machine learning algorithms that are scalable, transparent, robust and yield better performance. One of the most prominent examples of such an algorithm is XGBoost. Meanwhile, there is also a growing interest in building fair and unbiased models in these regulated domains and numerous bias-mitigation algorithms have been proposed to this end. However, most of these bias-mitigation methods are restricted to specific model families such as logistic regression or support vector machine models, thus leaving modelers with a difficult decision of choosing between fairness from the bias-mitigation algorithms and scalability, transparency, performance from algorithms such as XGBoost. We aim to leverage the best of both worlds by proposing a fair variant of XGBoost that enjoys all the advantages of XGBoost, while also matching the levels of fairness from the state-of-the-art bias-mitigation algorithms. Furthermore, the proposed solution requires very little in terms of changes to the original XGBoost library, thus making it easy for adoption. We provide an empirical analysis of our proposed method on standard benchmark datasets used in the fairness community.