论文标题
公平元学习的原始双重亚级别方法
A Primal-Dual Subgradient Approachfor Fair Meta Learning
论文作者
论文摘要
在培训期间学习概括为看不见的课程的问题(称为少数分类)引起了很大的关注。基于初始化的方法,例如基于梯度的模型不可知的元学习(MAML),通过“学习进行微调”来解决少数射击学习问题。这些方法的目的是学习适当的模型初始化,以便可以从一些标记的示例中学习新类的分类器,并具有少量的梯度更新步骤。很少有射击元学习众所周知,其快速适应能力和准确的概括是看不见的任务。通过公正的成果进行公平学习是人类智力的另一个重要标志,这很少在几次元学习中被触摸。在这项工作中,我们提出了一个原始的二次公平元学习框架,即PDFM,该框架学会了仅根据来自相关任务的数据使用几个示例来训练公平机器学习模型。关键想法是学习公平模型的原始和双重参数的良好初始化,以便它可以通过一些渐变更新步骤适应新的公平学习任务。 PDFM不是通过网格搜索手动调整双重参数作为超参数,而是通过亚级别的原始双偶对方法来优化原始参数和双参数的初始化,以共同进行公平的元学习。我们进一步实例化了使用均值差异和决策边界协方差的偏见控制的例子,分别是对每个任务的公平限制,分别是监督回归和分类。我们通过将方法应用于各种现实世界数据集来证明我们提出的方法的多功能性。我们的实验显示了对此环境的最佳先前工作的实质性改进。
The problem of learning to generalize to unseen classes during training, known as few-shot classification, has attracted considerable attention. Initialization based methods, such as the gradient-based model agnostic meta-learning (MAML), tackle the few-shot learning problem by "learning to fine-tune". The goal of these approaches is to learn proper model initialization, so that the classifiers for new classes can be learned from a few labeled examples with a small number of gradient update steps. Few shot meta-learning is well-known with its fast-adapted capability and accuracy generalization onto unseen tasks. Learning fairly with unbiased outcomes is another significant hallmark of human intelligence, which is rarely touched in few-shot meta-learning. In this work, we propose a Primal-Dual Fair Meta-learning framework, namely PDFM, which learns to train fair machine learning models using only a few examples based on data from related tasks. The key idea is to learn a good initialization of a fair model's primal and dual parameters so that it can adapt to a new fair learning task via a few gradient update steps. Instead of manually tuning the dual parameters as hyperparameters via a grid search, PDFM optimizes the initialization of the primal and dual parameters jointly for fair meta-learning via a subgradient primal-dual approach. We further instantiate examples of bias controlling using mean difference and decision boundary covariance as fairness constraints to each task for supervised regression and classification, respectively. We demonstrate the versatility of our proposed approach by applying our approach to various real-world datasets. Our experiments show substantial improvements over the best prior work for this setting.