论文标题
梯度掩盖平均联合学习
Gradient Masked Averaging for Federated Learning
论文作者
论文摘要
联合学习(FL)是一个新兴的范式,允许大量的客户使用异质数据来协调统一的全球模型的学习,而无需相互共享数据。联合学习的一个主要挑战是客户端数据的异质性,这可以降低标准FL算法的性能。标准FL算法涉及平均模型参数或梯度更新以近似服务器上的全局模型。但是,我们认为,在异质环境中,平均可以导致信息丢失,并导致由于主要客户梯度引起的偏见而导致概括不良。我们假设这是为了在非i.i.D数据集中更好地概括,算法应专注于学习不变的机制,而这种机制是恒定的,同时忽略了各个客户的虚假机制。受到分数过度概括的最新作品的启发,我们提出了一种梯度掩盖的平均方法,以替代客户更新的标准平均。对于大多数现有的联合算法,可以将用于客户更新的这种聚合技术用于置换。我们在多种FL算法上进行了广泛的实验,具有分布,现实世界,功能相关的分布和数量不平衡数据集,并表明它提供了一致的改进,尤其是在异类客户端的情况下。
Federated learning (FL) is an emerging paradigm that permits a large number of clients with heterogeneous data to coordinate learning of a unified global model without the need to share data amongst each other. A major challenge in federated learning is the heterogeneity of data across client, which can degrade the performance of standard FL algorithms. Standard FL algorithms involve averaging of model parameters or gradient updates to approximate the global model at the server. However, we argue that in heterogeneous settings, averaging can result in information loss and lead to poor generalization due to the bias induced by dominant client gradients. We hypothesize that to generalize better across non-i.i.d datasets, the algorithms should focus on learning the invariant mechanism that is constant while ignoring spurious mechanisms that differ across clients. Inspired from recent works in Out-of-Distribution generalization, we propose a gradient masked averaging approach for FL as an alternative to the standard averaging of client updates. This aggregation technique for client updates can be adapted as a drop-in replacement in most existing federated algorithms. We perform extensive experiments on multiple FL algorithms with in-distribution, real-world, feature-skewed out-of-distribution, and quantity imbalanced datasets and show that it provides consistent improvements, particularly in the case of heterogeneous clients.