论文标题
通过与ADMM有效沟通来改善保护隐私的垂直联合学习
Improving Privacy-Preserving Vertical Federated Learning by Efficient Communication with ADMM
论文作者
论文摘要
联合学习(FL)可以使分布式资源受限的设备共同培训共享模型,同时将培训数据保留为隐私目的。允许每个客户收集部分功能的垂直FL(VFL)最近吸引了密集的研究工作。我们确定了现有的VFL框架面临的主要挑战:服务器需要与客户进行每个培训步骤进行梯度交流,从而产生高沟通成本,从而导致快速消费隐私预算。为了应对这些挑战,在本文中,我们引入了一个带有多个头(VIM)的VFL框架,该框架将每个客户的单独贡献考虑在内,并使VFL优化目标有效地分解对服务器和客户本身可以迭代处理的子目标。特别是,我们提出了一种乘数的交替方向方法(ADMM)的方法来解决我们的优化问题,该方法允许客户在通信之前进行多个本地更新,从而降低了通信成本并在不同的隐私(DP)下导致更好的性能。我们为我们的框架提供了用户级的DP机制,以保护用户隐私。此外,我们表明VIM的副产品是学习的头部的权重反映了当地客户的重要性。我们进行了广泛的评估,并表明,在四个垂直FL数据集上,VIM与最先进的表现相比,vim的性能明显更高和更快的收敛性。我们还明确评估了本地客户的重要性,并表明VIM启用了诸如客户端级解释和客户端Denoising之类的功能。我们希望这项工作能够阐明一种有效的VFL培训和理解的新方法。
Federated learning (FL) enables distributed resource-constrained devices to jointly train shared models while keeping the training data local for privacy purposes. Vertical FL (VFL), which allows each client to collect partial features, has attracted intensive research efforts recently. We identified the main challenges that existing VFL frameworks are facing: the server needs to communicate gradients with the clients for each training step, incurring high communication cost that leads to rapid consumption of privacy budgets. To address these challenges, in this paper, we introduce a VFL framework with multiple heads (VIM), which takes the separate contribution of each client into account, and enables an efficient decomposition of the VFL optimization objective to sub-objectives that can be iteratively tackled by the server and the clients on their own. In particular, we propose an Alternating Direction Method of Multipliers (ADMM)-based method to solve our optimization problem, which allows clients to conduct multiple local updates before communication, and thus reduces the communication cost and leads to better performance under differential privacy (DP). We provide the user-level DP mechanism for our framework to protect user privacy. Moreover, we show that a byproduct of VIM is that the weights of learned heads reflect the importance of local clients. We conduct extensive evaluations and show that on four vertical FL datasets, VIM achieves significantly higher performance and faster convergence compared with the state-of-the-art. We also explicitly evaluate the importance of local clients and show that VIM enables functionalities such as client-level explanation and client denoising. We hope this work will shed light on a new way of effective VFL training and understanding.