论文标题
Transmia:使用转移影子训练的会员推理攻击
TransMIA: Membership Inference Attacks Using Transfer Shadow Training
论文作者
论文摘要
通过转移在不同的培训中获得的一些知识,对转移学习进行了广泛的研究和越来越受欢迎,以提高机器学习模型的准确性。但是,没有先前的工作指出,转移学习可以加强对机器学习模型的隐私攻击。在本文中,我们提出了Transmia(基于转移学习的会员推理攻击),这些攻击使用转移学习在对手能够访问转移模型的参数时对源模型进行成员推理攻击。特别是,我们提出了一种转移影子训练技术,在该技术中,当对手使用有限量的影子训练数据提供给对手时,对手采用转移模型的参数来构建阴影模型。我们使用两个真实数据集评估了攻击,并表明我们的攻击表现优于不使用我们的转移影子训练技术的最先进。我们还比较了基于学习/熵的方法和微调/冻结方法的四个组合,所有这些方法都采用了我们的转移影子训练技术。然后,我们根据置信价值的分布来检查这四种方法的性能,并讨论针对我们攻击的可能对策。
Transfer learning has been widely studied and gained increasing popularity to improve the accuracy of machine learning models by transferring some knowledge acquired in different training. However, no prior work has pointed out that transfer learning can strengthen privacy attacks on machine learning models. In this paper, we propose TransMIA (Transfer learning-based Membership Inference Attacks), which use transfer learning to perform membership inference attacks on the source model when the adversary is able to access the parameters of the transferred model. In particular, we propose a transfer shadow training technique, where an adversary employs the parameters of the transferred model to construct shadow models, to significantly improve the performance of membership inference when a limited amount of shadow training data is available to the adversary. We evaluate our attacks using two real datasets, and show that our attacks outperform the state-of-the-art that does not use our transfer shadow training technique. We also compare four combinations of the learning-based/entropy-based approach and the fine-tuning/freezing approach, all of which employ our transfer shadow training technique. Then we examine the performance of these four approaches based on the distributions of confidence values, and discuss possible countermeasures against our attacks.