论文标题

扬声器标识的音频分类器的生成性提取

Generative Extraction of Audio Classifiers for Speaker Identification

论文作者

Afonja, Tejumade, Bourtoule, Lucas, Chandrasekaran, Varun, Oore, Sageev, Papernot, Nicolas

论文摘要

机器学习模型,尤其是深层神经网络特别容易受到攻击,也许不再奇怪。经过充分研究的脆弱性是模型提取:攻击者试图通过训练代理模型来模仿受害者模型的决策界限来窃取受害者模型的现象。以前的工作证明了这种攻击及其毁灭性后果的有效性,但是这项工作的大部分主要用于图像和文本处理任务。我们的工作是对{\ em音频分类模型}进行模型提取的首次尝试。我们的动机是一个攻击者的目标是模仿受害者模型的行为,该模型受到训练以识别说话者。这在对安全敏感的域(例如生物识别身份验证)中尤其有问题。我们发现先前的模型提取技术,其中攻击者\ textit {naivery}使用代理数据集攻击潜在受害者的模型失败。因此,我们建议使用生成模型来创建一个足够大的合成攻击查询池。我们发现,我们的方法能够使用与代理数据集合成的查询基于\ texttt {voxceleb};我们的测试准确性为84.41 \%,预算为300万个查询。

It is perhaps no longer surprising that machine learning models, especially deep neural networks, are particularly vulnerable to attacks. One such vulnerability that has been well studied is model extraction: a phenomenon in which the attacker attempts to steal a victim's model by training a surrogate model to mimic the decision boundaries of the victim model. Previous works have demonstrated the effectiveness of such an attack and its devastating consequences, but much of this work has been done primarily for image and text processing tasks. Our work is the first attempt to perform model extraction on {\em audio classification models}. We are motivated by an attacker whose goal is to mimic the behavior of the victim's model trained to identify a speaker. This is particularly problematic in security-sensitive domains such as biometric authentication. We find that prior model extraction techniques, where the attacker \textit{naively} uses a proxy dataset to attack a potential victim's model, fail. We therefore propose the use of a generative model to create a sufficiently large and diverse pool of synthetic attack queries. We find that our approach is able to extract a victim's model trained on \texttt{LibriSpeech} using queries synthesized with a proxy dataset based off of \texttt{VoxCeleb}; we achieve a test accuracy of 84.41\% with a budget of 3 million queries.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源