论文标题

SA:抗剪辑和自剪接的综合语音检测的滑动攻击

SA: Sliding attack for synthetic speech detection with resistance to clipping and self-splicing

论文作者

JiaCheng, Deng, Li, Dong, Diqun, Yan, Rangding, Wang, Jiaming, Zeng

论文摘要

深度神经网络很容易受到对抗性示例的影响,这些例子误解了无法察觉的扰动模型。在音频中,尽管对抗性示例已经在白色框设置和黑色框设置上实现了令人难以置信的攻击成功率,但大多数现有的对抗性攻击都受到输入长度的限制。一个更实用的情况是,必须将对抗性示例剪切或自sple缩,然后输入黑框模型。因此,有必要探索如何在不同输入长度设置中提高可传递性。在本文中,我们以综合语音检测任务为例,并考虑两个代表性的SOTA模型。我们观察到,在不同模型中,具有相同样品值的片段梯度通过分析在裁剪或自拼插后通过将样品喂入模型中获得的梯度相似。受上述观察的启发,我们提出了一种称为滑动攻击的新的对抗攻击方法。具体而言,我们使每个采样点都意识到不同位置的梯度,这些梯度可以模拟对抗性示例输入到具有不同输入长度的黑盒模型的情况。因此,我们没有在梯度计算的每次迭代中直接使用当前梯度,而是经过以下三个步骤。首先,我们使用滑动窗口提取不同长度的子播种。然后,我们使用来自相邻域的数据来增强子播种。最后,我们将子细分市场馈送到不同的模型中,以获取骨料梯度以更新对抗性示例。经验结果表明,我们的方法可以显着提高剪辑或自拼图后对抗性示例的可传递性。此外,我们的方法还可以根据不同特征增强模型之间的可传递性。

Deep neural networks are vulnerable to adversarial examples that mislead models with imperceptible perturbations. In audio, although adversarial examples have achieved incredible attack success rates on white-box settings and black-box settings, most existing adversarial attacks are constrained by the input length. A More practical scenario is that the adversarial examples must be clipped or self-spliced and input into the black-box model. Therefore, it is necessary to explore how to improve transferability in different input length settings. In this paper, we take the synthetic speech detection task as an example and consider two representative SOTA models. We observe that the gradients of fragments with the same sample value are similar in different models via analyzing the gradients obtained by feeding samples into the model after cropping or self-splicing. Inspired by the above observation, we propose a new adversarial attack method termed sliding attack. Specifically, we make each sampling point aware of gradients at different locations, which can simulate the situation where adversarial examples are input to black-box models with varying input lengths. Therefore, instead of using the current gradient directly in each iteration of the gradient calculation, we go through the following three steps. First, we extract subsegments of different lengths using sliding windows. We then augment the subsegments with data from the adjacent domains. Finally, we feed the sub-segments into different models to obtain aggregate gradients to update adversarial examples. Empirical results demonstrate that our method could significantly improve the transferability of adversarial examples after clipping or self-splicing. Besides, our method could also enhance the transferability between models based on different features.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源