首页> 外文期刊>Computer speech and language >Adversarial attack and defense strategies for deep speaker recognition systems
【24h】

Adversarial attack and defense strategies for deep speaker recognition systems

机译:深层扬声器识别系统的对抗攻击与防御策略

获取原文
获取原文并翻译 | 示例

摘要

Robust speaker recognition, including in the presence of malicious attacks, is becoming increasingly important and essential, especially due to the proliferation of smart speakers and personal agents that interact with an individual's voice commands to perform diverse and even sensitive tasks. Adversarial attack is a recently revived domain which is shown to be effective in breaking deep neural network-based classifiers, specifically, by forcing them to change their posterior distribution by only perturbing the input samples by a very small amount. Although, significant progress in this realm has been made in the computer vision domain, advances within speaker recognition is still limited. We present an expository paper that considers several adversarial attacks to a deep speaker recognition system, employs strong defense methods as countermeasures, and reports a comprehensive set of ablation studies to better understand the problem. The experiments show that the speaker recognition systems are vulnerable to adversarial attacks, and the strongest attacks can reduce the accuracy of the system from 94% to even 0%. The study also compares the performances of the employed defense methods in detail, and finds adversarial training based on Projected Gradient Descent (PGD) to be the best defense method in our setting. We hope that the experiments presented in this paper provide baselines that can be useful for the research community interested in further studying adversarial robustness of speaker recognition systems.
机译:强大的说话人识别,包括恶意攻击的存在,正变得越来越重要和必要的,特别是由于智能扬声器和个人代理,与一个人的语音命令交互,以执行不同的甚至是敏感的任务的增殖。对抗性攻击是被证明是有效地打破了深基于神经网络的分类,具体而言,通过迫使他们只用极少量的扰动输入样本,以改变他们的后验分布的最近恢复域。虽然,在这一领域显著的进步已经在计算机视觉领域已经取得,说话人识别中的进展仍然有限。我们提出了一个说明性文件是考虑到深说话人识别系统的几个敌对攻击,采用了坚固的防御方法作为对策,并报告了一套全面消融研究,以更好地理解这个问题。该实验表明,该说话人识别系统很容易受到攻击敌对,和最强攻击可以从94%降低了系统的准确度,甚至0%。该研究还比较了详细的就业防御方法的性能,并认为基于投影梯度下降(PGD)对抗性训练,在我们的环境中最好的防御方法。我们希望,在这个文件中提出的实验提供了基线,可以为有兴趣进一步研究说话人识别系统的鲁棒性对抗性研究团体有用。

著录项

  • 来源
    《Computer speech and language》 |2021年第7期|101199.1-101199.14|共14页
  • 作者单位

    Electrical and Computer Engineering University of Southern California (USC) Los Angeles CA USA;

    Electrical and Computer Engineering University of Southern California (USC) Los Angeles CA USA;

    Electrical and Computer Engineering University of Southern California (USC) Los Angeles CA USA;

    Electrical and Computer Engineering University of Southern California (USC) Los Angeles CA USA;

    Electrical and Computer Engineering University of Southern California (USC) Los Angeles CA USA USC Information Sciences Institute Marina del Rey CA USA;

    Electrical and Computer Engineering University of Southern California (USC) Los Angeles CA USA USC Information Sciences Institute Marina del Rey CA USA;

  • 收录信息 美国《科学引文索引》(SCI);美国《工程索引》(EI);
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Adversarial attack; Deep neural network; Speaker recognition;

    机译:对抗攻击;深神经网络;扬声器认可;

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号