Downsampling Attack on Automatic Speaker Authentication System

S. Asha, P. Vinod, Varun G. Menon, A. Zemmari
{"title":"Downsampling Attack on Automatic Speaker Authentication System","authors":"S. Asha, P. Vinod, Varun G. Menon, A. Zemmari","doi":"10.1109/AICCSA53542.2021.9686767","DOIUrl":null,"url":null,"abstract":"Recent years have observed an exponential growth in the popularity of audio-based authentication systems. The benefit of a voice-based authentication system is that the person need not be physically present. Voice biometric system provides effective authentication in various domains like remote access control, authentication in mobile applications, customer care centers for call attests. Most of the existing authentication systems that recognize speakers formulate deep learning models for better classification. At the same time, research studies show that deep learning models are highly vulnerable to adversarial inputs. A breach in security on authentication systems are not generally acceptable. This paper exposes the vulnerabilities of audio-based authentication systems. Here, we propose a novel downsampling attack to the speaker recognition system. This attack can effectively trick the speaker recognition framework by causing inaccurate predictions. The proposed threat model achieved remarkable attack effectiveness of 75%. This system employs a custom human voice dataset recorded in real-time conditions to achieve real-time effectiveness during classification. We compare the attack accuracy of the proposed attack against the adversarial audios generated using the CleverHans toolbox. The proposed attack being a black box attack, is transferable to other deep learning systems also.","PeriodicalId":423896,"journal":{"name":"2021 IEEE/ACS 18th International Conference on Computer Systems and Applications (AICCSA)","volume":"322 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE/ACS 18th International Conference on Computer Systems and Applications (AICCSA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/AICCSA53542.2021.9686767","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Recent years have observed an exponential growth in the popularity of audio-based authentication systems. The benefit of a voice-based authentication system is that the person need not be physically present. Voice biometric system provides effective authentication in various domains like remote access control, authentication in mobile applications, customer care centers for call attests. Most of the existing authentication systems that recognize speakers formulate deep learning models for better classification. At the same time, research studies show that deep learning models are highly vulnerable to adversarial inputs. A breach in security on authentication systems are not generally acceptable. This paper exposes the vulnerabilities of audio-based authentication systems. Here, we propose a novel downsampling attack to the speaker recognition system. This attack can effectively trick the speaker recognition framework by causing inaccurate predictions. The proposed threat model achieved remarkable attack effectiveness of 75%. This system employs a custom human voice dataset recorded in real-time conditions to achieve real-time effectiveness during classification. We compare the attack accuracy of the proposed attack against the adversarial audios generated using the CleverHans toolbox. The proposed attack being a black box attack, is transferable to other deep learning systems also.
自动说话人认证系统的降采样攻击
近年来,基于音频的身份验证系统的普及呈指数级增长。基于语音的身份验证系统的好处是,这个人不需要亲自在场。语音生物识别系统为远程访问控制、移动应用认证、客户服务中心呼叫认证等领域提供了有效的认证。大多数现有的识别说话人的身份验证系统都制定了深度学习模型来进行更好的分类。与此同时,研究表明,深度学习模型极易受到对抗性输入的影响。身份验证系统的安全漏洞通常是不可接受的。本文揭示了基于音频的认证系统的漏洞。本文提出了一种针对说话人识别系统的降采样攻击方法。这种攻击可以通过导致不准确的预测来有效地欺骗说话人识别框架。所提出的威胁模型达到了75%的攻击效率。该系统采用实时录制的自定义人声数据集,在分类过程中实现实时性。我们将提出的攻击精度与使用CleverHans工具箱生成的对抗性音频进行了比较。提出的攻击是一种黑盒攻击,也可以转移到其他深度学习系统。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信