{"title":"Downsampling Attack on Automatic Speaker Authentication System","authors":"S. Asha, P. Vinod, Varun G. Menon, A. Zemmari","doi":"10.1109/AICCSA53542.2021.9686767","DOIUrl":null,"url":null,"abstract":"Recent years have observed an exponential growth in the popularity of audio-based authentication systems. The benefit of a voice-based authentication system is that the person need not be physically present. Voice biometric system provides effective authentication in various domains like remote access control, authentication in mobile applications, customer care centers for call attests. Most of the existing authentication systems that recognize speakers formulate deep learning models for better classification. At the same time, research studies show that deep learning models are highly vulnerable to adversarial inputs. A breach in security on authentication systems are not generally acceptable. This paper exposes the vulnerabilities of audio-based authentication systems. Here, we propose a novel downsampling attack to the speaker recognition system. This attack can effectively trick the speaker recognition framework by causing inaccurate predictions. The proposed threat model achieved remarkable attack effectiveness of 75%. This system employs a custom human voice dataset recorded in real-time conditions to achieve real-time effectiveness during classification. We compare the attack accuracy of the proposed attack against the adversarial audios generated using the CleverHans toolbox. The proposed attack being a black box attack, is transferable to other deep learning systems also.","PeriodicalId":423896,"journal":{"name":"2021 IEEE/ACS 18th International Conference on Computer Systems and Applications (AICCSA)","volume":"322 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE/ACS 18th International Conference on Computer Systems and Applications (AICCSA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/AICCSA53542.2021.9686767","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Recent years have observed an exponential growth in the popularity of audio-based authentication systems. The benefit of a voice-based authentication system is that the person need not be physically present. Voice biometric system provides effective authentication in various domains like remote access control, authentication in mobile applications, customer care centers for call attests. Most of the existing authentication systems that recognize speakers formulate deep learning models for better classification. At the same time, research studies show that deep learning models are highly vulnerable to adversarial inputs. A breach in security on authentication systems are not generally acceptable. This paper exposes the vulnerabilities of audio-based authentication systems. Here, we propose a novel downsampling attack to the speaker recognition system. This attack can effectively trick the speaker recognition framework by causing inaccurate predictions. The proposed threat model achieved remarkable attack effectiveness of 75%. This system employs a custom human voice dataset recorded in real-time conditions to achieve real-time effectiveness during classification. We compare the attack accuracy of the proposed attack against the adversarial audios generated using the CleverHans toolbox. The proposed attack being a black box attack, is transferable to other deep learning systems also.