Void: A fast and light voice liveness detection system
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Ahmed, M.E.[Ahmed, M.E.] | - |
dc.contributor.author | Kwak, I.-Y.[Kwak, I.-Y.] | - |
dc.contributor.author | Huh, J.H.[Huh, J.H.] | - |
dc.contributor.author | Kim, I.[Kim, I.] | - |
dc.contributor.author | Oh, T.[Oh, T.] | - |
dc.contributor.author | Kim, H.[Kim, H.] | - |
dc.date.accessioned | 2021-07-28T13:48:10Z | - |
dc.date.available | 2021-07-28T13:48:10Z | - |
dc.date.created | 2021-05-10 | - |
dc.date.issued | 2020 | - |
dc.identifier.issn | 0000-0000 | - |
dc.identifier.uri | https://scholarworks.bwise.kr/skku/handle/2021.sw.skku/6861 | - |
dc.description.abstract | Due to the open nature of voice assistants' input channels, adversaries could easily record people's use of voice commands, and replay them to spoof voice assistants. To mitigate such spoofing attacks, we present a highly efficient voice liveness detection solution called “Void.” Void detects voice spoofing attacks using the differences in spectral power between live-human voices and voices replayed through speakers. In contrast to existing approaches that use multiple deep learning models, and thousands of features, Void uses a single classification model with just 97 features. We used two datasets to evaluate its performance: (1) 255,173 voice samples generated with 120 participants, 15 playback devices and 12 recording devices, and (2) 18,030 publicly available voice samples generated with 42 participants, 26 playback devices and 25 recording devices. Void achieves equal error rate of 0.3% and 11.6% in detecting voice replay attacks for each dataset, respectively. Compared to a state of the art, deep learning-based solution that achieves 7.4% error rate in that public dataset, Void uses 153 times less memory and is about 8 times faster in detection. When combined with a Gaussian Mixture Model that uses Mel-frequency cepstral coefficients (MFCC) as classification features - MFCC is already being extracted and used as the main feature in speech recognition services - Void achieves 8.7% error rate on the public dataset. Moreover, Void is resilient against hidden voice command, inaudible voice command, voice synthesis, equalization manipulation attacks, and combining replay attacks with live-human voices achieving about 99.7%, 100%, 90.2%, 86.3%, and 98.2% detection rates for those attacks, respectively. © 2020 by The USENIX Association. All Rights Reserved. | - |
dc.language | 영어 | - |
dc.language.iso | en | - |
dc.publisher | USENIX Association | - |
dc.title | Void: A fast and light voice liveness detection system | - |
dc.type | Article | - |
dc.contributor.affiliatedAuthor | Ahmed, M.E.[Ahmed, M.E.] | - |
dc.contributor.affiliatedAuthor | Kim, H.[Kim, H.] | - |
dc.identifier.scopusid | 2-s2.0-85091928064 | - |
dc.identifier.bibliographicCitation | Proceedings of the 29th USENIX Security Symposium, pp.2685 - 2702 | - |
dc.relation.isPartOf | Proceedings of the 29th USENIX Security Symposium | - |
dc.citation.title | Proceedings of the 29th USENIX Security Symposium | - |
dc.citation.startPage | 2685 | - |
dc.citation.endPage | 2702 | - |
dc.type.rims | ART | - |
dc.type.docType | Conference Paper | - |
dc.description.journalClass | 1 | - |
dc.description.isOpenAccess | N | - |
dc.description.journalRegisteredClass | scopus | - |
dc.subject.keywordPlus | Classification (of information) | - |
dc.subject.keywordPlus | Deep learning | - |
dc.subject.keywordPlus | Errors | - |
dc.subject.keywordPlus | Gaussian distribution | - |
dc.subject.keywordPlus | Classification features | - |
dc.subject.keywordPlus | Classification models | - |
dc.subject.keywordPlus | Equal error rate | - |
dc.subject.keywordPlus | Gaussian Mixture Model | - |
dc.subject.keywordPlus | Liveness detection | - |
dc.subject.keywordPlus | Mel-frequency cepstral coefficients | - |
dc.subject.keywordPlus | Recording devices | - |
dc.subject.keywordPlus | Spoofing attacks | - |
dc.subject.keywordPlus | Speech recognition | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
(03063) 25-2, SUNGKYUNKWAN-RO, JONGNO-GU, SEOUL, KOREAsamsunglib@skku.edu
COPYRIGHT © 2021 SUNGKYUNKWAN UNIVERSITY ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.