Joint optimization of neural acoustic beamforming and dereverberation with x-vectors for robust speaker verification
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Yang, Joon-Young | - |
dc.contributor.author | Chang, Joon Hyuk | - |
dc.date.accessioned | 2021-07-30T05:13:58Z | - |
dc.date.available | 2021-07-30T05:13:58Z | - |
dc.date.created | 2021-05-13 | - |
dc.date.issued | 2019-09 | - |
dc.identifier.issn | 2308-457X | - |
dc.identifier.uri | https://scholarworks.bwise.kr/hanyang/handle/2021.sw.hanyang/3801 | - |
dc.description.abstract | In this paper, we investigate the deep neural network (DNN) supported acoustic beamforming and dereverberation as the front-end of the x-vector speaker verification (SV) framework in a noisy and reverberant environment. Firstly, a DNN for supporting either the classical beamforming (e.g. MVDR) or the dereverberation (e.g. WPE) algorithm is trained on multi-channel speech signals. Next, an x-vector speaker embedding network is trained on top of the enhanced speech features to classify the training speakers. Finally, after the separate training stages are over, either one or both of the DNN supported beamforming and dereverberation modules are serially connected to the x-vector network, and jointly trained to optimize the common objective of speaker classification. Experiments on the artificially generated speech dataset using simulated and real room impulse responses (RIRs) with various types of domestic noise samples show that jointly training the supportive neural network models along with the x-vector network within the classical speech enhancement framework brings significant performance gain for robust text-independent (TI) SV. | - |
dc.language | 영어 | - |
dc.language.iso | en | - |
dc.publisher | International Speech Communication Association | - |
dc.title | Joint optimization of neural acoustic beamforming and dereverberation with x-vectors for robust speaker verification | - |
dc.type | Article | - |
dc.contributor.affiliatedAuthor | Chang, Joon Hyuk | - |
dc.identifier.doi | 10.21437/Interspeech.2019-1356 | - |
dc.identifier.scopusid | 2-s2.0-85074681357 | - |
dc.identifier.wosid | 000831796404044 | - |
dc.identifier.bibliographicCitation | Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH, v.2019, no.September, pp.4075 - 4079 | - |
dc.relation.isPartOf | Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH | - |
dc.citation.title | Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH | - |
dc.citation.volume | 2019 | - |
dc.citation.number | September | - |
dc.citation.startPage | 4075 | - |
dc.citation.endPage | 4079 | - |
dc.type.rims | ART | - |
dc.type.docType | Conference Paper | - |
dc.description.journalClass | 1 | - |
dc.description.isOpenAccess | N | - |
dc.description.journalRegisteredClass | scopus | - |
dc.relation.journalResearchArea | Audiology & Speech-Language Pathology | - |
dc.relation.journalResearchArea | Computer Science | - |
dc.relation.journalWebOfScienceCategory | Audiology & Speech-Language Pathology | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Artificial Intelligence | - |
dc.subject.keywordAuthor | Acoustic beamforming | - |
dc.subject.keywordAuthor | Deep neural network | - |
dc.subject.keywordAuthor | Dereverberation | - |
dc.subject.keywordAuthor | Joint training | - |
dc.subject.keywordAuthor | Speaker verification | - |
dc.identifier.url | https://www.isca-speech.org/archive/interspeech_2019/yang19g_interspeech.html | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
222, Wangsimni-ro, Seongdong-gu, Seoul, 04763, Korea+82-2-2220-1365
COPYRIGHT © 2021 HANYANG UNIVERSITY.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.