Detailed Information

Cited 0 time in webofscience Cited 3 time in scopus
Metadata Downloads

Joint optimization of neural acoustic beamforming and dereverberation with x-vectors for robust speaker verification

Full metadata record
DC Field Value Language
dc.contributor.authorYang, Joon-Young-
dc.contributor.authorChang, Joon Hyuk-
dc.date.accessioned2021-07-30T05:13:58Z-
dc.date.available2021-07-30T05:13:58Z-
dc.date.created2021-05-13-
dc.date.issued2019-09-
dc.identifier.issn2308-457X-
dc.identifier.urihttps://scholarworks.bwise.kr/hanyang/handle/2021.sw.hanyang/3801-
dc.description.abstractIn this paper, we investigate the deep neural network (DNN) supported acoustic beamforming and dereverberation as the front-end of the x-vector speaker verification (SV) framework in a noisy and reverberant environment. Firstly, a DNN for supporting either the classical beamforming (e.g. MVDR) or the dereverberation (e.g. WPE) algorithm is trained on multi-channel speech signals. Next, an x-vector speaker embedding network is trained on top of the enhanced speech features to classify the training speakers. Finally, after the separate training stages are over, either one or both of the DNN supported beamforming and dereverberation modules are serially connected to the x-vector network, and jointly trained to optimize the common objective of speaker classification. Experiments on the artificially generated speech dataset using simulated and real room impulse responses (RIRs) with various types of domestic noise samples show that jointly training the supportive neural network models along with the x-vector network within the classical speech enhancement framework brings significant performance gain for robust text-independent (TI) SV.-
dc.language영어-
dc.language.isoen-
dc.publisherInternational Speech Communication Association-
dc.titleJoint optimization of neural acoustic beamforming and dereverberation with x-vectors for robust speaker verification-
dc.typeArticle-
dc.contributor.affiliatedAuthorChang, Joon Hyuk-
dc.identifier.doi10.21437/Interspeech.2019-1356-
dc.identifier.scopusid2-s2.0-85074681357-
dc.identifier.wosid000831796404044-
dc.identifier.bibliographicCitationProceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH, v.2019, no.September, pp.4075 - 4079-
dc.relation.isPartOfProceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH-
dc.citation.titleProceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH-
dc.citation.volume2019-
dc.citation.numberSeptember-
dc.citation.startPage4075-
dc.citation.endPage4079-
dc.type.rimsART-
dc.type.docTypeConference Paper-
dc.description.journalClass1-
dc.description.isOpenAccessN-
dc.description.journalRegisteredClassscopus-
dc.relation.journalResearchAreaAudiology & Speech-Language Pathology-
dc.relation.journalResearchAreaComputer Science-
dc.relation.journalWebOfScienceCategoryAudiology & Speech-Language Pathology-
dc.relation.journalWebOfScienceCategoryComputer Science, Artificial Intelligence-
dc.subject.keywordAuthorAcoustic beamforming-
dc.subject.keywordAuthorDeep neural network-
dc.subject.keywordAuthorDereverberation-
dc.subject.keywordAuthorJoint training-
dc.subject.keywordAuthorSpeaker verification-
dc.identifier.urlhttps://www.isca-speech.org/archive/interspeech_2019/yang19g_interspeech.html-
Files in This Item
Go to Link
Appears in
Collections
서울 공과대학 > 서울 융합전자공학부 > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Chang, Joon-Hyuk photo

Chang, Joon-Hyuk
COLLEGE OF ENGINEERING (SCHOOL OF ELECTRONIC ENGINEERING)
Read more

Altmetrics

Total Views & Downloads

BROWSE