Self-Distillation into Self-Attention Heads for Improving Transformer-based End-to-End Neural Speaker Diarization
- Authors
- Jeoung, Ye-Rin; Choi, Jeong-Hwan; Seong, Ju-Seok; Kyung, JeHyun; Chang, Joon-Hyuk
- Issue Date
- Aug-2023
- Publisher
- International Speech Communication Association
- Keywords
- end-to-end neural diarization; fine-tuning; self-attention mechanism; self-distillation; speaker diarization
- Citation
- Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH, v.2023-August, pp.3197 - 3201
- Indexed
- SCOPUS
- Journal Title
- Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH
- Volume
- 2023-August
- Start Page
- 3197
- End Page
- 3201
- URI
- https://scholarworks.bwise.kr/hanyang/handle/2021.sw.hanyang/191793
- DOI
- 10.21437/Interspeech.2023-1404
- ISSN
- 2308-457X
- Abstract
- In this study, we explore self-distillation (SD) techniques to improve the performance of the transformer-encoder-based self-attentive (SA) end-to-end neural speaker diarization (EEND). We first apply the SD approaches, introduced in the automatic speech recognition field, to the SA-EEND model to confirm their potential for speaker diarization. Then, we propose two novel SD methods for the SA-EEND, which distill the prediction output of the model or the SA heads of the upper blocks into the SA heads of the lower blocks. Consequently, we expect the high-level speaker-discriminative knowledge learned by the upper blocks to be shared across the lower blocks, thereby enabling the SA heads of the lower blocks to effectively capture the discriminative patterns of overlapped speech of multiple speakers. Experimental results on the simulated and CALLHOME datasets show that the SD generally improves the baseline performance, and the proposed methods outperform the conventional SD approaches.
- Files in This Item
-
Go to Link
- Appears in
Collections - 서울 공과대학 > 서울 융합전자공학부 > 1. Journal Articles
![qrcode](https://api.qrserver.com/v1/create-qr-code/?size=55x55&data=https://scholarworks.bwise.kr/hanyang/handle/2021.sw.hanyang/191793)
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.