Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

Self-Distillation into Self-Attention Heads for Improving Transformer-based End-to-End Neural Speaker Diarization

Authors
Jeoung, Ye-RinChoi, Jeong-HwanSeong, Ju-SeokKyung, JeHyunChang, Joon-Hyuk
Issue Date
Aug-2023
Publisher
International Speech Communication Association
Keywords
end-to-end neural diarization; fine-tuning; self-attention mechanism; self-distillation; speaker diarization
Citation
Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH, v.2023-August, pp.3197 - 3201
Indexed
SCOPUS
Journal Title
Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH
Volume
2023-August
Start Page
3197
End Page
3201
URI
https://scholarworks.bwise.kr/hanyang/handle/2021.sw.hanyang/191793
DOI
10.21437/Interspeech.2023-1404
ISSN
2308-457X
Abstract
In this study, we explore self-distillation (SD) techniques to improve the performance of the transformer-encoder-based self-attentive (SA) end-to-end neural speaker diarization (EEND). We first apply the SD approaches, introduced in the automatic speech recognition field, to the SA-EEND model to confirm their potential for speaker diarization. Then, we propose two novel SD methods for the SA-EEND, which distill the prediction output of the model or the SA heads of the upper blocks into the SA heads of the lower blocks. Consequently, we expect the high-level speaker-discriminative knowledge learned by the upper blocks to be shared across the lower blocks, thereby enabling the SA heads of the lower blocks to effectively capture the discriminative patterns of overlapped speech of multiple speakers. Experimental results on the simulated and CALLHOME datasets show that the SD generally improves the baseline performance, and the proposed methods outperform the conventional SD approaches.
Files in This Item
Go to Link
Appears in
Collections
서울 공과대학 > 서울 융합전자공학부 > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Chang, Joon-Hyuk photo

Chang, Joon-Hyuk
COLLEGE OF ENGINEERING (SCHOOL OF ELECTRONIC ENGINEERING)
Read more

Altmetrics

Total Views & Downloads

BROWSE