Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

Prediction of Closed Quotient During Vocal Phonation using GRU-type Neural Network with Audio Signals

Full metadata record
DC Field Value Language
dc.contributor.authorHan,Hyeonbin-
dc.contributor.authorLee,Keun Young-
dc.contributor.authorShin,Seong-Yoon-
dc.contributor.authorKim,Yoseup-
dc.contributor.authorJo,wanghyun-
dc.contributor.authorPark ,Jihoon-
dc.contributor.authorKim,Young-Min-
dc.date.accessioned2024-07-10T07:00:23Z-
dc.date.available2024-07-10T07:00:23Z-
dc.date.issued2024-06-
dc.identifier.issn2234-8255-
dc.identifier.issn2234-8883-
dc.identifier.urihttps://scholarworks.bwise.kr/erica/handle/2021.sw.erica/119845-
dc.description.abstractClosed quotient (CQ) represents the time ratio for which the vocal folds remain in contact during voice production. Because analyzing CQ values serves as an important reference point in vocal training for professional singers, these values have been measured mechanically or electrically by either inverse filtering of airflows captured by a circumferentially vented mask or post-processing of electroglottography waveforms. In this study, we introduced a novel algorithm to predict the CQ values only from audio signals. This has eliminated the need for mechanical or electrical measurement techniques. Our algorithm is based on a gated recurrent unit (GRU)-type neural network. To enhance the efficiency, we pre-processed an audio signal using the pitch feature extraction algorithm. Then, GRU-type neural networks were employed to extract the features. This was followed by a dense layer for the final prediction. The Results section reports the mean square error between the predicted and real CQ. It shows the capability of the proposed algorithm to predict CQ values.-
dc.format.extent8-
dc.language영어-
dc.language.isoENG-
dc.publisherThe Korean Institute of Information and Commucation Engineering-
dc.titlePrediction of Closed Quotient During Vocal Phonation using GRU-type Neural Network with Audio Signals-
dc.typeArticle-
dc.publisher.location대한민국-
dc.identifier.doi10.56977/jicce.2024.22.2.145-
dc.identifier.bibliographicCitationJournal of Information and Communication Convergence Engineering, v.22, no.2, pp 145 - 152-
dc.citation.titleJournal of Information and Communication Convergence Engineering-
dc.citation.volume22-
dc.citation.number2-
dc.citation.startPage145-
dc.citation.endPage152-
dc.type.docType정기학술지(Article(Perspective Article포함))-
dc.identifier.kciidART003091476-
dc.description.isOpenAccessN-
dc.description.journalRegisteredClasskciCandi-
dc.subject.keywordAuthorVocal phonation-
dc.subject.keywordAuthorGRU, Artificial neural network-
dc.subject.keywordAuthorElectroglottography-
dc.identifier.urlhttps://www.dbpia.co.kr/journal/articleDetail?nodeId=NODE11824246-
Files in This Item
Go to Link
Appears in
Collections
COLLEGE OF SCIENCE AND CONVERGENCE TECHNOLOGY > ERICA 수리데이터사이언스학과 > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Jo, Gwanghyun photo

Jo, Gwanghyun
ERICA 과학기술융합대학 (ERICA 수리데이터사이언스학과)
Read more

Altmetrics

Total Views & Downloads

BROWSE