Prediction of Closed Quotient During Vocal Phonation using GRU-type Neural Network with Audio Signals
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Han,Hyeonbin | - |
dc.contributor.author | Lee,Keun Young | - |
dc.contributor.author | Shin,Seong-Yoon | - |
dc.contributor.author | Kim,Yoseup | - |
dc.contributor.author | Jo,wanghyun | - |
dc.contributor.author | Park ,Jihoon | - |
dc.contributor.author | Kim,Young-Min | - |
dc.date.accessioned | 2024-07-10T07:00:23Z | - |
dc.date.available | 2024-07-10T07:00:23Z | - |
dc.date.issued | 2024-06 | - |
dc.identifier.issn | 2234-8255 | - |
dc.identifier.issn | 2234-8883 | - |
dc.identifier.uri | https://scholarworks.bwise.kr/erica/handle/2021.sw.erica/119845 | - |
dc.description.abstract | Closed quotient (CQ) represents the time ratio for which the vocal folds remain in contact during voice production. Because analyzing CQ values serves as an important reference point in vocal training for professional singers, these values have been measured mechanically or electrically by either inverse filtering of airflows captured by a circumferentially vented mask or post-processing of electroglottography waveforms. In this study, we introduced a novel algorithm to predict the CQ values only from audio signals. This has eliminated the need for mechanical or electrical measurement techniques. Our algorithm is based on a gated recurrent unit (GRU)-type neural network. To enhance the efficiency, we pre-processed an audio signal using the pitch feature extraction algorithm. Then, GRU-type neural networks were employed to extract the features. This was followed by a dense layer for the final prediction. The Results section reports the mean square error between the predicted and real CQ. It shows the capability of the proposed algorithm to predict CQ values. | - |
dc.format.extent | 8 | - |
dc.language | 영어 | - |
dc.language.iso | ENG | - |
dc.publisher | The Korean Institute of Information and Commucation Engineering | - |
dc.title | Prediction of Closed Quotient During Vocal Phonation using GRU-type Neural Network with Audio Signals | - |
dc.type | Article | - |
dc.publisher.location | 대한민국 | - |
dc.identifier.doi | 10.56977/jicce.2024.22.2.145 | - |
dc.identifier.bibliographicCitation | Journal of Information and Communication Convergence Engineering, v.22, no.2, pp 145 - 152 | - |
dc.citation.title | Journal of Information and Communication Convergence Engineering | - |
dc.citation.volume | 22 | - |
dc.citation.number | 2 | - |
dc.citation.startPage | 145 | - |
dc.citation.endPage | 152 | - |
dc.type.docType | 정기학술지(Article(Perspective Article포함)) | - |
dc.identifier.kciid | ART003091476 | - |
dc.description.isOpenAccess | N | - |
dc.description.journalRegisteredClass | kciCandi | - |
dc.subject.keywordAuthor | Vocal phonation | - |
dc.subject.keywordAuthor | GRU, Artificial neural network | - |
dc.subject.keywordAuthor | Electroglottography | - |
dc.identifier.url | https://www.dbpia.co.kr/journal/articleDetail?nodeId=NODE11824246 | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
55 Hanyangdeahak-ro, Sangnok-gu, Ansan, Gyeonggi-do, 15588, Korea+82-31-400-4269 sweetbrain@hanyang.ac.kr
COPYRIGHT © 2021 HANYANG UNIVERSITY. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.