Modeling Speech Emotion Recognition via Attention-Oriented Parallel CNN Encoders
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Fazliddin, Makhmudov | - |
dc.contributor.author | Kutlimuratov, Alpamis | - |
dc.contributor.author | Akhmedov, Farkhod | - |
dc.contributor.author | Abdallah, Mohamed S. | - |
dc.contributor.author | Cho, Young-Im | - |
dc.date.accessioned | 2023-01-19T00:40:19Z | - |
dc.date.available | 2023-01-19T00:40:19Z | - |
dc.date.created | 2023-01-18 | - |
dc.date.issued | 2022-12 | - |
dc.identifier.issn | 2079-9292 | - |
dc.identifier.uri | https://scholarworks.bwise.kr/gachon/handle/2020.sw.gachon/86640 | - |
dc.description.abstract | Meticulous learning of human emotions through speech is an indispensable function of modern speech emotion recognition (SER) models. Consequently, deriving and interpreting various crucial speech features from raw speech data are complicated responsibilities in terms of modeling to improve performance. Therefore, in this study, we developed a novel SER model via attention-oriented parallel convolutional neural network (CNN) encoders that parallelly acquire important features that are used for emotion classification. Particularly, MFCC, paralinguistic, and speech spectrogram features were derived and encoded by designing different CNN architectures individually for the features, and the encoded features were fed to attention mechanisms for further representation, and then classified. Empirical veracity executed on EMO-DB and IEMOCAP open datasets, and the results showed that the proposed model is more efficient than the baseline models. Especially, weighted accuracy (WA) and unweighted accuracy (UA) of the proposed model were equal to 71.8% and 70.9% in EMO-DB dataset scenario, respectively. Moreover, WA and UA rates were 72.4% and 71.1% with the IEMOCAP dataset. | - |
dc.language | 영어 | - |
dc.language.iso | en | - |
dc.publisher | MDPI | - |
dc.relation.isPartOf | ELECTRONICS | - |
dc.title | Modeling Speech Emotion Recognition via Attention-Oriented Parallel CNN Encoders | - |
dc.type | Article | - |
dc.type.rims | ART | - |
dc.description.journalClass | 1 | - |
dc.identifier.wosid | 000896175000001 | - |
dc.identifier.doi | 10.3390/electronics11234047 | - |
dc.identifier.bibliographicCitation | ELECTRONICS, v.11, no.23 | - |
dc.description.isOpenAccess | Y | - |
dc.identifier.scopusid | 2-s2.0-85143658341 | - |
dc.citation.title | ELECTRONICS | - |
dc.citation.volume | 11 | - |
dc.citation.number | 23 | - |
dc.contributor.affiliatedAuthor | Fazliddin, Makhmudov | - |
dc.contributor.affiliatedAuthor | Kutlimuratov, Alpamis | - |
dc.contributor.affiliatedAuthor | Akhmedov, Farkhod | - |
dc.contributor.affiliatedAuthor | Abdallah, Mohamed S. | - |
dc.contributor.affiliatedAuthor | Cho, Young-Im | - |
dc.type.docType | Article | - |
dc.subject.keywordAuthor | speech emotion recognition | - |
dc.subject.keywordAuthor | convolution neural network | - |
dc.subject.keywordAuthor | attention | - |
dc.subject.keywordAuthor | deep learning | - |
dc.subject.keywordAuthor | modeling | - |
dc.relation.journalResearchArea | Computer Science | - |
dc.relation.journalResearchArea | Engineering | - |
dc.relation.journalResearchArea | Physics | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Information Systems | - |
dc.relation.journalWebOfScienceCategory | Engineering, Electrical & Electronic | - |
dc.relation.journalWebOfScienceCategory | Physics, Applied | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
1342, Seongnam-daero, Sujeong-gu, Seongnam-si, Gyeonggi-do, Republic of Korea(13120)031-750-5114
COPYRIGHT 2020 Gachon University All Rights Reserved.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.