Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

Comparative Analysis of Vision Transformer Models for Facial Emotion Recognition Using Augmented Balanced Datasets

Full metadata record
DC FieldValueLanguage
dc.contributor.authorBobojanov, Sukhrob-
dc.contributor.authorKim, Byeong Man-
dc.contributor.authorArabboev, Mukhriddin-
dc.contributor.authorBegmatov, Shohruh-
dc.date.accessioned2024-01-11T05:30:29Z-
dc.date.available2024-01-11T05:30:29Z-
dc.date.issued2023-11-
dc.identifier.issn2076-3417-
dc.identifier.urihttps://scholarworks.bwise.kr/kumoh/handle/2020.sw.kumoh/26465-
dc.description.abstractFacial emotion recognition (FER) has a huge importance in the field of human-machine interface. Given the intricacies of human facial expressions and the inherent variations in images, which are characterized by diverse facial poses and lighting conditions, the task of FER remains a challenging endeavour for computer-based models. Recent advancements have seen vision transformer (ViT) models attain state-of-the-art results across various computer vision tasks, encompassing image classification, object detection, and segmentation. Moreover, one of the most important aspects of creating strong machine learning models is correcting data imbalances. To avoid biased predictions and guarantee reliable findings, it is essential to maintain the distribution equilibrium of the training dataset. In this work, we have chosen two widely used open-source datasets, RAF-DB and FER2013. As well as resolving the imbalance problem, we present a new, balanced dataset, applying data augmentation techniques and cleaning poor-quality images from the FER2013 dataset. We then conduct a comprehensive evaluation of thirteen different ViT models with these three datasets. Our investigation concludes that ViT models present a promising approach for FER tasks. Among these ViT models, Mobile ViT and Tokens-to-Token ViT models appear to be the most effective, followed by PiT and Cross Former models.-
dc.language영어-
dc.language.isoENG-
dc.publisherMDPI-
dc.titleComparative Analysis of Vision Transformer Models for Facial Emotion Recognition Using Augmented Balanced Datasets-
dc.typeArticle-
dc.publisher.location스위스-
dc.identifier.doi10.3390/app132212271-
dc.identifier.wosid001118096900001-
dc.identifier.bibliographicCitationAPPLIED SCIENCES-BASEL, v.13, no.22-
dc.citation.titleAPPLIED SCIENCES-BASEL-
dc.citation.volume13-
dc.citation.number22-
dc.type.docTypeArticle-
dc.description.isOpenAccessY-
dc.description.journalRegisteredClassscie-
dc.description.journalRegisteredClassscopus-
dc.relation.journalResearchAreaChemistry-
dc.relation.journalResearchAreaEngineering-
dc.relation.journalResearchAreaMaterials Science-
dc.relation.journalResearchAreaPhysics-
dc.relation.journalWebOfScienceCategoryChemistry, Multidisciplinary-
dc.relation.journalWebOfScienceCategoryEngineering, Multidisciplinary-
dc.relation.journalWebOfScienceCategoryMaterials Science, Multidisciplinary-
dc.relation.journalWebOfScienceCategoryPhysics, Applied-
dc.subject.keywordAuthorfacial emotion recognition-
dc.subject.keywordAuthorvision transformer-
dc.subject.keywordAuthordata augmentation-
dc.subject.keywordAuthorbalanced data-
dc.subject.keywordAuthorFER2013-
dc.subject.keywordAuthorRAF-DB-
Files in This Item
There are no files associated with this item.
Appears in
Collections
Department of Computer Software Engineering > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher KIM, BYEONG MAN photo

KIM, BYEONG MAN
College of Engineering (Department of Computer Software Engineering)
Read more

Altmetrics

Total Views & Downloads

BROWSE