Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

Deep-learning-based real-time silent speech recognition using facial electromyogram recorded around eyes for hands-free interfacing in a virtual reality environment

Authors
Cha, Ho-SeungChang, Won-DuIm, Chang-Hwan
Issue Date
Sep-2022
Publisher
SPRINGER LONDON LTD
Keywords
Deep learning; Facial electromyography; Human-computer interface; Myoelectric control; Silent speech recognition; Virtual reality
Citation
Virtual Reality, v.26, no.3, pp.1047 - 1057
Indexed
SCIE
SCOPUS
Journal Title
Virtual Reality
Volume
26
Number
3
Start Page
1047
End Page
1057
URI
https://scholarworks.bwise.kr/hanyang/handle/2021.sw.hanyang/185827
DOI
10.1007/s10055-021-00616-0
ISSN
1359-4338
Abstract
Speech recognition technology is a promising hands-free interfacing modality for virtual reality (VR) applications. However, it has several drawbacks, such as limited usability in a noisy environment or a public place and limited accessibility to those who cannot generate loud and clear voices. These limitations may be overcome by employing a silent speech recognition (SSR) technology utilizing facial electromyograms (fEMGs) in a VR environment. In the conventional SSR systems, however, fEMG electrodes were attached around the user's lips and neck, thereby creating new practical issues, such as the requirement of an additional wearable system besides the VR headset, necessity of a complex and time-consuming procedure for attaching the fEMG electrodes, and discomfort and limited facial muscle movements of the user. To solve these problems, we propose an SSR system using fEMGs measured by a few electrodes attached around the eyes of a user, which can also be easily incorporated into available VR headsets. To enhance the accuracy of classifying the fEMG signals recorded from limited recording locations relatively far from the phonatory organs, a deep neural network-based classification method was developed using similar fEMG data previously collected from other individuals and then transformed by dynamic positional warping. In the experiments, the proposed SSR system could classify six different fEMG patterns generated by six silently spoken words with an accuracy of 92.53%. To further demonstrate that our SSR system can be used as a hands-free control interface in practical VR applications, an online SSR system was implemented.
Files in This Item
Go to Link
Appears in
Collections
ETC > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Im, Chang Hwan photo

Im, Chang Hwan
COLLEGE OF ENGINEERING (서울 바이오메디컬공학전공)
Read more

Altmetrics

Total Views & Downloads

BROWSE