Text-driven Speech Animation with Emotion Control
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Chae, Wonseok | - |
dc.contributor.author | Kim, Yejin | - |
dc.date.available | 2021-03-17T06:51:54Z | - |
dc.date.created | 2021-02-26 | - |
dc.date.issued | 2020-08 | - |
dc.identifier.issn | 1976-7277 | - |
dc.identifier.uri | https://scholarworks.bwise.kr/hongik/handle/2020.sw.hongik/11604 | - |
dc.description.abstract | In this paper, we present a new approach to creating speech animation with emotional expressions using a small set of example models. To generate realistic facial animation, two example models called key visemes and expressions are used for lip-synchronization and facial expressions, respectively. The key visemes represent lip shapes of phonemes such as vowels and consonants while the key expressions represent basic emotions of a face. Our approach utilizes a text-to-speech (TTS) system to create a phonetic transcript for the speech animation. Based on a phonetic transcript, a sequence of speech animation is synthesized by interpolating the corresponding sequence of key visemes. Using an input parameter vector, the key expressions are blended by a method of scattered data interpolation. During the synthesizing process, an importance-based scheme is introduced to combine both lip-synchronization and facial expressions into one animation sequence in real time (over 120Hz). The proposed approach can be applied to diverse types of digital content and applications that use facial animation with high accuracy (over 90%) in speech recognition. | - |
dc.publisher | KSII-KOR SOC INTERNET INFORMATION | - |
dc.title | Text-driven Speech Animation with Emotion Control | - |
dc.type | Article | - |
dc.contributor.affiliatedAuthor | Kim, Yejin | - |
dc.identifier.doi | 10.3837/tiis.2020.08.018 | - |
dc.identifier.scopusid | 2-s2.0-85090490016 | - |
dc.identifier.wosid | 000567842500018 | - |
dc.identifier.bibliographicCitation | KSII TRANSACTIONS ON INTERNET AND INFORMATION SYSTEMS, v.14, no.8, pp.3473 - 3487 | - |
dc.relation.isPartOf | KSII TRANSACTIONS ON INTERNET AND INFORMATION SYSTEMS | - |
dc.citation.title | KSII TRANSACTIONS ON INTERNET AND INFORMATION SYSTEMS | - |
dc.citation.volume | 14 | - |
dc.citation.number | 8 | - |
dc.citation.startPage | 3473 | - |
dc.citation.endPage | 3487 | - |
dc.type.rims | ART | - |
dc.type.docType | Article | - |
dc.description.journalClass | 1 | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | scopus | - |
dc.description.journalRegisteredClass | kci | - |
dc.relation.journalResearchArea | Computer Science | - |
dc.relation.journalResearchArea | Telecommunications | - |
dc.relation.journalWebOfScienceCategory | Computer Science, Information Systems | - |
dc.relation.journalWebOfScienceCategory | Telecommunications | - |
dc.subject.keywordAuthor | Speech animation | - |
dc.subject.keywordAuthor | lip-synchronization | - |
dc.subject.keywordAuthor | emotional expressions | - |
dc.subject.keywordAuthor | facial expression synthesis | - |
dc.subject.keywordAuthor | example models | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
94, Wausan-ro, Mapo-gu, Seoul, 04066, Korea02-320-1314
COPYRIGHT 2020 HONGIK UNIVERSITY. ALL RIGHTS RESERVED.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.