Abstract
There have recently been many studies on face animation according to sound, but facial expressions have not yet accurately represented and clarified the semantic meaning of the text. Studies show that characters need to represent at least six basic emotions: happy, sad, fear, disgust, anger, surprise. However, creating for facial animation for virtual characters is time-consuming and requires high creativity. The main objective of this study is to create Facial Animations according to Vietnamese Semantics (FAVS) more easily. The method is based on the important numerical blendshapes of the 3D model. The input text after predicting the emotion will be passed to the lips synchronous and emotion animating to perform the 3D face animation. Do the comparison with 2 methods: create animation by direct control with real human face via webcam and using keyframe methods. Assess the emotional expression of 3D characters according to the above 3 approaches. Survey respondents were asked to recognize a 3D Virtually sensitive emotional pattern generated for each sentence of text input and give a confidence score for each sentence. Survey results show, negative emotions are the most recognizable, happy and excited are easily confused.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Ekman, P.: Are there basic emotions. Psychol. Rev. 3(99), 550–553 (1992)
Cowen, A.S., Keltner, D.: Self—report captures 27 distinct categories of emotion bridged by continuous gradients. Psychol. Cogn. Sci. 38(114), 7900–7909 (2017)
Gungor, T., Celik, K.: A comprehensive analysis of using semantic information in text categorization. In: IEEE INISTA, pp. 1–5 (2013)
Goddard, C.: Semantic Analysis: A Practical Introduction. Oxford University Press (1998)
ParallelDots, https://komprehend.io/emotion-analysis. Accessed 21 June 2023
Chen, C., Crivelli, C., Garrod, O.G.B., Schyns, P.G., Fernandez-Dols, J.M., Jack, R.E.: Distinct facial expressions represent pain and pleasure across cultures. Proc. Natl. Acad. Sci. 43(115), E10013–E10021 (2018)
Yu, H., Garrod, O.G.B., Schyns, P.G.: Perception-driven facial expression synthesis. Comput. Graph.. Graph. 3(36), 152–162 (2012)
Dahmani, S., Colotte, V., Girard, V., Ouni, S.: Conditional variational auto-encoder for text-driven expressive audio visual speech synthesis. In: INTERSPEECH 2019-20th (2019)
Karras, T., Aila, T., Laine, S., Herva, A., Lehtinen, J.: Audio-driven facial animation by joint end-to-end learning of pose and emotion. ACM Trans. Graph. 4(36), 1–12 (2017)
Tang, H., Fu, Y., Tu, J., Huang, T.S., Hasegawa-Johnson, M.: Eava: a 3d emotive audio-visual avatar. In: IEEE Workshop on Applications of Computer Vision, pp. 1–6 (2008)
Chen, C., Hensel, L.B., Duan, Y., Ince, R.A.A., Garrod, O.G.B., Beskow, J., Jack, R.E., Schyns, P.G.: Equipping social robots with culturally sensitive facial expressions of emotion using data-driven methods. In: 14th IEEE FG 2019, pp. 1–8 (2019)
Tsapatsoulis, N., Raouzaiou, A., Kollias, S.D., Cowie, R.I.D., Douglas-Cowie, E.: Emotion Recognition and Synthesis Based on MPEG-4 FAPs. Wiley (2002)
Liu, M., Duan, Y., Ince, R.A.A., Chen, C., Garrod, O.G.B., Schyns, P.G., Jack, R.E.: Building a generative space of facial expressions of emotions using psychological data driven methods. In: IVA ‘20, pp. 1–3 (2020)
Jack, R.E., Schyns, P.G.: Toward a social psychophysics of face communication. Annu. Rev. Psychol.. Rev. Psychol. 68, 269–297 (2017)
Ahumada Jr, A., Lovell, J.: Stimulus features in signal detection. J. Acoust. Soc. Am. 49(6B), 1751–1756 (1971)
Ekman, P.: Measuring facial movement. Environ. Psychol. Nonverbal Behav. 1, 56–75 (1976)
Li, X., Wu, Z., Meng, H.M., Jia, J.: Expressive speech driven talking avatar synthesis with DBLSTM using limited amount of emotional bimodal data. In: Interspeech, pp. 1477–1481 (2016)
Acknowledgement
The article is the product of the project code ĐH2022-TN07-01 funded by the University of Information and Communication Technology.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Chi, D.T., Thai, L.S., Van Huan, N. (2024). FAVS: 3D Facial Animation According to Vietnamese Semantic Analysis. In: Nghia, P.T., Thai, V.D., Thuy, N.T., Son, L.H., Huynh, VN. (eds) Advances in Information and Communication Technology. ICTA 2023. Lecture Notes in Networks and Systems, vol 848. Springer, Cham. https://doi.org/10.1007/978-3-031-50818-9_29
Download citation
DOI: https://doi.org/10.1007/978-3-031-50818-9_29
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-50817-2
Online ISBN: 978-3-031-50818-9
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)