Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Skip to main content

FAVS: 3D Facial Animation According to Vietnamese Semantic Analysis

  • Conference paper
  • First Online:
Advances in Information and Communication Technology (ICTA 2023)

Part of the book series: Lecture Notes in Networks and Systems ((LNNS,volume 848))

  • 295 Accesses

Abstract

There have recently been many studies on face animation according to sound, but facial expressions have not yet accurately represented and clarified the semantic meaning of the text. Studies show that characters need to represent at least six basic emotions: happy, sad, fear, disgust, anger, surprise. However, creating for facial animation for virtual characters is time-consuming and requires high creativity. The main objective of this study is to create Facial Animations according to Vietnamese Semantics (FAVS) more easily. The method is based on the important numerical blendshapes of the 3D model. The input text after predicting the emotion will be passed to the lips synchronous and emotion animating to perform the 3D face animation. Do the comparison with 2 methods: create animation by direct control with real human face via webcam and using keyframe methods. Assess the emotional expression of 3D characters according to the above 3 approaches. Survey respondents were asked to recognize a 3D Virtually sensitive emotional pattern generated for each sentence of text input and give a confidence score for each sentence. Survey results show, negative emotions are the most recognizable, happy and excited are easily confused.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Ekman, P.: Are there basic emotions. Psychol. Rev. 3(99), 550–553 (1992)

    Article  Google Scholar 

  2. Cowen, A.S., Keltner, D.: Self—report captures 27 distinct categories of emotion bridged by continuous gradients. Psychol. Cogn. Sci. 38(114), 7900–7909 (2017)

    Google Scholar 

  3. Gungor, T., Celik, K.: A comprehensive analysis of using semantic information in text categorization. In: IEEE INISTA, pp. 1–5 (2013)

    Google Scholar 

  4. Goddard, C.: Semantic Analysis: A Practical Introduction. Oxford University Press (1998)

    Google Scholar 

  5. ParallelDots, https://komprehend.io/emotion-analysis. Accessed 21 June 2023

  6. Chen, C., Crivelli, C., Garrod, O.G.B., Schyns, P.G., Fernandez-Dols, J.M., Jack, R.E.: Distinct facial expressions represent pain and pleasure across cultures. Proc. Natl. Acad. Sci. 43(115), E10013–E10021 (2018)

    Google Scholar 

  7. Yu, H., Garrod, O.G.B., Schyns, P.G.: Perception-driven facial expression synthesis. Comput. Graph.. Graph. 3(36), 152–162 (2012)

    Article  Google Scholar 

  8. Dahmani, S., Colotte, V., Girard, V., Ouni, S.: Conditional variational auto-encoder for text-driven expressive audio visual speech synthesis. In: INTERSPEECH 2019-20th (2019)

    Google Scholar 

  9. Karras, T., Aila, T., Laine, S., Herva, A., Lehtinen, J.: Audio-driven facial animation by joint end-to-end learning of pose and emotion. ACM Trans. Graph. 4(36), 1–12 (2017)

    Article  Google Scholar 

  10. Tang, H., Fu, Y., Tu, J., Huang, T.S., Hasegawa-Johnson, M.: Eava: a 3d emotive audio-visual avatar. In: IEEE Workshop on Applications of Computer Vision, pp. 1–6 (2008)

    Google Scholar 

  11. Chen, C., Hensel, L.B., Duan, Y., Ince, R.A.A., Garrod, O.G.B., Beskow, J., Jack, R.E., Schyns, P.G.: Equipping social robots with culturally sensitive facial expressions of emotion using data-driven methods. In: 14th IEEE FG 2019, pp. 1–8 (2019)

    Google Scholar 

  12. Tsapatsoulis, N., Raouzaiou, A., Kollias, S.D., Cowie, R.I.D., Douglas-Cowie, E.: Emotion Recognition and Synthesis Based on MPEG-4 FAPs. Wiley (2002)

    Google Scholar 

  13. Liu, M., Duan, Y., Ince, R.A.A., Chen, C., Garrod, O.G.B., Schyns, P.G., Jack, R.E.: Building a generative space of facial expressions of emotions using psychological data driven methods. In: IVA ‘20, pp. 1–3 (2020)

    Google Scholar 

  14. Jack, R.E., Schyns, P.G.: Toward a social psychophysics of face communication. Annu. Rev. Psychol.. Rev. Psychol. 68, 269–297 (2017)

    Article  Google Scholar 

  15. Ahumada Jr, A., Lovell, J.: Stimulus features in signal detection. J. Acoust. Soc. Am. 49(6B), 1751–1756 (1971)

    Google Scholar 

  16. Ekman, P.: Measuring facial movement. Environ. Psychol. Nonverbal Behav. 1, 56–75 (1976)

    Article  Google Scholar 

  17. Li, X., Wu, Z., Meng, H.M., Jia, J.: Expressive speech driven talking avatar synthesis with DBLSTM using limited amount of emotional bimodal data. In: Interspeech, pp. 1477–1481 (2016)

    Google Scholar 

Download references

Acknowledgement

The article is the product of the project code ĐH2022-TN07-01 funded by the University of Information and Communication Technology.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Do Thi Chi .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Chi, D.T., Thai, L.S., Van Huan, N. (2024). FAVS: 3D Facial Animation According to Vietnamese Semantic Analysis. In: Nghia, P.T., Thai, V.D., Thuy, N.T., Son, L.H., Huynh, VN. (eds) Advances in Information and Communication Technology. ICTA 2023. Lecture Notes in Networks and Systems, vol 848. Springer, Cham. https://doi.org/10.1007/978-3-031-50818-9_29

Download citation

Publish with us

Policies and ethics