Speech Driven 3D Face Animation

Kavitha Dhanushkodi *, Vasantha Kumaran A. P. **, Vignesh K. ***, Vimal Raj V. ****
*-**** Department of Computer Science and Engineering, SRM Valliammai Engineering College, Chennai, India.
Periodicity:March - May'2021
DOI : https://doi.org/10.26634/jit.10.2.18167

Abstract

When considering the problem of disabled persons, a new concept known as Speech driven 3D face animation employing deep learning and neural networks has come into use. Nonverbal behaviour cues, such as facial expressions, are still intact and provide information about what we are thinking, doing or reacting to. When it comes to computerbased movies and digital games, expressive face animation is a must-have feature. It is necessary to get audio input from the user, after which the matching characteristics of the audio are extracted. Once the expressions have been analysed, they are combined with the intermediate 3D model to complete the process. The relevant result is generated with the assistance of the neural renderer. An overview of the whole implementation is presented in this paper.

Keywords

Facial Animation, Deep Learning, Neural Networks, 3D Model.

How to Cite this Article?

Dhanushkodi, K., Kumaran, A. P. V., Vignesh, K., and Raj, V. V. (2021). Speech Driven 3D Face Animation. i-manager's Journal on Information Technology, 10(2), 13-21. https://doi.org/10.26634/jit.10.2.18167

References

[2]. Beskow, J. (1995). Rule-based visual speech synthesis. In EUROSPEECH '95. 4th European Conference on Speech Communication and Technology (pp. 299-302).
[6]. Chai, J. X., Xiao, J., & Hodgins, J. (2003, July). Visionbased control of 3Dfacial animation. In: Proceedings of the ACM SIGGRAPH/Eurographics Symposium on Computer Animation. Aire-la-Ville, Switzerland: Eurographics Association, 193−206.
[12]. Li, X., & Zhang, Z. (2008). Morphable linear fitting method for facial expression synthesis. Acta Automatica Sinica, 34(5), 593-597.
[19]. Sifakis, E., Neverov, I., & Fedkiw, R. (2005). Automatic determination of facial muscle activations from sparse motion capture marker data. ACM Transactions on Graphics, 24(3), 417−425
[22]. Waters, K., & Levergood, T. M. (1993). DECface: An automatic lip-synchronization algorithm for synthetic faces. Digital Equipment Corporation, Cambridge Research Laborartory.
If you have access to this article please login to view the article or kindly login to purchase the article

Purchase Instant Access

Single Article

North Americas,UK,
Middle East,Europe
India Rest of world
USD EUR INR USD-ROW
Online 15 15

Options for accessing this content:
  • If you would like institutional access to this content, please recommend the title to your librarian.
    Library Recommendation Form
  • If you already have i-manager's user account: Login above and proceed to purchase the article.
  • New Users: Please register, then proceed to purchase the article.