Three-dimensional facial adaptation for MPEG-4 talking heads

Grammalidis, N., Sarris, N., Deligianni, F. and Strintzis, M. G. (2002) Three-dimensional facial adaptation for MPEG-4 talking heads. Eurasip Journal on Applied Signal Processing, 2002(10), pp. 1005-1020. (doi: 10.1155/S1110865702206113)

Full text not currently available from Enlighten.


This paper studies a new method for three-dimensional (3D) facial model adaptation and its integration into a text-to-speech (TTS) system. The 3D facial adaptation requires a set of two orthogonal views of the user′s face with a number of feature points located on both views. Based on the correspondences of the feature points′ positions, a generic face model is deformed nonrigidly treating every facial part as a separate entity. A cylindrical texture map is then built from the two image views. The generated head models are compared to corresponding models obtained by the commonly used adaptation method that utilizes 3D radial bases functions. The generated 3D models are integrated into a talking head system, which consists of two distinct parts: a multilingual text to speech sub-system and an MPEG-4 compliant facial animation sub-system. Support for the Greek language has been added, while preserving lip and speech synchronization.

Item Type:Articles
Glasgow Author(s) Enlighten ID:Deligianni, Dr Fani
Authors: Grammalidis, N., Sarris, N., Deligianni, F., and Strintzis, M. G.
College/School:College of Science and Engineering > School of Computing Science
Journal Name:Eurasip Journal on Applied Signal Processing

University Staff: Request a correction | Enlighten Editors: Update this record