Gesture Generation with Low-Dimensional Embeddings

Chiu, C.-C. and Marsella, S. (2014) Gesture Generation with Low-Dimensional Embeddings. In: 2014 International Conference on Autonomous Agents and Multi-agent Systems (AAMAS '14), Paris, France, 05-09 May 2014, pp. 781-788. ISBN 9781450327381

Full text not currently available from Enlighten.

Publisher's URL:


There is a growing demand for embodied agents capable of engaging in face-to-face dialog using the same verbal and nonverbal behavior that people use. The focus of our work is generating coverbal hand gestures for these agents, gestures coupled to the content and timing of speech. A common approach to achieve this is to use motion capture of an actor or hand-crafted animations for each utterance. An alternative machine learning approach that saves development effort is to learn a general gesture controller that can generate behavior for novel utterances. However learning a direct mapping from speech to gesture movement faces the complexity of inferring the relation between the two time series of speech and gesture motion. We present a novel machine learning approach that decomposes the overall learning problem into learning two mappings: from speech to a gestural annotation and from gestural annotation to gesture motion. The combined model learns to synthesize natural gesture animation from speech audio. We assess the quality of generated animations by comparing them with the result generated by a previous approach that learns a direct mapping. Results from a human subject study show that our framework is perceived to be significantly better.

Item Type:Conference Proceedings
Glasgow Author(s) Enlighten ID:Marsella, Professor Stacy
Authors: Chiu, C.-C., and Marsella, S.
College/School:College of Medical Veterinary and Life Sciences > School of Psychology & Neuroscience

University Staff: Request a correction | Enlighten Editors: Update this record