University of Surrey

Test tubes in the lab Research in the ATI Dance Research

Model-based synthesis of visual speech movements from 3D video

Edge, JD, Hilton, A and Jackson, PJB (2009) Model-based synthesis of visual speech movements from 3D video In: SIGGRAPH '09, 2009-08-03 - 2009-08-07, Louisiana, USA.

Full text not available from this repository.

Abstract

We describe a method for the synthesis of visual speech movements using a hybrid unit selection/model-based approach. Speech lip movements are captured using a 3D stereo face capture system and split up into phonetic units. A dynamic parameterisation of this data is constructed which maintains the relationship between lip shapes and velocities; within this parameterisation a model of how lips move is built and is used in the animation of visual speech movements from speech audio input. The mapping from audio parameters to lip movements is disambiguated by selecting only the most similar stored phonetic units to the target utterance during synthesis. By combining properties of model-based synthesis (e.g., HMMs, neural nets) with unit selection we improve the quality of our speech synthesis.

Item Type: Conference or Workshop Item (UNSPECIFIED)
Authors :
NameEmailORCID
Edge, JDUNSPECIFIEDUNSPECIFIED
Hilton, AUNSPECIFIEDUNSPECIFIED
Jackson, PJBp.jackson@surrey.ac.ukUNSPECIFIED
Date : 2009
Identification Number : 10.1145/1599301.1599309
Depositing User : Symplectic Elements
Date Deposited : 17 May 2017 11:18
Last Modified : 17 May 2017 14:55
URI: http://epubs.surrey.ac.uk/id/eprint/830789

Actions (login required)

View Item View Item

Downloads

Downloads per month over past year


Information about this web site

© The University of Surrey, Guildford, Surrey, GU2 7XH, United Kingdom.
+44 (0)1483 300800