Abstract
The goal of this work is to use phonetic recognition to drive a synthetic image with speech. Phonetic units are identified by the phonetic recognition engine and mapped to mouth gestures, known as visemes, the visual counterpart of phonemes. The acoustic waveform and visemes are then sent to a synthetic image player, called FaceMe! where they are rendered synchronously. This paper provides background for the core technologies involved in this process and describes asynchronous and synchronous prototypes of a combined phonetic recognition/FaceMe! system which we use to render mouth gestures on an animated face.
Original language | English |
---|---|
Pages | 1995-1998 |
Number of pages | 4 |
State | Published - 1997 |
Externally published | Yes |
Event | 5th European Conference on Speech Communication and Technology, EUROSPEECH 1997 - Rhodes, Greece Duration: 22 Sep 1997 → 25 Sep 1997 |
Conference
Conference | 5th European Conference on Speech Communication and Technology, EUROSPEECH 1997 |
---|---|
Country/Territory | Greece |
City | Rhodes |
Period | 22/09/97 → 25/09/97 |
Bibliographical note
Publisher Copyright:© 1997 5th European Conference on Speech Communication and Technology, EUROSPEECH 1997. All rights reserved.