Audiovisual Speech Synthesis Based on Hidden Markov Models (2014)
Abstract / truncated to 115 words
In this dissertation, new methods for audiovisual speech synthesis using Hidden Markov Models (HMMs) are presented and their properties are investigated. The problem of audiovisual speech synthesis is to computationally generate both audible speech as well as a matching facial animation or video (a “visual speech signal”) for any given input text. This results in “talking heads” that can read any text to a user, with applications ranging from virtual agents in human-computer interaction to characters in animated films and computer games. For recording and playback of facial motion, an optical marker-based facial motion capturing hardware system and 3D animation software are employed, which represent the state of the art in the animation industry. For ...
speech synthesis – audiovisual speech synthesis – hidden markov models
Information
- Author
- Schabus, Dietmar
- Institution
- Graz University of Technology, Signal Processing and Speech Communication Laboratory
- Supervisors
- Publication Year
- 2014
- Upload Date
- Sept. 2, 2025
The current layout is optimized for mobile phones. Page previews, thumbnails, and full abstracts will remain hidden until the browser window grows in width.
The current layout is optimized for tablet devices. Page previews and some thumbnails will remain hidden until the browser window grows in width.