Sahandi, R., Vine, D. S. G. and Longster, J. A., 1998. Text-to-Visual Speech Synthesis. Informatica, Special issue: NLP & Multi-Agent Systems, 22 (4), pp. 445-450.
Full text not available from this repository.
The development of interactive multimedia systems, coupled with the advances in computer technology and high-speed communication systems, has made it possible for information to be presented to users more effectively and efficiently. Man-machine communication can be enhanced via the use of synthesised speech and computer animation in multimedia systems. Whilst synthetic speech is potentially a more natural communication medium, it can be improved by the addition of an animated human face synchronised with the synthetic speech. This facial display provides a number of visual cues relating to what the speaker is saying and the speaker's emotional state. This increases intelligibility if the synthetic speech is degraded with noise, and allows knowledge transfer for the hearing-impaired, through lip-reading. This paper provides an overview of existing speech synthesis and facial animation techniques, and discusses the limitations of each. The paper concludes with a description of a visual speech synthesis system developed at Bournemouth University, and a discussion of the audio-visual synchronisation issues.
|Uncontrolled Keywords:||Synthetic speech, Facial animation|
|Subjects:||Technology > Engineering > General Engineering|
|Group:||School of Design, Engineering & Computing > Design Simulation Research Centre|
|Deposited By:||Dr Reza Sahandi|
|Deposited On:||31 Aug 2009 17:46|
|Last Modified:||07 Mar 2013 15:12|
|Repository Staff Only -|
|BU Staff Only -|
|Help Guide -||Editing Your Items in BURO|