Speech-to-video synthesis using facial animation parameters. Aleksic, P. & Katsaggelos, A. In Proceedings 2003 International Conference on Image Processing (Cat. No.03CH37429), volume 2, pages III–1–4, 2003. IEEE.
Speech-to-video synthesis using facial animation parameters [link]Paper  doi  abstract   bibtex   
The presence of visual information in addition to audio could improve speech understanding in noisy environments. This additional information could be especially useful for people with impaired hearing who are able to speechread. This paper focuses on the problem of synthesizing the Facial Animation Parameters (FAPs), supported by the MPEG-4 standard for the visual representation of speech, from a narrowband acoustic speech (telephone) signal. A correlation Hidden Markov Model (CHMM) system for performing visual speech synthesis is proposed. The CHMM system integrates an independently trained acoustic HMM (AHMM) system and a visual HMM (VHMM) system, in order to realize speech-to-video synthesis. Objective experiments are performed by analyzing the synthesized FAPs and computing the time alignment errors. Time alignment errors are reduced by 40.5% compared to the conventional temporal scaling method.
@inproceedings{Petar2003,
abstract = {The presence of visual information in addition to audio could improve speech understanding in noisy environments. This additional information could be especially useful for people with impaired hearing who are able to speechread. This paper focuses on the problem of synthesizing the Facial Animation Parameters (FAPs), supported by the MPEG-4 standard for the visual representation of speech, from a narrowband acoustic speech (telephone) signal. A correlation Hidden Markov Model (CHMM) system for performing visual speech synthesis is proposed. The CHMM system integrates an independently trained acoustic HMM (AHMM) system and a visual HMM (VHMM) system, in order to realize speech-to-video synthesis. Objective experiments are performed by analyzing the synthesized FAPs and computing the time alignment errors. Time alignment errors are reduced by 40.5% compared to the conventional temporal scaling method.},
author = {Aleksic, P.S. and Katsaggelos, A.K.},
booktitle = {Proceedings 2003 International Conference on Image Processing (Cat. No.03CH37429)},
doi = {10.1109/ICIP.2003.1247166},
isbn = {0-7803-7750-8},
pages = {III--1--4},
publisher = {IEEE},
title = {{Speech-to-video synthesis using facial animation parameters}},
url = {http://ieeexplore.ieee.org/document/1247166/},
volume = {2},
year = {2003}
}

Downloads: 0