Exploiting Visual Information in Automatic Speech Processing. Aleksic, P. S., Potamianos, G., & Katsaggelos, A. K. In Handbook of Image and Video Processing, pages 1263–XXXIX. Elsevier, 2005.
Exploiting Visual Information in Automatic Speech Processing [link]Paper  doi  abstract   bibtex   
This chapter focuses on how the joint processing of visual and audio signals, both generated by a talking person, can provide valuable speech information to benefit a number of audiovisual speech processing applications crucial to human-computer interactions. The analysis of visual signals has been done followed by a description of various possible ways of representing and extracting the speech information available in them. It has been shown in the chapter that the obtained visual features can complement features extracted from the acoustic signal and that the two modality representations can be fused together to allow joint audiovisual speech processing. The general bimodal integration framework is subsequently applied to three problems-automatic speech recognition, talking face synthesis, and speaker identification and authentication. In all three cases, issues specific to the particular application have been discussed, several relevant systems that have been reported in the literature have been reviewed, and the results using the implementations developed at IBM Research and Northwestern University have been presented. © 2005 Elsevier Inc. All rights reserved.
@incollection{Petar2005,
abstract = {This chapter focuses on how the joint processing of visual and audio signals, both generated by a talking person, can provide valuable speech information to benefit a number of audiovisual speech processing applications crucial to human-computer interactions. The analysis of visual signals has been done followed by a description of various possible ways of representing and extracting the speech information available in them. It has been shown in the chapter that the obtained visual features can complement features extracted from the acoustic signal and that the two modality representations can be fused together to allow joint audiovisual speech processing. The general bimodal integration framework is subsequently applied to three problems-automatic speech recognition, talking face synthesis, and speaker identification and authentication. In all three cases, issues specific to the particular application have been discussed, several relevant systems that have been reported in the literature have been reviewed, and the results using the implementations developed at IBM Research and Northwestern University have been presented. {\textcopyright} 2005 Elsevier Inc. All rights reserved.},
author = {Aleksic, Petar S. and Potamianos, Gerasimos and Katsaggelos, Aggelos K.},
booktitle = {Handbook of Image and Video Processing},
doi = {10.1016/B978-012119792-6/50134-0},
isbn = {9780121197926},
pages = {1263--XXXIX},
publisher = {Elsevier},
title = {{Exploiting Visual Information in Automatic Speech Processing}},
url = {https://linkinghub.elsevier.com/retrieve/pii/B9780121197926501340},
year = {2005}
}

Downloads: 0