On the jointly unsupervised feature vector normalization and acoustic model compensation for robust speech recognition. Buera, L.; Miguel, A.; Lleida, E.; Saz, Ó.; and Ortega, A. In Interspeech 2007. Proceedings of the 8th Annual Conference of the International Speech Communication Association, pages 1046-1049. Antwerp, Belgium, August 27-31, 2007.
On the jointly unsupervised feature vector normalization and acoustic model compensation for robust speech recognition [link]Paper  abstract   bibtex   
To compensate the mismatch between training and testing conditions, an unsupervised hybrid compensation technique is proposed. It combines Multi-Environment Model based LInear Normalization (MEMLIN) with a novel acoustic model adaptation method based on rotation transformations. A set of rotation transformations is estimated between clean and MEMLIN-normalized data by linear regression in a training process. Thus, each MEMLIN-normalized frame is decoded using the expanded acoustic models, which are obtained from the reference ones and the set of rotation transformations. During the search algorithm, one of the rotation transformations is on-line selected for each frame according to the ML criterion in a modified Viterbi algorithm. Some experiments with Spanish SpeechDat Car database were carried out. MEMLIN over standard ETSI front-end parameters reaches 75.53% of mean improvement in WER, while the introduced hybrid solution goes up to 90.54%.
@incollection{buera_jointly_2007,
	Author = {Buera, Luis and Miguel, Antonio and Lleida, Eduardo and Saz, Óscar and Ortega, Alfonso},
	Booktitle = {Interspeech 2007. Proceedings of the 8th Annual Conference of the International Speech Communication Association},
	Date = {2007},
	Date-Modified = {2016-09-24 18:56:00 +0000},
	File = {Attachment:files/1663/Buera et al. - 2007 - On the jointly unsupervised feature vector normalization and acoustic model compensation for robust speech recogni.pdf:application/pdf},
	Keywords = {Spanish, speech recognition, speech technology},
	Pages = {1046-1049},
	Publisher = {Antwerp, Belgium, August 27-31, 2007},
	Title = {On the jointly unsupervised feature vector normalization and acoustic model compensation for robust speech recognition},
	Url = {http://www.isca-speech.org/archive/interspeech_2007/i07_1046.html},
	Abstract = {To compensate the mismatch between training and testing conditions, an unsupervised hybrid compensation technique is proposed. It combines Multi-Environment Model based LInear Normalization (MEMLIN) with a novel acoustic model adaptation method based on rotation transformations. A set of rotation transformations is estimated between clean and MEMLIN-normalized data by linear regression in a training process. Thus, each MEMLIN-normalized frame is decoded using the expanded acoustic models, which are obtained from the reference ones and the set of rotation transformations. During the search algorithm, one of the rotation transformations is on-line selected for each frame according to the ML criterion in a modified Viterbi algorithm. Some experiments with Spanish SpeechDat Car database were carried out. MEMLIN over standard ETSI front-end parameters reaches 75.53\% of mean improvement in WER, while the introduced hybrid solution goes up to 90.54\%.},
	Bdsk-File-1 = {YnBsaXN0MDDUAQIDBAUGJCVYJHZlcnNpb25YJG9iamVjdHNZJGFyY2hpdmVyVCR0b3ASAAGGoKgHCBMUFRYaIVUkbnVsbNMJCgsMDxJXTlMua2V5c1pOUy5vYmplY3RzViRjbGFzc6INDoACgAOiEBGABIAFgAdccmVsYXRpdmVQYXRoWWFsaWFzRGF0YV8QXy4uLy4uLy4uL0JpYmxpb2dyYWZpYS9QYXBlcnMvQnVlcmEvT24gdGhlIGpvaW50bHkgdW5zdXBlcnZpc2VkIGZlYXR1cmUgdmVjdG9yIG5vcm1hbGl6YXRpb24ucGRm0hcLGBlXTlMuZGF0YU8RAlYAAAAAAlYAAgAADE1hY2ludG9zaCBIRAAAAAAAAAAAAAAAAAAAAMv2H85IKwAAEIZoKh9PbiB0aGUgam9pbnRseSB1bnMjMTA4NjY4MkIucGRmAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAQhmgr1AnTEgAAAAAAAAAAAAMABAAACSAAAAAAAAAAAAAAAAAAAAAFQnVlcmEAABAACAAAy/YDrgAAABEACAAA1Am28gAAAAEAFBCGaCoQhmWOAAX8RwAF+5gAAMBGAAIAY01hY2ludG9zaCBIRDpVc2VyczoAam9hcXVpbV9sbGlzdGVycmk6AEJpYmxpb2dyYWZpYToAUGFwZXJzOgBCdWVyYToAT24gdGhlIGpvaW50bHkgdW5zIzEwODY2ODJCLnBkZgAADgB6ADwATwBuACAAdABoAGUAIABqAG8AaQBuAHQAbAB5ACAAdQBuAHMAdQBwAGUAcgB2AGkAcwBlAGQAIABmAGUAYQB0AHUAcgBlACAAdgBlAGMAdABvAHIAIABuAG8AcgBtAGEAbABpAHoAYQB0AGkAbwBuAC4AcABkAGYADwAaAAwATQBhAGMAaQBuAHQAbwBzAGgAIABIAEQAEgBuVXNlcnMvam9hcXVpbV9sbGlzdGVycmkvQmlibGlvZ3JhZmlhL1BhcGVycy9CdWVyYS9PbiB0aGUgam9pbnRseSB1bnN1cGVydmlzZWQgZmVhdHVyZSB2ZWN0b3Igbm9ybWFsaXphdGlvbi5wZGYAEwABLwAAFQACABj//wAAgAbSGxwdHlokY2xhc3NuYW1lWCRjbGFzc2VzXU5TTXV0YWJsZURhdGGjHR8gVk5TRGF0YVhOU09iamVjdNIbHCIjXE5TRGljdGlvbmFyeaIiIF8QD05TS2V5ZWRBcmNoaXZlctEmJ1Ryb290gAEACAARABoAIwAtADIANwBAAEYATQBVAGAAZwBqAGwAbgBxAHMAdQB3AIQAjgDwAPUA/QNXA1kDXgNpA3IDgAOEA4sDlAOZA6YDqQO7A74DwwAAAAAAAAIBAAAAAAAAACgAAAAAAAAAAAAAAAAAAAPF},
	Bdsk-Url-1 = {http://www.isca-speech.org/archive/interspeech_2007/i07_1046.html}}
Downloads: 0