Title :
Bayesian learning for models of human speech perception
Author :
Hasegawa-Johnson, Mark
Author_Institution :
Dept of ECE, Illinois Univ., IL, USA
fDate :
28 Sept.-1 Oct. 2003
Abstract :
Human speech recognition error rates are 30 times lower than machine error rates. Psychophysical experiments have pinpointed a number of specific human behaviors that may contribute to accurate speech recognition, but previous attempts to incorporate such behaviors into automatic speech recognition have often failed because the resulting models could not be easily trained from data. This paper describes Bayesian learning methods for computational models for human speech perception. Specifically, the linked computational models proposed in this paper seek to imitate the following human behaviors: independence of distinctive feature errors, perceptual magnet effect, the vowel sequence illusion, sensitivity to energy onsets and offsets, and redundant use of asynchronous acoustic correlates. The proposed models differ from many previous computational psychological models in that the desired behavior is learned from data, using a constrained optimization algorithm (the EM algorithm), rather than being coded into the model as a series of fixed rules.
Keywords :
Bayes methods; human factors; learning (artificial intelligence); optimisation; speech recognition; Bayesian learning; asynchronous acoustic correlates; constrained optimization algorithm; distinctive feature errors independence; human speech perception; perceptual magnet effect; psychophysical experiments; speech recognition; vowel sequence illusion; Automatic speech recognition; Bayesian methods; Computational modeling; Error analysis; Humans; Mathematical model; Psychology; Signal processing algorithms; Speech processing; Speech recognition;
Conference_Titel :
Statistical Signal Processing, 2003 IEEE Workshop on
Print_ISBN :
0-7803-7997-7
DOI :
10.1109/SSP.2003.1289432