DocumentCode :
1551540
Title :
Perceptual Confusions Among Consonants, Revisited—Cross-Spectral Integration of Phonetic-Feature Information and Consonant Recognition
Author :
Christiansen, Thomas U. ; Greenberg, Steven
Author_Institution :
Dept. of Electr. Eng., Tech. Univ. of Denmark, Lyngby, Denmark
Volume :
20
Issue :
1
fYear :
2012
Firstpage :
147
Lastpage :
161
Abstract :
The perceptual basis of consonant recognition was experimentally investigated through a study of how information associated with phonetic features (Voicing, Manner, and Place of Articulation) combines across the acoustic-frequency spectrum. The speech signals, 11 Danish consonants embedded in Consonant + Vowel + Liquid syllables, were partitioned into 3/4-octave bands (“slits”) centered at 750 Hz, 1500 Hz, and 3000 Hz, and presented individually and in two- or three-slit combinations. The amount of information transmitted (IT) was calculated from consonant-confusion matrices for each feature and slit combination. The growth of IT was measured as a function of the number of slits presented and their center frequency for the phonetic features and consonants. The IT associated with Voicing, Manner, and Consonants sums nearly linearly for two-band stimuli irrespective of their center frequency. Adding a third band increases the IT by an amount somewhat less than predicted by linear cross-spectral integration (i.e., a compressive function). In contrast, for Place of Articulation, the IT gained through addition of a second or third slit is far more than predicted by linear, cross-spectral summation. This difference is mirrored in a measure of error-pattern similarity across bands-Symmetric Redundancy. Consonants, as well as Voicing and Manner, share a moderate degree of redundancy between bands. In contrast, the cross-spectral redundancy associated with Place is close to zero, which means the bands are essentially independent in terms of decoding this feature. Because consonant recognition and Place decoding are highly correlated (correlation coefficient r2 = 0.99), these results imply that the auditory processes underlying consonant recognition are not strictly linear. This may account for why conventional cross-spectral integration speech models, such as the Articulation Index, Speech Intelligibility Index, and the Speech Transmission Index do- - not predict intelligibility and segment recognition well under certain conditions (e.g., discontiguous frequency bands and audio-visual speech).
Keywords :
correlation methods; natural language processing; spectral analysis; speech intelligibility; speech processing; speech recognition; Danish consonants; acoustic-frequency spectrum; articulation index; audio-visual speech; auditory processes; center frequency; compressive function; consonant + vowel + liquid syllables; consonant recognition; consonant-confusion matrices; conventional cross-spectral integration speech models; correlation coefficient; cross-spectral redundancy; cross-spectral summation; discontiguous frequency bands; error-pattern similarity across bands; information transmission; linear cross-spectral integration; linear summation; perceptual confusions; phonetic features; phonetic-feature information; place decoding; place of articulation; segment recognition; speech intelligibility index; speech signals; speech transmission index; symmetric redundancy; two-band stimuli; voicing and manner; Artificial intelligence; Indexes; Modulation; Redundancy; Signal to noise ratio; Speech; Speech recognition; Consonant recognition; cross-spectral integration; information theory; phonetic features; speech perception;
fLanguage :
English
Journal_Title :
Audio, Speech, and Language Processing, IEEE Transactions on
Publisher :
ieee
ISSN :
1558-7916
Type :
jour
DOI :
10.1109/TASL.2011.2159202
Filename :
5872010
Link To Document :
بازگشت