DocumentCode
2153226
Title
Robust representations of cortical speech and language information
Author
Baker, Janet M. ; Chan, Alexander M. ; Marinkovic, Ksenija ; Halgren, Eric ; Cash, Sydney S.
Author_Institution
Med. Sch., Dept. of Otology & Laryngology, Harvard Univ., Cambridge, MA, USA
fYear
2011
fDate
22-27 May 2011
Firstpage
785
Lastpage
788
Abstract
Cortical recordings with high temporal resolution enable the tracking of neuronal excitation in response to stimuli. Here intra and extracranial recordings are analyzed from experiments presenting varied speech and language stimuli to human subjects. These studies demonstrate that information about speech and language is widely distributed across the brain, both spatially and temporally. Analyses using machine learning techniques can be used to track the space and time-course of performance in recognizing different words (83% on 10 spoken words), semantic categories (76% on 2 categories), etc.
Keywords
learning (artificial intelligence); medical signal processing; neurophysiology; signal representation; signal resolution; speech recognition; brain; cortical recording; extracranial recording; intracranial recording; language information representation; machine learning technique; neuronal excitation tracking; robust cortical speech representation; speech recognition; temporal resolution; Accuracy; Decoding; Electroencephalography; Machine learning; Semantics; Speech; Support vector machines; brain; categorization; machine learning; semantics; speech recognition;
fLanguage
English
Publisher
ieee
Conference_Titel
Acoustics, Speech and Signal Processing (ICASSP), 2011 IEEE International Conference on
Conference_Location
Prague
ISSN
1520-6149
Print_ISBN
978-1-4577-0538-0
Electronic_ISBN
1520-6149
Type
conf
DOI
10.1109/ICASSP.2011.5946521
Filename
5946521
Link To Document