TEMPORAL CODING OF SPEECH IN HUMAN AUDITORY CORTEX

dc.contributor.advisorSimon, Jonathan Zen_US
dc.contributor.authorDing, Naien_US
dc.contributor.departmentElectrical Engineeringen_US
dc.contributor.publisherDigital Repository at the University of Marylanden_US
dc.contributor.publisherUniversity of Maryland (College Park, Md.)en_US
dc.date.accessioned2012-10-10T11:11:56Z
dc.date.available2012-10-10T11:11:56Z
dc.date.issued2012en_US
dc.description.abstractHuman listeners can reliably recognize speech in complex listening environments. The underlying neural mechanisms, however, remain unclear and cannot yet be emulated by any artificial system. In this dissertation, we study how speech is represented in the human auditory cortex and how the neural representation contributes to reliable speech recognition. Cortical activity from normal hearing human subjects is noninvasively recorded using magnetoencephalography, during natural speech listening. It is first demonstrated that neural activity from auditory cortex is precisely synchronized to the slow temporal modulations of speech, when the speech signal is presented in a quiet listening environment. How this neural representation is affected by acoustic interference is then investigated. Acoustic interference degrades speech perception via two mechanisms, informational masking and energetic masking, which are addressed respectively by using a competing speech stream and a stationary noise as the interfering sound. When two speech streams are presented simultaneously, cortical activity is predominantly synchronized to the speech stream the listener attends to, even if the unattended, competing speech stream is 8 dB more intense. When speech is presented together with spectrally matched stationary noise, cortical activity remains precisely synchronized to the temporal modulations of speech until the noise is 9 dB more intense. Critically, the accuracy of neural synchronization to speech predicts how well individual listeners can understand speech in noise. Further analysis reveals that two neural sources contribute to speech synchronized cortical activity, one with a shorter response latency of about 50 ms and the other with a longer response latency of about 100 ms. The longer-latency component, but not the shorter-latency component, shows selectivity to the attended speech and invariance to background noise, indicating a transition from encoding the acoustic scene to encoding the behaviorally important auditory object, in auditory cortex. Taken together, we have demonstrated that during natural speech comprehension, neural activity in the human auditory cortex is precisely synchronized to the slow temporal modulations of speech. This neural synchronization is robust to acoustic interference, whether speech or noise, and therefore provides a strong candidate for the neural basis of acoustic background invariant speech recognition.en_US
dc.identifier.urihttp://hdl.handle.net/1903/12988
dc.subject.pqcontrolledEngineeringen_US
dc.subject.pqcontrolledElectrical engineeringen_US
dc.subject.pqcontrolledNeurosciencesen_US
dc.subject.pquncontrolledauditory scene analysisen_US
dc.subject.pquncontrolledhuman auditory cortexen_US
dc.subject.pquncontrolledmagnetoencephalography (MEG)en_US
dc.subject.pquncontrolledspectro-temporal response function (STRF)en_US
dc.subject.pquncontrolledspeechen_US
dc.subject.pquncontrolledtemporal processingen_US
dc.titleTEMPORAL CODING OF SPEECH IN HUMAN AUDITORY CORTEXen_US
dc.typeDissertationen_US

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Ding_umd_0117E_13274.pdf
Size:
5.03 MB
Format:
Adobe Portable Document Format