Electrical & Computer Engineering
Permanent URI for this communityhttp://hdl.handle.net/1903/2234
Browse
3 results
Search Results
Item CORTICAL REPRESENTATIONS OF INTELLIGIBLE AND UNINTELLIGIBLE SPEECH: EFFECTS OF AGING AND LINGUISTIC CONTENT(2023) Karunathilake , I.M Dushyanthi; Simon, Jonathan Z.; Electrical Engineering; Digital Repository at the University of Maryland; University of Maryland (College Park, Md.)Speech communication requires real-time processing of rapidly varying acoustic sounds across various speech landmarks while recruiting complex cognitive processes to derive the intended meaning. Behavioral studies have highlighted that speech comprehension is altered by factors like aging, linguistic content, and intelligibility, yet the systematic neural mechanisms underlying these changes are not well understood. This thesis aims to explore how the neural bases are modulated by each of these factors using three different experiments, by comparing speech representation in the cortical responses, measured by Magnetoencephalography (MEG). We use neural encoding (Temporal Response Functions (TRFs)) and decoding (reconstruction accuracy) models which describe the mapping between stimulus features and the cortical responses, which are instrumental in understanding cortical temporal processing mechanisms in the brain.Firstly, we investigate age-related changes in timing and fidelity of the cortical representation of speech-in-noise. Understanding speech in a noisy environment becomes more challenging with age, even for healthy aging. Our findings demonstrate that some of the age-related difficulties in understanding speech in noise experienced by older adults are accompanied by age-related temporal processing differences in the auditory cortex. This is an important step towards incorporating neural measures to both diagnostic evaluation and treatments aimed at speech comprehension problems in older adults. Next, we investigate how the cortical representation of speech is influenced by the linguistic content by comparing neural responses to four types of continuous speech-like passages: non-speech, non-words, scrambled words, and narrative. We find neural evidence for emergent features of speech processing from acoustics to linguistic processes at the sentential level as incremental steps in the processing of speech input occur. We also show the gradual computation of hierarchical speech features over time, encompassing both bottom-up and top-down mechanisms. Top-down driven mechanisms at linguistic level demonstrates N400-like response, suggesting involvement of predictive coding mechanisms. Finally, we find potential neural markers of speech intelligibility using a priming paradigm, where intelligibility is varied while keeping the acoustic structure constant. Our findings suggest that segmentation of sounds into words emerges with better speech intelligibility and most strongly at ~400 ms in prefrontal cortex (PFC), in line with engagement of top-down mechanisms associated with priming. Taken together, this thesis furthers our understanding on neural mechanisms underlying speech comprehension and potential objective neural markers to evaluate the level of speech comprehension.Item Effects of Aging on Cortical Representations of Continuous Speech(2022) Karunathilake, I.M Dushyanthi; Simon, Jonathan Z.Understanding speech in a noisy environment is crucial in day-to-day interactions, and yet becomes more challenging with age, even for healthy aging. Age-related changes in the neural mechanisms that enable speech-in-noise listening have been investigated previously; however, the extent to which age affects the timing and fidelity of encoding of target and interfering speech streams are not well understood. Using magnetoencephalography (MEG), we investigated how continuous speech is represented in auditory cortex in the presence of interfering speech, in younger and older adults. Cortical representations were obtained from neural responses that time-locked to the speech envelopes using speech envelope reconstruction and temporal response functions (TRFs). TRFs showed three prominent peaks corresponding to auditory cortical processing stages: early (~50 ms), middle (~100 ms) and late (~200 ms). Older adults showed exaggerated speech envelope representations compared to younger adults. Temporal analysis revealed both that the age-related exaggeration starts as early as ~50 ms, and that older adults needed a substantially longer integration time window to achieve their better reconstruction of the speech envelope. As expected, with increased speech masking, envelope reconstruction for the attended talker decreased and all three TRF peaks were delayed, with aging contributing additionally to the reduction. Interestingly, for older adults the late peak was delayed, suggesting that this late peak may receive contributions from multiple sources. Together these results suggest that there are several mechanisms at play compensating for age-related temporal processing deficits at several stages, but which are not able to fully reestablish unimpaired speech perception.Item Cortical Processing of Arithmetic and Simple Sentences in an Auditory Attention Task - Dataset(2021) Kulasingham, Joshua P.; Joshi, Neha H.; Rezaeizadeh, Mohsen; Simon, Jonathan Z.Cortical processing of arithmetic and of language rely on both shared and task-specific neural mechanisms, which should also be dissociable from the particular sensory modality used to probe them. Here, spoken arithmetical and non-mathematical statements were employed to investigate neural processing of arithmetic, compared to general language processing, in an attention-modulated cocktail party paradigm. Magnetoencephalography (MEG) data were recorded from 22 human subjects listening to audio mixtures of spoken sentences and arithmetic equations while selectively attending to one of the two speech streams. Short sentences and simple equations were presented diotically at fixed and distinct word/symbol and sentence/equation rates. Critically, this allowed neural responses to acoustics, words, and symbols to be dissociated from responses to sentences and equations. Indeed, the simultaneous neural processing of the acoustics of words and symbols were observed in auditory cortex for both streams. Neural responses to sentences and equations, however, were predominantly to the attended stream, originating primarily from left temporal, and parietal areas, respectively. Additionally, these neural responses were correlated with behavioral performance in a deviant detection task. Source-localized Temporal Response Functions revealed distinct cortical dynamics of responses to sentences in left temporal areas and equations in bilateral temporal, parietal, and motor areas. Finally, the target of attention could be decoded from MEG responses, especially in left superior parietal areas. In short, the neural responses to arithmetic and language are especially well segregated during the cocktail party paradigm, and the correlation with behavior suggests that they may be linked to successful comprehension or calculation.