Theses and Dissertations from UMD

Permanent URI for this communityhttp://hdl.handle.net/1903/2

New submissions to the thesis/dissertation collections are added automatically as they are received from the Graduate School. Currently, the Graduate School deposits all theses and dissertations from a given semester after the official graduation date. This means that there may be up to a 4 month delay in the appearance of a give thesis/dissertation in DRUM

More information is available at Theses and Dissertations at University of Maryland Libraries.

Browse

Search Results

Now showing 1 - 6 of 6
  • Thumbnail Image
    Item
    Speech Segregation and Representation in the Ferret Auditory and Frontal Cortices
    (2022) Joshi, Neha Hemant; Shamma, Shihab; Electrical Engineering; Digital Repository at the University of Maryland; University of Maryland (College Park, Md.)
    The problem of separating overlapping streams of sound, and selectively attending to a sound of interest is ubiquitous in humans and animals alike as a means for survival and communication. This problem, known as the cocktail party problem, is the focus of this thesis, where we explore the neural correlates of two-speaker segregation in the auditory and frontal cortex, using the ferret as an animal model. While speech segregation has been studied extensively in humans using various non-invasive imaging as well as some restricted invasive techniques, these do not provide a way to obtain neural data at the single-unit level. In animal models, streaming studies have been limited to simple stimuli like tone streams, or sound in noise. In this thesis, we extend this work to understand how complex auditory stimuli such as human speech is encoded at the single-unit and population level in both the auditory cortex, as well as the frontal cortex of the ferret. In the first part of the thesis, we explore current literature in auditory streaming and design a behavioral task using the ferret as an animal model to perform stream segregation. We train ferrets to selectively listen to one speaker over another, and perform a task to indicate detection of the attended speaker. We show the validity of this behavioral task, and the reliability with which the animal performs this task of two speaker stream segregation. In the second part, we collect neurophysiological data which is post-processed to obtain data from single units in both the auditory cortex (the primary auditory cortex, and the secondary region which includes the dorsal posterior ectosylvian gyrus) as well as the dorsolateral aspect of the frontal cortex of the ferret. We analyse the data and present findings of how the auditory and frontal cortices encode the information required to reliably segregate the speaker of relevance from the mixture of two speakers, and the insights provided into stream segregation mechanisms and the cocktail party solved by animals using neural decoding approaches. We finally demonstrate that stream segregation has already begun at the level of the primary auditory cortex. In agreement with previous attention-modulated neural studies in the auditory cortex, we show that this stream segregation is more pronounced in the secondary cortex, where we see clear enhancement of the attended speaker, and suppression of the unattended speaker. We explore the contribution of various areas within the primary and secondary regions, and how it relates to speaker selectivity of individual neuronal units. We also study the neural encoding of top-down attention modulation in the ferret frontal cortex. Finally, we discuss the conclusions from these results in the broader context of their relevance to the field, and what future directions it may hold for the field.
  • Thumbnail Image
    Item
    Time-locked Cortical Processing of Speech in Complex Environments
    (2021) Kulasingham, Joshua Pranjeevan; Simon, Jonathan Z; Electrical Engineering; Digital Repository at the University of Maryland; University of Maryland (College Park, Md.)
    Our ability to communicate using speech depends on complex, rapid processing mechanisms in the human brain. These cortical processes make it possible for us to easily understand one another even in noisy environments. Measurements of neural activity have found that cortical responses time-lock to the acoustic and linguistic features of speech. Investigating the neural mechanisms that underlie this ability could lead to a better understanding of human cognition, language comprehension, and hearing and speech impairments. We use Magnetoencephalography (MEG), which non-invasively measures the magnetic fields that arise from neural activity, to further explore these time-locked cortical processes. One method for detecting this activity is the Temporal Response Function (TRF), which models the impulse response of the neural system to continuous stimuli. Prior work has found that TRFs reflect several stages of speech processing in the cortex. Accordingly, we use TRFs to investigate cortical processing of both low-level acoustic and high-level linguistic features of continuous speech. First, we find that cortical responses time-lock at high gamma frequencies (~100 Hz) to the acoustic envelope modulations of the low pitch segments of speech. Older and younger listeners show similar high gamma responses, even though slow envelope TRFs show age-related differences. Next, we utilize frequency domain analysis, TRFs and linear decoders to investigate cortical processing of high-level structures such as sentences and equations. We find that the cortical networks involved in arithmetic processing dissociate from those underlying language processing, although bothinvolve several overlapping areas. These processes are more separable when subjects selectively attend to one speaker over another distracting speaker. Finally, we compare both conventional and novel TRF algorithms in terms of their ability to estimate TRF components, which may provide robust measures for analyzing group and task differences in auditory and speech processing. Overall, this work provides insights into several stages of time-locked cortical processing of speech and highlights the use of TRFs for investigating neural responses to continuous speech in complex environments.
  • Thumbnail Image
    Item
    THE ACOUSTIC QUALITIES THAT INFLUENCE AUDITORY OBJECT AND EVENT RECOGNITION
    (2019) Ogg, Mattson Wallace; Slevc, L. Robert; Neuroscience and Cognitive Science; Digital Repository at the University of Maryland; University of Maryland (College Park, Md.)
    Throughout the course of a given day, human listeners encounter an immense variety of sounds in their environment. These are quickly transformed into mental representations of objects and events in the world, which guide more complex cognitive processes and behaviors. Through five experiments in this dissertation, I investigated the rapid formation of auditory object and event representations (i.e., shortly after sound onset) with a particular focus on understanding what acoustic information the auditory system uses to support this recognition process. The first three experiments analyzed behavioral (dissimilarity ratings in Experiment 1; duration-gated identification in Experiment 2) and neural (MEG decoding in Experiment 3) responses to a diverse array of natural sound recordings as a function of the acoustic qualities of the stimuli and their temporal development alongside participants’ concurrently developing responses. The findings from these studies highlight the importance of acoustic qualities related to noisiness, spectral envelope, spectrotemporal change over time, and change in fundamental frequency over time for sound recognition. Two additional studies further tested these results via syntheszied stimuli that explicitly manipulated these acoustic cues, interspersed among a new set of natural sounds. Findings from these acoustic manipulations as well as replications of my previous findings (with new stimuli and tasks) again revealed the importance of aperiodicity, spectral envelope, spectral variability and fundamental frequency in sound-category representations. Moreover, analyses of the synthesized stimuli suggested that aperiodicity is a particularly robust cue for some categories and that speech is difficult to characterize acoustically, at least based on this set of acoustic dimensions and synthesis approach. While the study of the perception of these acoustic cues has a long history, a fuller understanding of how these qualities contribute to natural auditory object recognition in humans has been difficult to glean. This is in part because behaviorally important categories of sound (studied together in this work) have previously been studied in isolation. By bringing these literatures together over these five experiments, this dissertation begins to outline a feature space that encapsulates many different behaviorally relevant sounds with dimensions related to aperiodicity, spectral envelope, spectral variability and fundamental frequency.
  • Thumbnail Image
    Item
    INTEGRATION OF INTRA-AUDITORY MODALITIES FOR THE ENHANCEMENT OF MOTOR PERFORMANCE AND COORDINATION IN A CONSTANT FORCE PRODUCTION TASK
    (2015) Koh, Kyung; Shim, Jae Kun; Kinesiology; Digital Repository at the University of Maryland; University of Maryland (College Park, Md.)
    One of most fundamental problems in the field of neuromechanics is to understand how the central nervous system (CNS) integrates multiple sources of sensory information and coordinates multiple effectors in human movement. Much attention has been directed to the integration of multiple modalities between sensory organs (e.g., visual and auditory, visual and tactile, or visual and proprioceptor), while little is known about the integration of multiple modalities within one sensory (i.e., intra-sensory integration), especially regarding the auditory sensory. This dissertation investigated the mechanisms of intra-auditory integration for the control of multiple fingers during constant force production tasks, specifically regarding how the CNS utilizes multiple sources in auditory feedback, how the CNS deals with uncertainty in auditory feedback, and how the CNS adapts or learns a motor task using auditory feedback. The specific aims of this dissertation included: 1) development of analytical tools for the quantification of motor performance and coordination in a hierarchical structure of motor variability; 2) investigation into the effect of intra-auditory integration on motor performance and coordination (Experiment I); 3) investigation of the role of uncertainty in auditory information on the effectiveness of intra-auditory integration in motor performance and coordination (Experiment II); and 4) investigation of the auditory-motor learning in the context of motor performance and coordination (Experiment III). Results from Experiments I & II have indicated that the CNS can integrate frequency and intensity of auditory information to enhance motor performance and coordination among fingers. Intra-auditory integration was found to be most effective when uncertainty in auditory feedback was moderate between two extreme levels of uncertainty (low and high uncertainty). Results from Experiment III indicate that practice leads to the enhancement of performance by reducing individual finger variability without changes in inter-finger coordination. Further, the enhancement of performance and coordination after practice was specific to the intra-auditory modality that was available during practice. This dissertation discusses the mechanisms responsible for the changes in motor performance and coordination with auditory feedback and directions for future research are suggested.
  • Thumbnail Image
    Item
    Analysis of Gamma-Band Auditory Responses in Schizophrenia
    (2015) Walsh, Benjamin Bryan; Simon, Jonathan Z; Electrical Engineering; Digital Repository at the University of Maryland; University of Maryland (College Park, Md.)
    Schizophrenia is a debilitating mental illness that affects 1% of the general population. One characteristic symptom is auditory hallucinations, which is experienced by almost all patients sometime in their lifetime. To investigate differences in auditory response in general, 50 schizophrenic patients and 50 age and sex-matched healthy controls were presented with auditory click trains at 40 Hz. Responses are recorded using electroencephalography (EEG). Magnitude and phase of responses at 40 Hz are computed using Gabor filters. Supporting previous literature, a significant difference in inter-trial phase coherence (ITC) and overall power is found between patients and controls, in particular near stimulus onset. Additionally, this study also investigated inter-subject phase coherence (ISC). This study finds that ISC is in fact higher for patients, in particular near stimulus onset. One possible explanation is that while healthy controls develop a preferred phase for perception, schizophrenic patients exhibit phase that is primarily stimulus-driven.
  • Thumbnail Image
    Item
    Memory-related cognitive modulation of human auditory cortex: Magnetoencephalography-based validation of a computational model
    (2008-04-09) Rong, Feng; Contreras-Vidal, José L; Neuroscience and Cognitive Science; Digital Repository at the University of Maryland; University of Maryland (College Park, Md.)
    It is well known that cognitive functions exert task-specific modulation of the response properties of human auditory cortex. However, the underlying neuronal mechanisms are not well understood yet. In this dissertation I present a novel approach for integrating 'bottom-up' (neural network modeling) and 'top-down' (experiment) methods to study the dynamics of cortical circuits correlated to shortterm memory (STM) processing that underlie the task-specific modulation of human auditory perception during performance of the delayed-match-to-sample (DMS) task. The experimental approach measures high-density magnetoencephalography (MEG) signals from human participants to investigate the modulation of human auditory evoked responses (AER) induced by the overt processing of auditory STM during task performance. To accomplish this goal, a new signal processing method based on independent component analysis (ICA) was developed for removing artifact contamination in the MEG recordings and investigating the functional neural circuits underlying the task-specific modulation of human AER. The computational approach uses a large-scale neural network model based on the electrophysiological knowledge of the involved brain regions to simulate system-level neural dynamics related to auditory object processing and performance of the corresponding tasks. Moreover, synthetic MEG and functional magnetic resonance imaging (fMRI) signals were simulated with forward models and compared to current and previous experimental findings. Consistently, both simulation and experimental results demonstrate a DMSspecific suppressive modulation of the AER and corresponding increased connectivity between the temporal auditory and frontal cognitive regions. Overall, the integrated approach illustrates how biologically-plausible neural network models of the brain can increase our understanding of brain mechanisms and their computations at multiple levels from sensory input to behavioral output with the intermediate steps defined.