Neuromorphic model for sound source segregation

dc.contributor.advisorShamma, Shihaben_US
dc.contributor.authorKrishnan, Lakshmien_US
dc.contributor.departmentElectrical Engineeringen_US
dc.contributor.publisherDigital Repository at the University of Marylanden_US
dc.contributor.publisherUniversity of Maryland (College Park, Md.)en_US
dc.date.accessioned2016-06-22T05:36:40Z
dc.date.available2016-06-22T05:36:40Z
dc.date.issued2015en_US
dc.description.abstractWhile humans can easily segregate and track a speaker's voice in a loud noisy environment, most modern speech recognition systems still perform poorly in loud background noise. The computational principles behind auditory source segregation in humans is not yet fully understood. In this dissertation, we develop a computational model for source segregation inspired by auditory processing in the brain. To support the key principles behind the computational model, we conduct a series of electro-encephalography experiments using both simple tone-based stimuli and more natural speech stimulus. Most source segregation algorithms utilize some form of prior information about the target speaker or use more than one simultaneous recording of the noisy speech mixtures. Other methods develop models on the noise characteristics. Source segregation of simultaneous speech mixtures with a single microphone recording and no knowledge of the target speaker is still a challenge. Using the principle of temporal coherence, we develop a novel computational model that exploits the difference in the temporal evolution of features that belong to different sources to perform unsupervised monaural source segregation. While using no prior information about the target speaker, this method can gracefully incorporate knowledge about the target speaker to further enhance the segregation.Through a series of EEG experiments we collect neurological evidence to support the principle behind the model. Aside from its unusual structure and computational innovations, the proposed model provides testable hypotheses of the physiological mechanisms of the remarkable perceptual ability of humans to segregate acoustic sources, and of its psychophysical manifestations in navigating complex sensory environments. Results from EEG experiments provide further insights into the assumptions behind the model and provide motivation for future single unit studies that can provide more direct evidence for the principle of temporal coherence.en_US
dc.identifierhttps://doi.org/10.13016/M2NB68
dc.identifier.urihttp://hdl.handle.net/1903/18155
dc.language.isoenen_US
dc.subject.pqcontrolledElectrical engineeringen_US
dc.subject.pqcontrolledNeurosciencesen_US
dc.subject.pquncontrolledauditory EEGen_US
dc.subject.pquncontrolledauditory scene analysisen_US
dc.subject.pquncontrolledcocktail party problemen_US
dc.subject.pquncontrolledsound source segregationen_US
dc.subject.pquncontrolledtemporal coherenceen_US
dc.titleNeuromorphic model for sound source segregationen_US
dc.typeDissertationen_US

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Krishnan_umd_0117E_16881.pdf
Size:
6.8 MB
Format:
Adobe Portable Document Format