Optimal Point-Spread-Function Engineering with Dynamic Optics and Event Cameras

dc.contributor.advisorMetzler, Christopher Aen_US
dc.contributor.authorShah, Sachinen_US
dc.contributor.departmentComputer Scienceen_US
dc.contributor.publisherDigital Repository at the University of Marylanden_US
dc.contributor.publisherUniversity of Maryland (College Park, Md.)en_US
dc.date.accessioned2024-07-02T05:43:24Z
dc.date.available2024-07-02T05:43:24Z
dc.date.issued2024en_US
dc.description.abstractComputational imaging systems co-design optics and algorithms to observe phenomena beyond the reach of traditional cameras. Point-spread-function (PSF) engineering is a powerful technique wherein a custom phase mask is integrated into an optical system to encode additional information into captured images. Used in combination with deep learning, such systems now offer state-of-the-art performance at three-dimensional molecule localization, extended depth-of-field imaging, lensless imaging, and other tasks. Recent hardware breakthroughs are unlocking unprecedented ultrafast capabilities such as micro-electromechanical system based spatial light modulators will allow us to module light at kilohertz rates and neuromorphic event cameras will enable kilohertz lower-power and high-dynamic-range capture. Unfortunately, existing theories and algorithms are unable to fully harness these new capabilities. This work answers a natural question: Can one encode additional information and achieve superior performance by leveraging the ultrafast capabilities of spatial light modulators and event cameras. We first prove that the set of PSFs described by static phase masks is non-convex and that, as a result, time-averaged PSFs generated by dynamic phase masks displayed on a spatial light modulator are fundamentally more expressive. We then derive the theoretical limits on three-dimensional tracking with PSF-engineered event cameras. Using these bounds, we design new optimal phase masks and binary amplitude masks. We demonstrate the efficacy of our designs through extensive simulations and validate our method with a simple lab prototype.en_US
dc.identifierhttps://doi.org/10.13016/fo1l-j1dl
dc.identifier.urihttp://hdl.handle.net/1903/33051
dc.language.isoenen_US
dc.subject.pqcontrolledComputer scienceen_US
dc.subject.pqcontrolledOpticsen_US
dc.subject.pqcontrolledArtificial intelligenceen_US
dc.subject.pquncontrolledComputational Imagingen_US
dc.subject.pquncontrolledEvent Camerasen_US
dc.subject.pquncontrolledMachine Learningen_US
dc.titleOptimal Point-Spread-Function Engineering with Dynamic Optics and Event Camerasen_US
dc.typeThesisen_US

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Shah_umd_0117N_24332.pdf
Size:
10.07 MB
Format:
Adobe Portable Document Format