UNDERSTANDING FAILURE MODES OF DEEP LEARNING MODELS

dc.contributor.advisorGoldstein, Tomen_US
dc.contributor.authorSomepalli, Gowthamien_US
dc.contributor.departmentComputer Scienceen_US
dc.contributor.publisherDigital Repository at the University of Marylanden_US
dc.contributor.publisherUniversity of Maryland (College Park, Md.)en_US
dc.date.accessioned2025-08-08T11:33:01Z
dc.date.issued2024en_US
dc.description.abstractIn the past few years, deep learning models have advanced significantly, achieving remarkable results in challenging problems across multiple domains such as vision and language. Despite their proficiency, these models exhibit numerous failure modes and intriguing behaviors that cannot be mitigated solely through scaling. A comprehensive understanding of these failure modes is imperative for their safe application and utilization. The first part of the thesis focuses on memorization in neural networks as a failure mode. We begin by asking, ``Do diffusion models create unique works of art, or are they replicating content directly from their training sets?" We discuss our proposed framework for comparing generated images with training images and detecting content replication. While it is widely believed that duplicated images in the training set are responsible for content replication at inference time, we demonstrate that the model's text conditioning plays a similarly important role. Motivated by our findings, we propose several techniques to reduce data replication at both training and inference times. We then explore the question, ``Do diffusion models memorize style?" First, we train a feature extractor to capture the style elements in a given image. Then, we use this feature extractor to study how effectively recent generative models can replicate style in their generations. The second part of the thesis focuses on the failure mode -- reproducibility in neural networks. We present visualization techniques for neural network decision boundaries to study reproducibility across different classification architectures and training methodologies. We show that architectural changes significantly alter decision boundaries, whereas similar architectures tend to yield similar decision regions, especially in wider models. Additionally, we also investigate the double descent phenomenon using these visualizations and propose a few hypotheses on how decision regions change with model scaling.en_US
dc.identifierhttps://doi.org/10.13016/gaom-mbci
dc.identifier.urihttp://hdl.handle.net/1903/34029
dc.language.isoenen_US
dc.subject.pqcontrolledComputer scienceen_US
dc.subject.pquncontrolledComputer Visionen_US
dc.subject.pquncontrolledDiffusion Modelsen_US
dc.subject.pquncontrolledGenerative Modelsen_US
dc.subject.pquncontrolledMachine Learningen_US
dc.titleUNDERSTANDING FAILURE MODES OF DEEP LEARNING MODELSen_US
dc.typeDissertationen_US

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Somepalli_umd_0117E_24859.pdf
Size:
39.51 MB
Format:
Adobe Portable Document Format