Neural Networks as Databases - From Data to Model Compression

dc.contributor.advisorShrivastava, Abhinaven_US
dc.contributor.authorMaiya, Shishira Raghunathen_US
dc.contributor.departmentComputer Scienceen_US
dc.contributor.publisherDigital Repository at the University of Marylanden_US
dc.contributor.publisherUniversity of Maryland (College Park, Md.)en_US
dc.date.accessioned2025-09-15T05:37:56Z
dc.date.issued2025en_US
dc.description.abstractThe past decade has witnessed an exponential increase in data and the rise of deep learning systems to handle it. These systems primarily analyze, learn and interpret the data by excelling in exploiting patterns. This leads us to the question - can we use these patterns to efficiently store the data as well ? Essentially this would mean moving from the paradigm of data compression towards model compression. This thesis introduces and investigates a variety of methods in the realm of model compression, with the eventual goal of replacing data compression itself. To explore this, we start with the most information rich, yet sparse media form - videos. In the first part of this thesis, we introduce a framework for representing videos as continuous functions using Implicit Neural Representations (INRs), which aim to represent any given signal as a mapping between the spatial/temporal coordinate space to its values. We propose an auto-regressive framework that exploits the redundancies of a video for efficient compression and real time decoding - a first in this field. We then build upon this framework by introducing a shared video prior that captures common patterns across video frames, significantly improving the encoding times by 10-20$\times$, while improving the compression rates. Despite these value additions, INR's aren't really representations of underlying data unless the resulting model weights also encode some semantics. Based on this intuition, we introduce a hypernetwork-based INR system enables us to perform semantic tasks like video retrieval and understanding along with compression. Having established that the problem of data compression is actually a model compression problem, I will then present a case study on a widely used model compression method: pruning. We take a popular pruning method - the Lottery Ticket Hypothesis which offers extreme sparsity and study its effects on fundamental vision tasks like classification, detection and segmentation. Finally, we take a look at the proposed future directions in which we explore enhanced network and algorithmic designs with random networks for greater compression and meta-learning for achieving faster video encoding.en_US
dc.identifierhttps://doi.org/10.13016/yyrf-kusd
dc.identifier.urihttp://hdl.handle.net/1903/34656
dc.language.isoenen_US
dc.subject.pqcontrolledComputer scienceen_US
dc.subject.pqcontrolledArtificial intelligenceen_US
dc.subject.pquncontrolledComputer visionen_US
dc.subject.pquncontrolledDeep Learningen_US
dc.subject.pquncontrolledMachine Learningen_US
dc.titleNeural Networks as Databases - From Data to Model Compressionen_US
dc.typeDissertationen_US

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Maiya_umd_0117E_25489.pdf
Size:
69.87 MB
Format:
Adobe Portable Document Format