UMD Theses and Dissertations
Permanent URI for this collectionhttp://hdl.handle.net/1903/3
New submissions to the thesis/dissertation collections are added automatically as they are received from the Graduate School. Currently, the Graduate School deposits all theses and dissertations from a given semester after the official graduation date. This means that there may be up to a 4 month delay in the appearance of a given thesis/dissertation in DRUM.
More information is available at Theses and Dissertations at University of Maryland Libraries.
Browse
3 results
Search Results
Item Analyzing Inverse Design Problems from a Topological Perspective(2024) Chen, Qiuyi; Fuge, Mark; Mechanical Engineering; Digital Repository at the University of Maryland; University of Maryland (College Park, Md.)Inverse design (ID) problems are inverse problems that aim to rapidly retrieve the subset of valid designs having the desired performances and properties under the given conditions. In practice, this can be solved by training generative models to approximate and sample the posterior distributions of designs. However, little has been done to understand their mechanisms and limitations from a theoretical perspective. This dissertation leverages theoretical tools from general and differential topology to answer these three questions of inverse design: what does a set of valid designs look like? How helpful are the data-driven generative models for retrieving the desired designs from this set? What topological properties affect the subset of desired designs? The dissertation proceeds by dismantling inverse (design) problems into two major subjects: that is, the representing and probing of a given set of valid designs (or data), and the retrieval of the desired designs (or data) from this given set. It draws inspiration from topology and geometry to investigate them and makes the main contributions below: 1. Chapter 3 details a novel representation learning method called Least Volume, which has properties similar to nonlinear PCA for representing datasets. It can minimize the representation's dimension automatically and, as shown in Chapter 4, conducts contrastive learning when applied to labeled datasets. 2. Two conditional generative models are developed to generate performant 2-D airfoils and 3-D heat sinks in Chapter 5 and 6 respectively. They can produce realistic designs to warm-start further optimization, with the relevant chapters detailing their acceleration effects. 3. Lastly, Chapter 7 describes how to use Least volume to solve high-dimensional inverse problems efficiently. Specifically, using examples from physic system identification, the chapter uncovers the correlation between the inverse problem's uncertainty and its intrinsic dimensions.Item Denoising the Design Space: Diffusion Models for Accelerated Airfoil Shape Optimization(2024) Diniz, Cashen; Fuge, Mark D; Mechanical Engineering; Digital Repository at the University of Maryland; University of Maryland (College Park, Md.)Generative models offer the possibility to accelerate and potentially substitute parts of the often expensive traditional design optimization process. We present Aero-DDM, a novel application of a latent denoising diffusion model (DDM) capable of generating airfoil geometries conditioned on flow parameters and an area constraint. Additionally, we create a novel, diverse dataset of optimized airfoil designs that better reflects a realistic design space than has been done in previous work. Aero-DDM is applied to this dataset, and key metrics are assessed both statistically and with an open-source computational fluid dynamics (CFD) solver to determine the performance of the generated designs. We compare our approach to an optimal transport GAN, and demonstrate that our model can generate designs with superior performance statistically, in aerodynamic benchmarks, and in warm-start scenarios. We also extend our diffusion model approach, and demonstrate that the number of steps required for inference can be reduced by as much as ~86%, compared to an optimized version of the baseline inference process, without meaningful degradation in design quality, simply by using the initial design to start the denoising process.Item Leveraging Deep Generative Models for Estimation and Recognition(2023) PNVR, Koutilya; Jacobs, David W.; Electrical Engineering; Digital Repository at the University of Maryland; University of Maryland (College Park, Md.)Generative models are a class of statistical models that estimate the joint probability distribution on a given observed variable and a target variable. In computer vision, generative models are typically used to model the joint probability distribution of a set of real image samples assumed to be on a complex high-dimensional image manifold. The recently proposed deep generative architectures such as Variational Autoencoders (VAEs), Generative Adversarial Networks (GANs), and diffusion models (DMs) were shown to generate photo-realistic images of human faces and other objects. These generative models also became popular for other generative tasks such as image editing, text-to-image, etc. As appealing as the perceptual quality of the generated images has become, the use of generative models for discriminative tasks such as visual recognition or geometry estimation has not been well studied. Moreover, with different kinds of powerful generative models getting popular lately, it's important to study their significance in other areas of computer vision. In this dissertation, we demonstrate the advantages of using generative models for applications that go beyond just photo-realistic image generation: Unsupervised Domain Adaptation (UDA) between synthetic and real datasets for geometry estimation; Text-based image segmentation for recognition. In the first half of the dissertation, we propose a novel generative-based UDA method for combining synthetic and real images when training networks to determine geometric information from a single image. Specifically, we use a GAN model to map both synthetic and real domains into a shared image space by translating just the domain-specific task-related information from respective domains. This is connected to a primary network for end-to-end training. Ideally, this results in images from two domains that present shared information to the primary network. Compared to previous approaches, we demonstrate an improved domain gap reduction and much better generalization between synthetic and real data for geometry estimation tasks such as monocular depth estimation and face normal estimation. In the second half of the dissertation, we showcase the power of a recent class of generative models for improving an important recognition task: text-based image segmentation. Specifically, large-scale pre-training tasks like image classification, captioning, or self-supervised techniques do not incentivize learning the semantic boundaries of objects. However, recent generative foundation models built using text-based latent diffusion techniques may learn semantic boundaries. This is because they must synthesize intricate details about all objects in an image based on a text description. Therefore, we present a technique for segmenting real and AI-generated images using latent diffusion models (LDMs) trained on internet-scale datasets. First, we show that the latent space of LDMs (z-space) is a better input representation compared to other feature representations like RGB images or CLIP encodings for text-based image segmentation. By training the segmentation models on the latent z-space, which creates a compressed representation across several domains like different forms of art, cartoons, illustrations, and photographs, we are also able to bridge the domain gap between real and AI-generated images. We show that the internal features of LDMs contain rich semantic information and present a technique in the form of LD-ZNet to further boost the performance of text-based segmentation. Overall, we show up to 6% improvement over standard baselines for text-to-image segmentation on natural images. For AI-generated imagery, we show close to 20% improvement compared to state-of-the-art techniques.