Efficient Optimization Algorithms for Nonconvex Machine Learning Problems
dc.contributor.advisor | Huang, Heng HH | en_US |
dc.contributor.author | Xian, Wenhan | en_US |
dc.contributor.department | Computer Science | en_US |
dc.contributor.publisher | Digital Repository at the University of Maryland | en_US |
dc.contributor.publisher | University of Maryland (College Park, Md.) | en_US |
dc.date.accessioned | 2024-09-23T06:14:21Z | |
dc.date.available | 2024-09-23T06:14:21Z | |
dc.date.issued | 2024 | en_US |
dc.description.abstract | In recent years, the success of the AI revolution has led to the training of larger neural networks on vast amounts of data to achieve superior performance. These powerful machine learning models have enabled the creation of remarkable AI products. Optimization, as the core of machine learning, becomes especially crucial because most machine learning problems can ultimately be formulated as optimization problems, which require minimizing a loss function with respect to model parameters based on training samples. To enhance the efficiency of optimization algorithms, distributed learning has emerged as a popular solution for addressing large-scale machine learning tasks. In distributed learning, multiple worker nodes collaborate to train a global model. However, a key challenge in distributed learning is the communication cost. This thesis introduces a novel adaptive gradient algorithm with gradient sparsification to address this issue. Another significant challenge in distributed learning is the communication overhead on the central parameter server. To mitigate this bottleneck, decentralized distributed (serverless) learning has been proposed, where each worker node only needs to communicate with its neighbors. This thesis investigates core nonconvex optimization problems in decentralized settings, including constrained optimization, minimax optimization, and second-order optimality. Efficient optimization algorithms are proposed to solve these problems. Additionally, the convergence analysis of minimax optimization under the generalized smooth condition is explored. A generalized algorithm is proposed, which can be applied to a broader range of applications. | en_US |
dc.identifier | https://doi.org/10.13016/r6ko-uppv | |
dc.identifier.uri | http://hdl.handle.net/1903/33413 | |
dc.language.iso | en | en_US |
dc.subject.pqcontrolled | Computer science | en_US |
dc.subject.pquncontrolled | Decentralized Learning | en_US |
dc.subject.pquncontrolled | Distributed Learning | en_US |
dc.subject.pquncontrolled | Machine Learning | en_US |
dc.subject.pquncontrolled | Optimization | en_US |
dc.title | Efficient Optimization Algorithms for Nonconvex Machine Learning Problems | en_US |
dc.type | Dissertation | en_US |
Files
Original bundle
1 - 1 of 1