Convergence Rates of Gradient Descent-ascent Dynamics under Computation Constraints in Solving Min-max Optimization
dc.contributor.author | Do, Duy Anh | en |
dc.contributor.committeechair | Doan, Thinh Thanh | en |
dc.contributor.committeemember | L'Afflitto, Andrea | en |
dc.contributor.committeemember | Boker, Almuatazbellah M. | en |
dc.contributor.department | Electrical Engineering | en |
dc.date.accessioned | 2025-04-11T08:00:17Z | en |
dc.date.available | 2025-04-11T08:00:17Z | en |
dc.date.issued | 2025-04-10 | en |
dc.description.abstract | This thesis is dedicated to providing a new analysis of the convergence rates of the Gradient Descent-ascent (GDA) method for solving the Min-max optimization (MMO) problems under computation constraints. In particular, we focus our study on two main classes of MMO: a continuous-time variant of the centralized Min-max problem where the GDA update only has access to the gradients of the objective function after some delay, as well as the Federated Min-max Learning (FML) problem, a special setting within the family of decentralized MMO under quantization constraints. Also known as the saddle point problem due to its objective being to find a saddle point of a function, the MMO problem, as well as its decentralized counterpart, has received wide attention due to its significant impact in different fields such as stochastic control and training generative adversarial networks. The GDA method is one of the most celebrated algorithms to find the saddle point of such functions, due to its computational efficiency and ease of implementation. Therefore, it is important that we delve into the analysis of the GDA approach in the MMO problem, not only in the traditional settings where information can be readily and perfectly available to the computation units but also under more practical scenarios. In this thesis, we focus our study on two special types of constraints. Firstly, understanding that calculating gradients of a function instantly is practically impossible, we tackle the continuous-time variant of centralized GDA under delayed gradients. We utilize the singular perturbation approach to obtain convergence rates in two non-convex settings of the objective function, namely, the two-sided and one-sided Polyak - Lojasiewicz (PL) conditions, by designing a coupling Lyapunov function to capture the interaction between the gradient descent and ascent dynamics subject to asynchronous gradients. Secondly, we study an important framework within decentralized MMO named Federated Min-max Learning, in which limited communication bandwidth requires information exchanged to be quantized. We apply a similar two-time-scale GDA technique to obtain convergence rates in three different settings, namely, the strongly-convex-strongly-concave case, and when it is subject to the two-sided and one-sided PL conditions. Finally, we provide numerical simulations to demonstrate the efficiency of our theoretical results. | en |
dc.description.abstractgeneral | The Min-max optimization (MMO) problem, with the goal being to find the best of the worst point in a 2-variables function is an important problem because of its significance in different areas such as automated control and machine learning. A common method to solve MMO is called Gradient Descent-Ascent (GDA), which involves iteratively updating the value of the variables along the direction of the textit{gradients}, a rate of changes of the function. However, in practical scenarios, solving these problems is often complicated by constraints, such as delays in processing information or limitations in communication. This thesis focuses on understanding how these constraints affect the performance of the GDA method. In particular, this thesis provides a new analysis of how fast the GDA method finds the solution in two different settings. The first setting considers a scenario where we only have access to the gradients after some delay, which is more practical given that calculating the gradients of a function instantly is not feasible. The second setting focuses on decentralized environments, where multiple devices or agents work together to solve the problem but must communicate through limited bandwidth, requiring the information exchanged to be approximated. | en |
dc.description.degree | Master of Science | en |
dc.format.medium | ETD | en |
dc.identifier.other | vt_gsexam:42607 | en |
dc.identifier.uri | https://hdl.handle.net/10919/125164 | en |
dc.language.iso | en | en |
dc.publisher | Virginia Tech | en |
dc.rights | In Copyright | en |
dc.rights.uri | http://rightsstatements.org/vocab/InC/1.0/ | en |
dc.subject | Min-max Optimization | en |
dc.subject | Federated Min-max Learning | en |
dc.subject | Two-time-scale Gradient Descent-ascent | en |
dc.subject | Delays | en |
dc.subject | Random Quantization | en |
dc.subject | Singular Pertubation Approach | en |
dc.title | Convergence Rates of Gradient Descent-ascent Dynamics under Computation Constraints in Solving Min-max Optimization | en |
dc.type | Thesis | en |
thesis.degree.discipline | Electrical Engineering | en |
thesis.degree.grantor | Virginia Polytechnic Institute and State University | en |
thesis.degree.level | masters | en |
thesis.degree.name | Master of Science | en |
Files
Original bundle
1 - 1 of 1