Foundations of Multiple-Time-Scale Stochastic Approximation for Fast and Resilient Distributed Optimization Algorithms
dc.contributor.author | Dutta, Amit | en |
dc.contributor.committeechair | Doan, Thinh Thanh | en |
dc.contributor.committeemember | Reed, Jeffrey H. | en |
dc.contributor.committeemember | Stilwell, Daniel J. | en |
dc.contributor.committeemember | L'Afflitto, Andrea | en |
dc.contributor.committeemember | Williams, Ryan K. | en |
dc.contributor.department | Electrical Engineering | en |
dc.date.accessioned | 2025-06-03T08:03:46Z | en |
dc.date.available | 2025-06-03T08:03:46Z | en |
dc.date.issued | 2025-06-02 | en |
dc.description.abstract | This dissertation establishes a theoretical framework for analyzing and designing fast and resilient distributed optimization algorithms for large-scale networks. The central focus is on understanding and leveraging two-time-scale dynamics, which may arise naturally from the underlying network structure or be introduced through algorithmic design. Motivated by challenges in large-scale decentralized networks—such as communication constraints, agent failures, and dynamic participation—this work develops and analyzes four key aspects. First, we study consensus-based gradient methods over networks with a clustered structure, characterized by dense intra-cluster and sparse inter-cluster communication. This setting naturally gives rise to textit{two-time-scale} dynamics. We develop a framework to analyze these dynamics and, using the singular perturbation method, derive new convergence rate results that reveal how local and global interactions influence the scalability and performance of distributed algorithms over clustered networks. Second, we analyze a two-time-scale distributed gradient descent algorithm for networks where agents communicate using quantized information. By averaging the quantized values at a faster time scale and performing gradient updates at a slower time scale, the algorithm effectively mitigates the impact of quantization errors. We introduce a novel analysis and step-size selection strategy that allows us to establish exact convergence to the optimal solution at the optimal rate. Third, we develop a resilient distributed algorithm for networks exposed to malicious agents. By introducing separate update speeds for the gradient estimation and gradient descent steps, the algorithm exhibits a two-time-scale behavior. We show that this design effectively mitigates the impact of both malicious agents and stochastic gradients, while still ensuring exact convergence at the optimal rate. Finally, we study open networks, where agents can dynamically join or leave. By leveraging redundancy in local objectives, we convert the time-varying problem into an effectively static one. This enables us to design both a local stochastic gradient descent algorithm and a gradient-balancing protocol for resource allocation, each of which achieves optimal and exact convergence. | en |
dc.description.abstractgeneral | This dissertation focuses on making distributed systems—such as networks of drones, sensors, or computers—work together efficiently and reliably, even under real-world challenges. In these systems, each unit (or agent) has access to only part of the information and must coordinate with others to solve a larger optimization problem. The central goal of this work is to develop optimization algorithms that are both fast and resilient. In practice, distributed systems face several difficulties: agents may fail or behave maliciously, communication may be limited or noisy, and the network may change dynamically as agents join or leave. To address these challenges, this dissertation presents four key contributions. First, it studies how agents can quickly reach consensus when the network has a natural cluster structure. These clusters lead to what are known as two-time-scale dynamics—faster communication within clusters and slower communication across them. By leveraging this structure, the distributed protocols are shown to have improved convergence speed and scalability. Second, it addresses communication constraints by demonstrating how agents can accurately solve optimization problems even when limited to exchanging quantized information. By leveraging a two-time-scale approach, the method effectively mitigates the impact of quantization and ensures optimal convergence. Third, it develops an algorithm that is robust to faulty or adversarial agents. By introducing separate update speeds for gradient estimation and local optimization steps—resulting in two-time-scale dynamics—the method effectively suppresses malicious behavior while ensuring optimal and exact convergence. Finally, it introduces distributed methods for open networks, where agents may join or leave at any time. Using the so-called redundancy property in the agents' objectives, the approach ensures that the optimization process remains stable and achieves exact convergence, regardless of network changes. | en |
dc.description.degree | Doctor of Philosophy | en |
dc.format.medium | ETD | en |
dc.identifier.other | vt_gsexam:43937 | en |
dc.identifier.uri | https://hdl.handle.net/10919/134989 | en |
dc.language.iso | en | en |
dc.publisher | Virginia Tech | en |
dc.rights | In Copyright | en |
dc.rights.uri | http://rightsstatements.org/vocab/InC/1.0/ | en |
dc.subject | Distributed optimization | en |
dc.subject | Cluster networks | en |
dc.subject | Federated optimization | en |
dc.subject | Byzantine fault-tolerance | en |
dc.subject | Quantized communication | en |
dc.subject | Two-time-scale methods | en |
dc.title | Foundations of Multiple-Time-Scale Stochastic Approximation for Fast and Resilient Distributed Optimization Algorithms | en |
dc.type | Dissertation | en |
thesis.degree.discipline | Electrical Engineering | en |
thesis.degree.grantor | Virginia Polytechnic Institute and State University | en |
thesis.degree.level | doctoral | en |
thesis.degree.name | Doctor of Philosophy | en |
Files
Original bundle
1 - 1 of 1