Efficient Reinforcement Learning for Control

dc.contributor.authorBaddam, Vasanth Reddyen
dc.contributor.committeechairEldardiry, Hoda Mohameden
dc.contributor.committeechairBoker, Almuatazbellah M.en
dc.contributor.committeememberGumussoy, Suaten
dc.contributor.committeememberCho, Jin-Heeen
dc.contributor.committeememberWatson, Layne T.en
dc.contributor.departmentComputer Science and#38; Applicationsen
dc.date.accessioned2025-07-02T08:00:34Zen
dc.date.available2025-07-02T08:00:34Zen
dc.date.issued2025-07-01en
dc.description.abstractThe landscape of control systems has evolved rapidly with the emergence of Reinforcement Learning (RL), offering promising solutions to a wide range of dynamic decision-making problems. However, the application of RL to real-world control systems is often hindered by computational inefficiencies, scalability issues, and a lack of structure in learning mech- anisms. This thesis explores a central question: How can we design reinforcement learning algorithms that are not only effective but also computationally effi- cient and scalable for control systems of increasing complexity? To address this, we present a progression of approaches—starting with time-scale decomposition in small- scale systems and moving towards structured and adaptive learning strategies for large-scale, multi-agent control problems. Each chapter builds upon the previous one by introducing new methods tailored to the complexity and scale of the environment, culminating in a unified framework for efficient RL-driven controlen
dc.description.abstractgeneralThe landscape of control systems has evolved rapidly with the emergence of Reinforcement Learning (RL), offering promising solutions to a wide range of dynamic decision-making problems. However, the application of RL to real-world control systems is often hindered by computational inefficiencies, scalability issues, and a lack of structure in learning mech- anisms. This thesis explores a central question: How can we design reinforcement learning algorithms that are not only effective but also computationally effi- cient and scalable for control systems of increasing complexity? To address this, we present a progression of approaches—starting with time-scale decomposition in small- scale systems and moving towards structured and adaptive learning strategies for large-scale, multi-agent control problems. Each chapter builds upon the previous one by introducing new methods tailored to the complexity and scale of the environment, culminating in a unified framework for efficient RL-driven controlen
dc.description.degreeDoctor of Philosophyen
dc.format.mediumETDen
dc.identifier.othervt_gsexam:43548en
dc.identifier.urihttps://hdl.handle.net/10919/135747en
dc.language.isoenen
dc.publisherVirginia Techen
dc.rightsCreative Commons Attribution-NonCommercial 4.0 Internationalen
dc.rights.urihttp://creativecommons.org/licenses/by-nc/4.0/en
dc.subjectOptimal Control; Reinforcement Learningen
dc.titleEfficient Reinforcement Learning for Controlen
dc.typeDissertationen
thesis.degree.disciplineComputer Science & Applicationsen
thesis.degree.grantorVirginia Polytechnic Institute and State Universityen
thesis.degree.leveldoctoralen
thesis.degree.nameDoctor of Philosophyen

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Baddam_V_D_2025.pdf
Size:
7.33 MB
Format:
Adobe Portable Document Format