Deep Reinforcement Learning for Multirotor Flight Control: A Comparative Study of Sim-to-Real Training and Real-World Performance

dc.contributor.authorThomas, Patrick Jamesen
dc.contributor.committeechairSchroeder, Kevin Kenten
dc.contributor.committeechairBlack, Jonathan T.en
dc.contributor.committeememberWoolsey, Craig A.en
dc.contributor.committeememberL'Afflitto, Andreaen
dc.contributor.departmentAerospace and Ocean Engineeringen
dc.date.accessioned2025-12-19T09:00:55Zen
dc.date.available2025-12-19T09:00:55Zen
dc.date.issued2025-12-18en
dc.description.abstractThis dissertation investigates Deep Reinforcement Learning (DRL) for low-level flight control of multirotor unmanned aerial vehicles (UAVs), focusing on factors that most influence sim-to-real policy transfer. Using the Proximal Policy Optimization (PPO) algorithm, extensive ablation studies evaluated the effects of domain randomization, reward weighting, observation representation, and neural network architecture. Policies were trained in a MuJoCo-based simulator and deployed via TensorFlow Lite Micro inference on PX4 flight controllers. Domain randomization of actuator and mass properties yielded the best balance between positional accuracy and attitude stability, achieving reductions of 40–50 % in position error and roll–pitch oscillation. Policies required only the current vehicle state and previous commanded action, maintaining less than 0.1 m steady-state error and less than 5 % overshoot. Among activation functions tested, ReLU outperformed tanh and ELU, lowering steady-state error by up to 36 % and inference time by 26 %. Post-training quantization further reduced inference latency by ≈40 % with negligible performance loss. Incorporating trajectory tracking during training decreased tracking error by ≈75 % and eliminated temporal lag. The optimal training configuration generalized effectively to multiple vehicle morphologies, including quadcopter, hexacopter, and coaxial octocopter platforms. In addition, the process was successfully extended to an omnidirectional multi-rotor vehicle (OMV). For the OMV, a learned Multi-Layer Perceptron (MLP) controller outperformed both adaptive and PID-based baselines when commanded to track a complex reference attitude, demonstrating stable six-degree-of-freedom trajectory tracking in experimental flights. Collectively, these results provide new insight into parameter sensitivities within the DRL training pipeline and establish a reproducible methodology for sim-to-real policy transfer in aerial robotics.en
dc.description.abstractgeneralUnmanned aerial vehicles (UAVs)—more commonly known as drones—have become indispensable tools across industries ranging from filmmaking and infrastructure inspection to logistics and defense. However, their ability to operate safely and reliably in the real world depends on how well their onboard flight controllers can adapt to uncertain and changing conditions. Traditional control systems struggle when real-world conditions differ from their mathematical models. This dissertation explores how deep reinforcement learning (DRL)—a branch of machine learning where agents learn through trial and error—can be used to train flight controllers that are more robust and adaptable. Using advanced simulation techniques, this research identifies which aspects of the training process—such as environmental variability, reward functions, and neural network design—most affect real-world performance when controllers are transferred from simulation to physical flight. The work also extends these techniques to an omnidirectional multirotor vehicle, capable of moving and rotating freely in all directions. The findings offer a framework for developing more reliable and generalizable learning-based flight controllers, contributing to the broader goal of making autonomous aerial systems more capable in complex, uncertain environments.en
dc.description.degreeDoctor of Philosophyen
dc.format.mediumETDen
dc.identifier.othervt_gsexam:45302en
dc.identifier.urihttps://hdl.handle.net/10919/140039en
dc.language.isoenen
dc.publisherVirginia Techen
dc.rightsCreative Commons Attribution 4.0 Internationalen
dc.rights.urihttp://creativecommons.org/licenses/by/4.0/en
dc.subjectUAVen
dc.subjectDeep Reinforcement Learningen
dc.subjectControlsen
dc.titleDeep Reinforcement Learning for Multirotor Flight Control: A Comparative Study of Sim-to-Real Training and Real-World Performanceen
dc.typeDissertationen
thesis.degree.disciplineAerospace Engineeringen
thesis.degree.grantorVirginia Polytechnic Institute and State Universityen
thesis.degree.leveldoctoralen
thesis.degree.nameDoctor of Philosophyen

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Thomas_PJ_D_2025.pdf
Size:
27.07 MB
Format:
Adobe Portable Document Format