Cooperative Payload Transportation by UAVs: A Model-Based Deep Reinforcement Learning (MBDRL) Application
dc.contributor.author | Khursheed, Shahwar Atiq | en |
dc.contributor.committeechair | Williams, Ryan K. | en |
dc.contributor.committeemember | Boker, Almuatazbellah M. | en |
dc.contributor.committeemember | Doan, Thinh Thanh | en |
dc.contributor.department | Electrical Engineering | en |
dc.date.accessioned | 2024-08-21T08:00:26Z | en |
dc.date.available | 2024-08-21T08:00:26Z | en |
dc.date.issued | 2024-08-20 | en |
dc.description.abstract | We propose a Model-Based Deep Reinforcement Learning (MBDRL) framework for collaborative paylaod transportation using Unmanned Aerial Vehicles (UAVs) in Search and Rescue (SAR) missions, enabling heavier payload conveyance while maintaining vehicle agility. Our approach extends the single-drone application to a novel multi-drone one, using the Probabilistic Ensembles with Trajectory Sampling (PETS) algorithm to model the unknown stochastic system dynamics and uncertainty. We use the Multi-Agent Reinforcement Learning (MARL) framework via a centralized controller in a leader-follower configuration. The agents utilize the approximated transition function in a Model Predictive Controller (MPC) configured to maximize the reward function for waypoint navigation, while a position-based formation controller ensures stable flights of these physically linked UAVs. We also developed an Unreal Engine (UE) simulation connected to an offboard planner and controller via a Robot Operating System (ROS) framework that is transferable to real robots. This work achieves stable waypoint navigation in a stochastic environment with a sample efficiency following that seen in single UAV work. This work has been funded by the National Science Foundation (NSF) under Award No. 2046770. | en |
dc.description.abstractgeneral | We apply the Model-Based Deep Reinforcement Learning (MBDRL) framework to the novel application of a UAV team transporting a suspended payload during Search and Rescue missions. Collaborating UAVs can transport heavier payloads while staying agile, reducing the need for human involvement. We use the Probabilistic Ensemble with Trajectory Sampling (PETS) algorithm to model uncertainties and build on the previously used single UAVpayload system. By utilizing the Multi-Agent Reinforcement Learning (MARL) framework via a centralized controller, our UAVs learn to transport the payload to a desired position while maintaining stable flight through effective cooperation. We also develop a simulation in Unreal Engine (UE) connected to a controller using a Robot Operating System (ROS) architecture, which can be transferred to real robots. Our method achieves stable navigation in unpredictable environments while maintaining the sample efficiency observed in single UAV scenarios. | en |
dc.description.degree | Master of Science | en |
dc.format.medium | ETD | en |
dc.identifier.other | vt_gsexam:41310 | en |
dc.identifier.uri | https://hdl.handle.net/10919/120971 | en |
dc.language.iso | en | en |
dc.publisher | Virginia Tech | en |
dc.rights | Creative Commons Attribution 4.0 International | en |
dc.rights.uri | http://creativecommons.org/licenses/by/4.0/ | en |
dc.subject | Unmanned aerial vehicles | en |
dc.subject | Unreal Engine | en |
dc.subject | Model-based deep reinforcement learning | en |
dc.subject | Cooperative multi-agent systems | en |
dc.subject | Motion planning | en |
dc.subject | Payload transportation | en |
dc.title | Cooperative Payload Transportation by UAVs: A Model-Based Deep Reinforcement Learning (MBDRL) Application | en |
dc.type | Thesis | en |
thesis.degree.discipline | Electrical Engineering | en |
thesis.degree.grantor | Virginia Polytechnic Institute and State University | en |
thesis.degree.level | masters | en |
thesis.degree.name | Master of Science | en |
Files
Original bundle
1 - 1 of 1