Improved 2D Camera-Based Multi-Object Tracking for Autonomous Vehicles
dc.contributor.author | Shinde, Omkar Mahesh | en |
dc.contributor.committeechair | Eskandarian, Azim | en |
dc.contributor.committeechair | Akbari Hamed, Kaveh | en |
dc.contributor.committeemember | Taheri, Saied | en |
dc.contributor.department | Mechanical Engineering | en |
dc.date.accessioned | 2025-03-07T09:01:07Z | en |
dc.date.available | 2025-03-07T09:01:07Z | en |
dc.date.issued | 2025-03-06 | en |
dc.description.abstract | Effective multi-object tracking is crucial for autonomous vehicles to navigate safely and efficiently in dynamic environments. To make autonomous vehicles more affordable one area to address is the computational limitations of the sensors, therefore, cameras are often the first choice sensor. Three challenges in implementation of multi-object tracking in autonomous vehicles are: 1) In these vehicles, sensors like cameras are not static, which can cause motion blur in the frames and make tracking inefficient. 2) Traditional methods for motion compensation, such as those used in Kalman Filter-based Multi-Object Tracking, require extensive parameter tuning to match features between consecutive frames accurately. 3) Simple intersection over union (IoU) metric is insufficient for reliable identification in such environments. This thesis proposes a novel methodology for 2D multi-object tracking in autonomous vehicles using a camera-based Tracking-by-Detection (TBD) approach, emphasizing four key innovations: (1) A real-time deblurring module to mitigate motion blur, ensuring clearer frames for accurate detection; (2) deep learning-based motion compensation module that adapts dynamically to varying motion patterns, enhancing robustness; (3) adaptive cost function for association, incorporating object appearance and temporal consistency to improve upon traditional IoU metrics; (4) The integration of the Unscented Kalman Filter to effectively address non-linearities in the tracking process, enhancing state estimation accuracy. To maintain a Simple Online and Realtime (SORT) framework, we enhance detection by fine-tuning YOLOv8 and YOLOv9 models using autonomous driving datasets like BDD100K and KITTI, which are specifically tailored for these scenarios. Additionally, we incorporate a non-linear approach using the UKF to better capture the influence of various tracking dynamics, further improving tracking performance. Our evaluations show that the proposed methodology significantly outperforms existing state-of-the-art methods while maintaining the same inference rate as the baseline SORT model. These advancements not only improve the accuracy and reliability of multi-object tracking but also reduce the computational burden associated with parameter tuning and motion compensation. Consequently, this work presents a robust and efficient tracking solution for autonomous vehicles, making it viable for real-world deployment under both computational and cost constraints. | en |
dc.description.abstractgeneral | Tracking multiple objects is really important for self-driving cars to move safely in busy places. Cameras are often the best choice because they are cheaper and easier to use, but using cameras comes with three main challenges: (1) When cars move, cameras can make blurry images, which makes it harder to see and track things; (2) Traditional tracking methods, like Kalman Filters, need a lot of adjustments to work well; (3) Simple methods, like checking if objects overlap (called Intersection over Union), are not always good enough in crowded, complicated places. This thesis presents a new way to track lots of things using cameras, with four big improvements: (1) A real-time deblurring system that fixes blurry pictures so the camera can see things more clearly; (2) A smart system that uses deep learning to follow movement better; (3) A better way to match objects by using not just their positions but also how they look and move over time, which is better than old IoU methods; (4) A special tool called the Unscented Kalman Filter that helps track objects more accurately when their movements aren't simple or straight. To keep everything simple, fast, and real-time, we use object detectors to help find objects, and we train them with special self-driving datasets like BDD100K and KITTI. These datasets are great for showing the kinds of situations self-driving cars deal with. The Unscented Kalman Filter helps us track objects with more complicated movements, making everything more accurate. Our study show that this new way works much better than older methods, without making the system slower. These improvements make tracking more reliable and cut down on the time needed for tuning and adjusting. Overall, this work provides a strong and simple solution for tracking things in self-driving cars, even if the computer isn't super powerful or the budget is small. | en |
dc.description.degree | Master of Science | en |
dc.format.medium | ETD | en |
dc.identifier.other | vt_gsexam:42578 | en |
dc.identifier.uri | https://hdl.handle.net/10919/124823 | en |
dc.language.iso | en | en |
dc.publisher | Virginia Tech | en |
dc.rights | In Copyright | en |
dc.rights.uri | http://rightsstatements.org/vocab/InC/1.0/ | en |
dc.subject | Multi-Object Tracking | en |
dc.subject | 2D Perception | en |
dc.subject | Autonomous Driving | en |
dc.title | Improved 2D Camera-Based Multi-Object Tracking for Autonomous Vehicles | en |
dc.type | Thesis | en |
thesis.degree.discipline | Mechanical Engineering | en |
thesis.degree.grantor | Virginia Polytechnic Institute and State University | en |
thesis.degree.level | masters | en |
thesis.degree.name | Master of Science | en |