Browsing by Author "Yuan, Jiehao"
Now showing 1 - 2 of 2
Results Per Page
Sort Options
- Head-up Displays Improve Drivers' Performance and Subjective Perceptions with the In-Vehicle Gesture Interaction SystemCao, Yusheng; Li, Lingyu; Yuan, Jiehao; Jeon, Myounghoon (Taylor & Francis, 2024-01-01)In-vehicle infotainment systems can cause various distractions, increasing the risk of car accidents. To address this problem, mid-air gesture systems have been introduced. This study investigated the potential of a novel interface that integrates a Head-Up Display (HUD) with auditory displays (spearcons: compressed speech) in a gesture-based menu navigation system to minimize visual distraction and improve driving and secondary task performance. The experiment involved 24 participants who navigated through 12 menu items using mid-air gestures while driving on a simulated road under four conditions: HUD (with, without spearcons) and Head-Down Display (HDD) (with, without spearcons). Results showed that the HUD condition significantly outperformed the HDD condition in participants’ level 1 situation awareness, perceived workload, menu navigation performance, and system usability. However, there were trade-offs on visual fixation duration on the menu, and lane deviation. These findings will guide future research in developing safer and more effective HUD-supported in-vehicle gesture interaction systems.
- Increasing Driving Safety and In-Vehicle Gesture-Based Menu Navigation Accuracy with a Heads-up DisplayCao, Yusheng; Li, Lingyu; Yuan, Jiehao; Jeon, Myounghoon (ACM, 2022)More and more novel functions are being integrated into the vehicle infotainment system to allow individuals to perform secondary tasks with high accuracy and low accident risks. Mid-air gesture interactions are one of them. The current paper will present novel designs to solve a specific issue with this method of interaction: visual distraction within the car. In this study, a Heads-up display (HUD) will be integrated with a gesture-based menu navigation system to allow drivers to see menu selections without looking away from the road. An experiment will be conducted to investigate the potential of this system in improving drivers' overall safety and gesture interaction accuracy. The experiment will recruit 24 participants to test the system. Participants will provide subjective feedback about the directions for conducting future research and improving the overall experience, as well as objective performance data.