4.6 Article

A Driver Gaze Estimation Method Based on Deep Learning

期刊

SENSORS
卷 22, 期 10, 页码 -

出版社

MDPI
DOI: 10.3390/s22103959

关键词

advanced driver-assistance technique; You Only Look Once (YOLO); Inception-v3; CNN; InceptionResNet-v2

资金

  1. National Key Research and Development Program [2018YFB1600202, 2021YFB1600205]
  2. National Natural Science Foundation of China [52178407]

向作者/读者索取更多资源

This research proposes a method to build an efficient Advanced Driver Assistance Technique (ADAT) by utilizing the driver's attention. A real-time gaze tracking system is used to obtain the driver's gaze information, and a new model is developed to accurately detect and estimate head pose directions and eye directions.
Car crashes are among the top ten leading causes of death; they could mainly be attributed to distracted drivers. An advanced driver-assistance technique (ADAT) is a procedure that can notify the driver about a dangerous scenario, reduce traffic crashes, and improve road safety. The main contribution of this work involved utilizing the driver's attention to build an efficient ADAT. To obtain this attention value, the gaze tracking method is proposed. The gaze direction of the driver is critical toward understanding/discerning fatal distractions, pertaining to when it is obligatory to notify the driver about the risks on the road. A real-time gaze tracking system is proposed in this paper for the development of an ADAT that obtains and communicates the gaze information of the driver. The developed ADAT system detects various head poses of the driver and estimates eye gaze directions, which play important roles in assisting the driver and avoiding any unwanted circumstances. The first (and more significant) task in this research work involved the development of a benchmark image dataset consisting of head poses and horizontal and vertical direction gazes of the driver's eyes. To detect the driver's face accurately and efficiently, the You Only Look Once (YOLO-V4) face detector was used by modifying it with the Inception-v3 CNN model for robust feature learning and improved face detection. Finally, transfer learning in the InceptionResNet-v2 CNN model was performed, where the CNN was used as a classification model for head pose detection and eye gaze angle estimation; a regression layer to the InceptionResNet-v2 CNN was added instead of SoftMax and the classification output layer. The proposed model detects and estimates head pose directions and eye directions with higher accuracy. The average accuracy achieved by the head pose detection system was 91%; the model achieved a RMSE of 2.68 for vertical and 3.61 for horizontal eye gaze estimations.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据