Abstract
Research and development of autonomous driving technology is accelerating in the automotive industry. Currently, drivers of such vehicles are considered to pay less attention to environmental conditions while being driven, due to potential overestimations of autonomous driving functionality and its reliability. In this paper, methods to quantitatively measure the driver’s gaze behavior are proposed, followed by investigation methodology and results on how auditory warning signals influence the behavior, where the difference between novice and experienced drivers is also compared.
References
Sodnik, J., Dicke, C., Tomazic, S., Billinghurst, M.: A user study of auditory versus visual interfaces for use while driving. Int. J. Hum.-Comput. Stud. 66, 318–332 (2008)
Larsson, P., Niemand, M.: Using sound to reduce visual distraction from in-vehicle human-machine interfaces. Traffic Inj. Prev. 16(1), S25–S30 (2015)
SAE International: “Automated Driving Levels of Driving Automation are Defined in New SAE International Standard J3016”
Wallach, H.: On sound localization. J. Acoust. Soc. Am. 10, 270–274 (1939)
Toyama, T., Kieninger, T., Shafait, F., Dengel, A.: Gaze guided object recognition using a head-mounted eye tracker. In: Proceedings of 7th ACM Symposium on Eye Tracking Research & Applications (ETRA2012), pp. 91–98 (2012)
Takemura, K., Kohashi, Y., Suenaga, T., Takamatsu, J., Ogasawara, T.: Estimating 3D point-of-regard and visualizing gaze trajectories under natural head movements. In: Proceedings of 6th ACM Symposium on Eye Tracking Research & Applications (ETRA2010), pp. 157–160 (2012)
Chanijani, S.S.M., Al-Naser, M., Bukhari, S.S., Borth, D., Allen, S.E.M., Dengel, A.: An eye movement study on scientific papers using wearable eye tracking technology. In: 9th International Conference on Mobile Computing and Ubiquitous Networking (ICMU) (2016)
NAC Image Technology, EMR-dStream. http://www.eyemark.jp/product/emr_dstream/
Tomi, A.B., Rambli, D.R.A.: Automated calibration for optical see-through head mounted display using display screen space based eye tracking. In: 3rd International Conference on Computer and Information Science (ICCOINS), pp. 448–453 (2016)
Huang, C.W., Tan, W.C.: An approach of head movement compensation when using a head mounted eye tracker. In: International Conference of Consumer Electronics-Taiwan (2016)
Kocejko, T., Bujnowski, A., Ruminski, J., Bylinska, E., Wtorek, J.: Head movement compensation algorithm in multi-display communication by gaze. In: 7th International Conference on Human System Interactions (HSI), pp. 88–94 (2014)
Sohn, B., Lee, J., Chae, H., Yu, W.: Localization system for mobile robot using wireless communication with IR landmark. In: Proceedings of the 1st International Conference on Robot Communication and Coordination, pp. 1–6 (2007)
Hu, M.K.: Visual pattern recognition by moment invariants. IRE Trans. Inf. Theory 8(2), 179–187 (1962)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer International Publishing AG
About this paper
Cite this paper
Osawa, R., Imafuku, S., Shirayama, S. (2017). A Basic Study of Gaze Behavior Measurement Methodology for Drivers in Autonomous Vehicles. In: Kurahashi, S., Ohta, Y., Arai, S., Satoh, K., Bekki, D. (eds) New Frontiers in Artificial Intelligence. JSAI-isAI 2016. Lecture Notes in Computer Science(), vol 10247. Springer, Cham. https://doi.org/10.1007/978-3-319-61572-1_21
Download citation
DOI: https://doi.org/10.1007/978-3-319-61572-1_21
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-61571-4
Online ISBN: 978-3-319-61572-1
eBook Packages: Computer ScienceComputer Science (R0)