Abstract
Eye gaze tracking is an increasingly important technology in the field of human-computer interaction. Individuals’ preferences, tendencies, and attention can be measured by processing the data obtained from face and eye images. This technology is used in advertising, market research, web page design, education, learning methods, and various neurological-psychiatric studies of medical research. Many different methods have been used in eye gaze tracking tasks. Today, commonly model-shape and appearance-based methods are used. Model-shape based methods require less workload than appearance-based methods. But it is more sensitive to environmental conditions. Appearance-based methods require powerful hardware, but they are less susceptible to environmental conditions. Developments in technology have paved the way for applying appearance-based models in eye gaze tracking. In this paper, a CNN-based real-time eye tracking system was designed to overcome environmental problems in eye gaze tracking. The designed system is used to determine the areas of interest of the user in web pages. The performance of the designed CNN-based system is evaluated during the training and testing phases. In the training phase, the difference between the desired and determined points on the screen is 32 pixels and in testing phase, the difference between the desired and determined points on the screen is 53 pixels. The results of the test trials have shown that the proposed system could be used successfully in eye tracking studies on web pages.















Similar content being viewed by others
References
Akhtar N, Ragavendran U (2020) Interpretation of intelligence in CNN-pooling processes: a methodological survey. Neural Comput Appl 32(3):879–898
Aljaafreh A, Alaqtash M, Al-Oudat N, Abukhait J, Saleh M (2020) A low-cost webcam-based eye tracker and saccade measurement system. Int J Circuits Syst Signal Process 14. https://doi.org/10.46300/9106.2020.14.16
Baluja, S. and Pomerleau, D. (1993) Non-Intrusive Gaze Tracking using Artificial Neural Networks. In Cowan J.D., Tesauro, G. and Alspector, J. (eds.) Advances in Neural Information Processing Systems, (NIPS) 6. 1994. pp. 753–760. Morgan Kaufmann Publishers, San Francisco, CA.
Barea R, Boquete L, Mazo M, López E (2002) System for assisted mobility using eye movements based on electrooculography. IEEE Trans Neural Syst Rehabil Eng 10(4):209–218
Chen J, Ji Q (2008) 3D gaze estimation with a single camera without IR illumination. Proceedings-International Conference on Pattern Recognition, pp 1–4. https://doi.org/10.1109/icpr.2008.4761343
Chen HH, Hwang BJ, Wu JS, Liu PT (2020) The effect of different deep network architectures upon CNN-based gaze tracking. Algorithms 13(5). https://doi.org/10.3390/A13050127
Collewijn H, van der Mark F, Jansen TC (1975) Precise recording of human eye movements. Vision Res 15(3):447–450. https://doi.org/10.1016/0042-6989(75)90098-X
Cooles TF, Edwards GJ, Taylor CJ (2001) Active appearance models. IEEE Trans Pattern Anal Mach Intell 23(6):681–685. https://doi.org/10.1109/34.927467
Cun YL, Guyon I, Jackel LD, Henderson D, Boser B, Howard RE, Denker JS, Hubbard W, Graf HP (1989) Handwritten digit recognition: applications of neural network chips and automatic learning. IEEE Commun Mag 27(11):41–46. https://doi.org/10.1109/35.41400
dlib C + + Library. http://www.dlib.net. Accessed 23 Sept 2020
Damji O, Lee-Nobbee P, Borkenhagen D, Cheng A (2019) Analysis of eye-tracking behaviours in a pediatric trauma simulation. Can J Emerg Med 21(1):138–140
de Araújo FMA, Fonseca Ferreira NM, Mascarenhas VTOC, Adad Filho JA, Viana Filho PRF (2020) Eye tracking in framework for the development of games for people with motor disabilities. Lect Notes Electr Eng 621:239–247. https://doi.org/10.1007/978-981-15-1465-4_25
Doğan F, Türkoğlu İ (2019) Derin Öğrenme Modelleri ve Uygulama Alanlarına İlişkin Bir Derleme. DÜMF Mühendislik Dergisi 10(2):409–445. https://doi.org/10.24012/dumf.411130
Donuk K, Hanbay D (2020) Video Based Real-Time Eye Tracking. In: 28th IEEE Signal Processing and Communications Applications (SIU-2020), pp 21–24, Turkey
Duchowski AT (2017) In eye tracking methodology: Theory and practice: 3rd edn. Springer International Publishing, Berlin. https://doi.org/10.1007/978-3-319-57883-5
Dutta P, Upadhyay P, De M, Khalkar RG (2020) Medical Image Analysis using Deep Convolutional Neural Networks: CNN architectures and transfer learning. Proceedings of the 5th International Conference on Inventive Computation Technologies, ICICT 2020, pp 175–180
Elmas, Ç. (2011). “Yapay Zekâ Uygulamaları: Yapay Sinir Ağı, Bulanık Mantık, Genetik Algoritma”, Seçkin Yayıncılık,Ankara.
Eye Tracking System Technology For Everyone - UX Testing | Gazepoint. https://www.gazept.com. Accessed 30 Dec 2020
Fukushima K (1980) Neocognitron: A self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position. Biol Cybern 36(4):193–202
Gurgu E, Gurgu I-A, Tonis RBM (2020) Neuromarketing for a better understanding of consumer needs and emotions. Indep J Manage Prod 11(1):208
Hansen DW, Ji Q (2010) In the eye of the beholder: a survey of models for eyes and gaze. IEEE Trans Pattern Anal Mach Intell 32(3):478–500
Hinton GE, Salakhutdinov RR (2006) Reducing the dimensionality of data with neural networks. Science 313(5786):504–507. https://doi.org/10.1126/science.1127647
Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780
Home - OpenCV. https://opencv.org/. Accessed 30 Dec 2020
Hung JC, Wang CC (2020) Exploring the website object layout of responsive web design: results of eye tracking evaluations. J Supercomput 1–23. https://doi.org/10.1007/s11227-020-03283-1
i·bug-Facial point annotations. https://ibug.doc.ic.ac.uk/resources/facial-point-annotations/. Accessed 23 Sept 2020
Ioffe S, Szegedy C (2015) Batch normalization: Accelerating deep network training by reducing internal covariate shift. 32nd International Conference on Machine Learning, ICML 2015, vol 1
Ishikawa T, Baker S, Matthews I, Kanade T (2004) Passive driver gaze tracking with active appearance models. Proc. World Congress on Intelligent Transportation Systems
Ivakhnenko, A.; Lapa, V. (1965) Cybernetic predicting devices; CCM Information Corp.: New York, NY, USA.
Janssen RGJ, Rothkrantz L, Wiggers IP, Geers IH (2010) Real time eye blink detection using a configurable processor. Ph.D. Dissertation, Doctoral Dissertation, TU Delft, Delft University of Technology
Jie HJ, Wanda P (2020) Runpool: A dynamic pooling layer for convolution neural network. Int J Comput Intell Syst 13(1):66–76. https://doi.org/10.2991/ijcis.d.200120.002
Kazemi V, Sullivan J (2014) One millisecond face alignment with an ensemble of regression trees. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition. https://doi.org/10.1109/CVPR.2014.241
Kim M, Wang O, Ng N (2017) Convolutional neural network architectures for gaze. Estimation on Mobile Devices, Stanford
Kimme C, Ballard D, Sklansky J (1975) Finding circles by an array of accumulators. Commun ACM 18(2):120–122. https://doi.org/10.1145/360666.360677
Kingma DP, Ba JL (2015) Adam: A method for stochastic optimization. 3rd International Conference on Learning Representations, ICLR 2015- Conference Track Proceedings
Krafka K, Khosla A, Kellnhofer P, Kannan H, Bhandarkar S, Matusik W, Torralba A (2016) Eye tracking for everyone. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition. https://doi.org/10.1109/CVPR.2016.239
Krizhevsky A, Sutskever I, Hinton GE (2012) ImageNet classification with deep convolutional neural networks. Adv Neural Inf Process Syst 2:1097–1105
LeCun Y, Boser B, Denker J, Henderson D, Howard R, Hubbard W, Jackel L (1989) Handwritten digit recognition with a back-propagation network. pp 396–404
LeCun Y, Bottou L, Bengio Y, Haffner P (1998)Gradient-based learning applied to document recognition. Proc IEEE 86(11):2278–2324. https://doi.org/10.1109/5.726791
Lu F, Sugano Y, Okabe T, Sato Y (2014) Adaptive linear regression for appearance-based gaze estimation. IEEE Trans Pattern Anal Mach Intell 36(10):2033–2046
Majaranta P, Bulling A (2014) Eye Tracking and Eye-BasedHuman–Computer Interaction. 39–65. https://doi.org/10.1007/978-1-4471-6392-3_3
Martens J, Sutskever I (2011) Learning recurrent neural networks with Hessian-free optimization. Proceedings of the 28th International Conference on Machine Learning, ICML 2011
Massin L, Nourrit V, Lahuec C, Seguin F, Adam L, Daniel E, De J-L, De B, Tocnaye LA (2020) Development of a new scleral contact lens with encapsulated photodetectors for eye tracking. Opt Express 28(19):28635–28647. https://doi.org/10.1364/OE.399823
Mora KAF, Monay F, Odobez JM (2014) EYEDIAP: A database for the development and evaluation of gaze estimation algorithms from RGB and RGB-D cameras. Eye Tracking Research and Applications Symposium (ETRA), pp 255–258. https://doi.org/10.1145/2578153.2578190
Morimoto CH, Amir A, Flickner M (2002) Detecting eye position and gaze from a single camera and 2 light sources. Proceedings - International Conference on Pattern Recognition 16(4):314–317
Nair V, Hinton GE (2010) Rectified linear units improve restricted Boltzmann machines. In: ICML’10: Proceedings of the 27th International Conference on International Conference on Machine Learning, pp 807–814
Niemann J, Fussenecker C, Schlösser M (2019) Eye tracking for quality control in automotive manufacturing. Commun Comput Inform Sci 1060:289–298
Pfiffelmann J, Dens N, Soulez S (2020) Personalized advertisements with integration of names and photographs: An eye-tracking experiment. J Bus Res 111:196–207
Rayner K (1998) Eye movements in reading and information processing: 20 years of research. Psychol Bull 124(3). https://doi.org/10.1037/0033-2909.124.3.372
Reyes F, Fajardo A, Hernandez A (2020) Convolutional neural network for automatic speech recognition of Filipino language. Int J Adv Trends Comput Sci Eng 9(1.1 Special Issue):34–40. https://doi.org/10.30534/ijatcse/2020/0791.12020
Robinson DA (1963) A method of measuring eye movement using a scleral search coil in a magnetic field. IEEE Trans Bio-med Electron 10(4):137–145
Rosenblatt F (1958) The perceptron: A probabilistic model for information storage and organization in the brain. Psychol Rev 65(6):386–408. https://doi.org/10.1037/h0042519
Salakhutdinov, R. and Hinton, G. (2009). Deep Boltzmann machines. In Proceedings of the Twelfth International Conference on Artificial Intelligence and Statistics (AISTATS 2009), vol. 8
Saraee E, Jalal M, Betke M (2020) Visual complexity analysis using deep intermediate-layer features. Comput Vis Image Underst 195:102949. https://doi.org/10.1016/j.cviu.2020.102949
Schmidhuber J (2015) Deep Learning in neural networks: An overview. Neural Netw 61:85–117. https://doi.org/10.1016/j.neunet.2014.09.003
Şeker A, Diri B, Balık H (2017) Derin Öğrenme Yöntemleri ve Uygulamaları Hakkında Bir İnceleme. Gazi J Eng Sci 3(3):47–64
Sewell W, Komogortsev O (2010)Real-time eye gaze tracking with an unmodified commodity webcam employing a neural network. Conference on Human Factors in Computing Systems – Proceedings, pp 3739–3744. https://doi.org/10.1145/1753846.1754048
Shokishalov Z, Wang H (2019) Applying eye tracking in information security. Procedia Comput Sci 150:347–351. https://doi.org/10.1016/j.procs.2019.02.062
Slanzi G, Balazs JA, Velásquez JD (2017) Combining eye tracking, pupil dilation and EEG analysis for predicting web users click intention. Inform Fusion 35:51–57
Specializing in eye tracking - EyeLink Eye Trackers - SR Research. https://www.sr-research.com. Accessed 30 Dec 2020
Sugano Y, Matsushita Y, Sato Y (2014)Learning-by-synthesis for appearance-based 3D gaze estimation. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp 1821–1828. https://doi.org/10.1109/CVPR.2014.235
Sultana F, Sufian A, Dutta P (2020) A review of object detection models based on convolutional neural network. In: Advances in Intelligent Systems and Computing 1157:1–16. Springer
Sun S, Hu B, Yu Z, Song XN (2020) A stochastic max pooling strategy for convolutional neural network trained by noisy samples. Intern J Comput Commun Contr 15(1):9
Suzuki S, Be KA (1985) Topological structural analysis of digitized binary images by border following. Comput Vis Graphics Image Process 30(1):32–46
Tobii is the world leader in eye tracking. https://www.tobii.com. Accessed 30 Dec 2020
Tsantekidis A, Passalis N, Tefas A, Kanniainen J, Gabbouj M, Iosifidis A (2020) Using Deep Learning for price prediction by exploiting stationary limit order book features. Appl Soft Comput J 93. https://doi.org/10.1016/j.asoc.2020.106401
Uhm KH, Kang MC, Kim JY, Ko SJ (2020) Improving the robustness of gaze tracking under unconstrained illumination conditions. Multimed Tools Appl 79(29–30):20603–20616
Velásquez JD (2013) Combining eye-tracking technologies with web usage mining for identifying Website Keyobjects. Eng Appl Artif Intell 26(5–6):1469–1478
Vincent P, Larochelle H, Bengio Y, Manzagol PA (2008) Extracting and composing robust features with denoising autoencoders. Proceedings of the 25th International Conference on Machine Learning, pp 1096–1103. https://doi.org/10.1145/1390156.1390294
Viola P, Jones M (2001) Rapid object detection using a boosted cascade of simple features. Proc IEEE Comput Soc Conf Comput Vision Pattern Recogn 1:511–518
Welcome to Colaboratory - Colaboratory. https://colab.research.google.com. Accessed 3 Nov 2020
Whitmire E, Trutoiu L, Cavin R, Perek D, Scally B, Phillips J, Patel S (2016) EyeContact: Scleral coil eye tracking for virtual reality. International Symposium on Wearable Computers, pp 184–191
Wilkinson L, Friendly M (2009) History corner the history of the cluster heat map. Am Stat 63(2):179–184. https://doi.org/10.1198/tas.2009.0033
Wong ET, Yean S, Hu Q, Lee BS, Liu J, Deepu R, Workshops C (2019) PerCom Workshops 2019. https://doi.org/10.1109/PERCOMW.2019.8730846
Xu L-Q, Machin D, Sheppard P (1998) A novel approach to real-time non-intrusive gaze finding. In: Proc. BMVC. https://doi.org/10.5244/c.12.43
Ye L, Zhu M, Xia S, Pan H (2014) Cascaded convolutional neural network for eye detection under complex scenarios. Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture. Notes in Bioinformatics) 8833:473–480. https://doi.org/10.1007/978-3-319-12484-1_54
Yoo DH, Chung MJ (2005) A novel non-intrusive eye gaze estimation using cross-ratio under large head motion. Comput Vis Image Underst 98(1):25–51
Zhang X, Sugano Y, Fritz M, Bulling A (2015)Appearance-based gaze estimation in the wild. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp 4511–4520. https://doi.org/10.1109/CVPR.2015.7299081
Zhang K, Zhang Z, Li Z, Qiao Y (2016) Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Signal Process Lett 23(10):1499–1503
Zhang Z, Zhao Y, Liao X, Shi W, Li K, Zou Q, Peng S (2019) Deep learning in omics: A survey and guideline. Brief Funct Genomics 18(1):41–57. https://doi.org/10.1093/bfgp/ely030
Zhang X, Wang L, Su Y (2020) Visual place recognition: a survey from deep learning perspective pattern recognition. https://doi.org/10.1016/j.patcog.2020.107760
Zhou Y (2020) Psychological analysis of online teaching in colleges based on eye-tracking technology. Rev Argenti Clin Psicol 29(2):523–529. https://doi.org/10.24205/03276716.2020.272
Funding
This study was supported by Inonu University Scientific Research Projects Coordination Unit (BAP) with the project coded FDK-2020-2110.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that there is no conflict to interest related to this paper.
Ethical approval
Not required.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Donuk, K., Ari, A. & Hanbay, D. A CNN based real-time eye tracker for web mining applications. Multimed Tools Appl 81, 39103–39120 (2022). https://doi.org/10.1007/s11042-022-13085-7
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11042-022-13085-7