skip to main content
10.1145/3334480.3382856acmconferencesArticle/Chapter ViewAbstractPublication PageschiConference Proceedingsconference-collections
abstract

A Preliminary Study on Performance Evaluation of Multi-View Multi-Modal Gaze Estimation under Challenging Conditions

Authors Info & Claims
Published:25 April 2020Publication History

ABSTRACT

In this paper, we address gaze estimation under practical and challenging conditions. Multi-view and multi-modal learning have been considered useful for various complex tasks; however, an in-depth analysis or a large-scale dataset on multi-view, multi-modal gaze estimation under a long-distance setup with a low illumination is still very limited. To address these limitations, first, we construct a dataset of images captured under challenging conditions. And we propose a simple deep learning architecture that can handle multi-view multi-modal data for gaze estimation. Finally, we conduct a performance evaluation of the proposed network with the constructed dataset to understand the effects of multiple views of a user and multi-modality (RGB, depth, and infrared). We report various findings from our preliminary experimental results and expect this would be helpful for gaze estimation studies to deal with challenging conditions.

References

  1. National Institute on Aging. 2019. Social isolation, loneliness in older people pose health risks. Retrieved January 5, 2020 from https://www.nia.nih.gov/news/social-isolationloneliness-older-people-pose-health-risksGoogle ScholarGoogle Scholar
  2. Xucong Zhang, Yusuke Sugano, Mario Fritz, and Andreas Bulling. 2019. MPIIGaze: Real-World Dataset and Deep Appearance-Based Gaze Estimation. IEEE Transactions on Pattern Analysis and Machine Intelligence 41, 1: 162--175.Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Yusuke Sugano, Yasuyuki Matsushita, and Yoichi Sato. 2014. Learning-by-synthesis for appearance-based 3D gaze estimation. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, IEEE Computer Society, 1821--1828. https://doi.org/10.1109/CVPR.2014.235Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Tobias Fischer, Hyung Jin Chang, and Yiannis Demiris. 2018. RT-GENE: Real-time eye gaze estimation in natural environments. In proceedings of the European Conference on Computer Vision (ECCV '18), 334--352. https://doi.org/10.1007/978--3-030-01249--6_21Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Petr Kellnhofer, Adria Recasens, Simon Stent, Wojciech Matusik, and Antonio Torralba. 2019. Gaze360: Physically Unconstrained Gaze Estimation in the Wild. arXiv:1910.10088Google ScholarGoogle Scholar
  6. Benoit Massé, Stéphane Lathuilière, Pablo Mesejo, and Radu Horaud. 2019. Extended Gaze Following: Detecting Objects in Videos Beyond the Camera Field of View. arXiv:1902.10953Google ScholarGoogle Scholar
  7. Dongze Lian, Lina Hu, Weixin Luo, Yanyu Xu, and Lixin Duan. 2019. Multiview Multitask Gaze Estimation With Deep Convolutional Neural Networks. IEEE Transactions on Neural Networks and Learning Systems 30, 10: 3010--3023.Google ScholarGoogle ScholarCross RefCross Ref
  8. Jie Shiou Tsai and Chang Hong Lin. 2018. Gaze direction estimation using only a depth camera. In Proceedings of the International Conference on Intelligent Green Building and Smart Grid (IGBSG), Institute of Electrical and Electronics Engineers Inc., 1--2. https://doi.org/10.1109/IGBSG.2018.8393539Google ScholarGoogle ScholarCross RefCross Ref
  9. Kai Su, Dongdong Yu, Zhenqi Xu, Xin Geng, and Changhu Wang. 2019. Multi-Person Pose Estimation with Enhanced Channel-wise and Spatial Information. arXiv:1905.03466Google ScholarGoogle Scholar
  10. Hao Tang, Dan Xu, Yan Yan, Jason J.Corso, Philip H.S. Torr, and Nicu Sebe. 2020. Multi-Channel Attention Selection GANs for Guided Image-toImage Translation. arXiv:2002.01048Google ScholarGoogle Scholar
  11. Chaoqun Hong, Jun Yu, Jian Zhang, Xiongnan Jin, and KYong-Ho Lee. 2018. Multimodal Face-Pose Estimation with Multitask Manifold Deep Learning. IEEE Transactions on Industrial Informatics 15, 7:3952--3961. https://doi.org/10.1109/TII.2018.2884211Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. A Preliminary Study on Performance Evaluation of Multi-View Multi-Modal Gaze Estimation under Challenging Conditions

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      CHI EA '20: Extended Abstracts of the 2020 CHI Conference on Human Factors in Computing Systems
      April 2020
      4474 pages
      ISBN:9781450368193
      DOI:10.1145/3334480

      Copyright © 2020 Owner/Author

      Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 25 April 2020

      Check for updates

      Qualifiers

      • abstract

      Acceptance Rates

      Overall Acceptance Rate6,164of23,696submissions,26%

      Upcoming Conference

      CHI '24
      CHI Conference on Human Factors in Computing Systems
      May 11 - 16, 2024
      Honolulu , HI , USA

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format .

    View HTML Format