skip to main content
10.1145/2557500.2557528acmconferencesArticle/Chapter ViewAbstractPublication PagesiuiConference Proceedingsconference-collections
poster

A mixed reality head-mounted text translation system using eye gaze input

Published: 24 February 2014 Publication History

Abstract

Efficient text recognition has recently been a challenge for augmented reality systems. In this paper, we propose a system with the ability to provide translations to the user in real-time. We use eye gaze for more intuitive and efficient input for ubiquitous text reading and translation in head mounted displays (HMDs). The eyes can be used to indicate regions of interest in text documents and activate optical-character-recognition (OCR) and translation functions. Visual feedback and navigation help in the interaction process, and text snippets with translations from Japanese to English text snippets, are presented in a see-through HMD. We focus on travelers who go to Japan and need to read signs and propose two different gaze gestures for activating the OCR text reading and translation function. We evaluate which type of gesture suits our OCR scenario best. We also show that our gaze-based OCR method on the extracted gaze regions provide faster access times to information than traditional OCR approaches. Other benefits include that visual feedback of the extracted text region can be given in real-time, the Japanese to English translation can be presented in real-time, and the augmentation of the synchronized and calibrated HMD in this mixed reality application are presented at exact locations in the augmented user view to allow for dynamic text translation management in head-up display systems.

References

[1]
J. M. Henderson. Human gaze control during real-world scene perception. Trends in Cognitive Sciences, 7(11):498--504, 2003.
[2]
M. Hopmann, P. Salamin, N. Chauvin, F. Vexo, and D. Thalmann. Natural activation for gesture recognition systems. In CHI fi11 Extended Abstracts on Human Factors in Computing Systems, CHI EA fi11, pages 173--183, New York, NY, USA, 2011. ACM.
[3]
R. J. K. Jacob. Eye movement-based human-computer interaction techniques: Toward non-command interfaces. In Advances in Human-Computer Interaction, volume 4, pages 151--190. Ablex Publishing Co, 1993.
[4]
T. Kobayahsi, M. Iwamura, and K. Kise. An anytime algorithm for faster camera-based character recognition. In Proceedings of the 12th International Conference on Document Analysis and Recognition (ICDAR), 2013.
[5]
T. Kobayashi, T. Toyamaya, F. Shafait, M. Iwamura, K. Kise, and A. Dengel. Recognizing words in scenes with a head-mounted eye-tracker. IAPR International Workshop on Document Analysis Systems, pages 333--338, 2012.
[6]
M. Kumar, A. Paepcke, and T. Winograd. Eyepoint: Practical pointing and selection using gaze and keyboard. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, CHI '07, pages 421--430, New York, NY, USA, 2007. ACM.
[7]
J. Orlosky, K. Kiyokawa, and H. Takemura. Dynamic text management for see-through wearable and heads-up display systems. In Proceedings of the International Conference on Intelligent user interfaces, IUI '13, pages 363--370, New York, NY, USA, 2013. ACM.
[8]
Z. Prasov and J. Y. Chai. What's in a gazefi: the role of eye-gaze in reference resolution in multimodal conversational interfaces. In Proceedings of the 13th international conference on Intelligent user interfaces, IUI '08, pages 20--29, New York, NY, USA, 2008. ACM.
[9]
P. Qvarfordt and S. Zhai. Conversing with the user based on eye-gaze patterns. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, CHI '05, pages 221--230, New York, NY, USA, 2005. ACM.
[10]
K. Rayner. Eye movements in reading and information processing: 20 years of research. Psychological Bulletin, pages 372--422, 1998.
[11]
J. Shi and C. Tomasi. Good features to track. In Computer Vision and Pattern Recognition, 1994. Proceedings CVPR '94., 1994 IEEE Computer Society Conference on, pages 593--600, 1994.
[12]
D. Sonntag and T. Toyama. On-Body IE: A Head-Mounted Multimodal Augmented Reality System for Learning and Recalling Faces. In 9th International Conference on Intelligent Environments (IE), pages 151--156, 2013.
[13]
T. Toyama, T. Kieninger, F. Shafait, and A. Dengel. Gaze guided object recognition using a head-mounted eye tracker. In Proc. of the Symposium on Eye Tracking Research and Applications, pages 91--98, 2012.
[14]
W. Wahlster, editor. VERBMOBIL: Foundations of Speech-to-Speech Translation. Springer, 2000.

Cited By

View all
  • (2025)Adaptive Real-Time Translation Assistance Through Eye-TrackingAI10.3390/ai60100056:1(5)Online publication date: 2-Jan-2025
  • (2024)MRTranslate: Bridging Language Barriers in the Physical World Using a Mixed Reality Point-and-Translate SystemProceedings of the 2024 International Conference on Advanced Visual Interfaces10.1145/3656650.3656652(1-9)Online publication date: 3-Jun-2024
  • (2024)A natural bare-hand interface-enabled interactive AR assembly guidanceThe International Journal of Advanced Manufacturing Technology10.1007/s00170-024-13922-z133:7-8(3193-3207)Online publication date: 12-Jun-2024
  • Show More Cited By

Index Terms

  1. A mixed reality head-mounted text translation system using eye gaze input

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    IUI '14: Proceedings of the 19th international conference on Intelligent User Interfaces
    February 2014
    386 pages
    ISBN:9781450321846
    DOI:10.1145/2557500
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 24 February 2014

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. augmented reality and projection
    2. mobile and embedded devices
    3. smart environments
    4. ubiquitous computing
    5. visualization

    Qualifiers

    • Poster

    Conference

    IUI'14
    Sponsor:

    Acceptance Rates

    IUI '14 Paper Acceptance Rate 46 of 191 submissions, 24%;
    Overall Acceptance Rate 746 of 2,811 submissions, 27%

    Upcoming Conference

    IUI '25

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)39
    • Downloads (Last 6 weeks)4
    Reflects downloads up to 28 Feb 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2025)Adaptive Real-Time Translation Assistance Through Eye-TrackingAI10.3390/ai60100056:1(5)Online publication date: 2-Jan-2025
    • (2024)MRTranslate: Bridging Language Barriers in the Physical World Using a Mixed Reality Point-and-Translate SystemProceedings of the 2024 International Conference on Advanced Visual Interfaces10.1145/3656650.3656652(1-9)Online publication date: 3-Jun-2024
    • (2024)A natural bare-hand interface-enabled interactive AR assembly guidanceThe International Journal of Advanced Manufacturing Technology10.1007/s00170-024-13922-z133:7-8(3193-3207)Online publication date: 12-Jun-2024
    • (2023)Multi-Functional Glasses for the Blind and Visually Impaired: Design and DevelopmentProceedings of the Human Factors and Ergonomics Society Annual Meeting10.1177/2169506723119245067:1(995-1001)Online publication date: 19-Oct-2023
    • (2022)AtAwAR Translate: Attention-Aware Language Translation Application in Augmented Reality for Mobile PhonesSensors10.3390/s2216616022:16(6160)Online publication date: 17-Aug-2022
    • (2022)SOCRAR: Semantic OCR through Augmented RealityProceedings of the 12th International Conference on the Internet of Things10.1145/3567445.3567453(25-32)Online publication date: 7-Nov-2022
    • (2022)Interactive Extended Reality Techniques in Information VisualizationIEEE Transactions on Human-Machine Systems10.1109/THMS.2022.321131752:6(1338-1351)Online publication date: Dec-2022
    • (2022)Design Eye-Tracking Augmented Reality Headset to Reduce Cognitive Load in Repetitive Parcel Scanning TaskIEEE Transactions on Human-Machine Systems10.1109/THMS.2022.317995452:4(578-590)Online publication date: Aug-2022
    • (2021)Ubiquitous Interactions for Heads-Up Computing: Understanding Users’ Preferences for Subtle Interaction Techniques in Everyday SettingsProceedings of the 23rd International Conference on Mobile Human-Computer Interaction10.1145/3447526.3472035(1-15)Online publication date: 27-Sep-2021
    • (2020)Effects of position of real-time translation on AR glassesProceedings of Mensch und Computer 202010.1145/3404983.3405523(251-257)Online publication date: 6-Sep-2020
    • Show More Cited By

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media