Skip to main content

Augmented Reality Using Gesture and Speech Accelerates User Interaction

  • Conference paper
  • First Online:
Advanced Communication and Intelligent Systems (ICACIS 2022)

Abstract

A user interaction technique is needed to manipulate the Augmented Reality (AR) content to make AR deliver realism. However, these problems persist, especially when they involve natural user interaction such as real hands and speech. Speech might be the most crucial part, especially in detecting the user’s voice. The user’s voice usually contains noises that come from the surroundings and every people produces different kinds of voice vibrations. Different genders also produce different voice vibrations. Robust real hand gestures are required to ensure the AR can accelerate the natural user interface. By providing users with more than one way to interact with AR, it tends to produce an efficient way of interaction. This chapter discusses the approach involves three main phases. In the first phase, study the interaction metaphor using gesture and speech in AR. After that, the second phase is carried out to perform a test application for the gesture and speech interaction. This chapter presents an appropriate interaction method that accelerates user interaction, it describes the implementation of gesture and speech interaction to the next level to see how AR could accelerate user interaction.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 99.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 129.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Farahani, N., et al.: Exploring virtual reality technology and the Oculus Rift for the examination of digital pathology slides. J. Pathol. Inform. 7, 22 (2016)

    Google Scholar 

  2. Ratcliffe, J., Soave, F., Bryan-Kinns, N., Tokarchuk, L., Farkhatdinov, I.: Extended Reality (XR) remote research: a survey of drawbacks and opportunities. In: Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems, pp. 1–13, May 2021

    Google Scholar 

  3. Tepper, O.M., et al.: Mixed reality with HoloLens: where virtual reality meets augmented reality in the operating room. Plast. Reconstr. Surg. 140(5), 1066–1070 (2017)

    Article  Google Scholar 

  4. Nor’a, M.N.A., Ismail, A.W., Aladin, M.Y.F.: Interactive augmented reality pop-up book with natural gesture interaction for handheld. In: Lee, N. (eds.) Encyclopedia of Computer Graphics and Games, pp. 1–10. Springer, Cham (2019). https://doi.org/10.1007/978-3-319-08234-9_365-1

  5. Liu, W.: Natural user interface-next mainstream product user interface. In: 2010 IEEE 11th International Conference on Computer-Aided Industrial Design & Conceptual Design 1, vol. 1, pp. 203–205. IEEE, November 2010

    Google Scholar 

  6. Norman, D.A.: Natural user interfaces are not natural. Interactions 17(3), 6–10 (2010)

    Google Scholar 

  7. Fu, L.P., Landay, J., Nebeling, M., Xu, Y., Zhao, C.: Redefining natural user interface. In: Extended Abstracts of the 2018 CHI Conference on Human Factors in Computing Systems, pp. 1–3, April 2018

    Google Scholar 

  8. Chun, J., Lee, S.: A vision-based 3D hand interaction for marker-based AR. Int. J. Multimedia Ubiquitous Eng. 7(3), 51–58 (2012)

    Google Scholar 

  9. Quek, F., et al.: Multimodal human discourse: gesture and speech. ACM Trans. Comput. Hum. Interact. 9(3), 171–193 (2002)

    Google Scholar 

  10. Piumsomboon, T., Altimira, D., Kim, H., Clark, A., Lee, G., Billinghurst, M.: Grasp-Shell vs gesture-speech: a comparison of direct and indirect natural interaction techniques in augmented reality. In: ISMAR 2014 - IEEE International Symposium on Mixed and Augmented Reality - Science and Technology 2014, Proceedings, pp. 73–82 (2014)

    Google Scholar 

  11. Malkawi, A.M., Srinivasan, R.S.: Multimodal human-computer interaction for immersive visualization: integrating speech-gesture recognitions and augmented reality for indoor environments multimodal HCI-based visualization model immersive (2004)

    Google Scholar 

  12. Ismail, A.W., Billinghurst, M., Sunar, M.S., Yusof, C.S.: Designing an augmented reality multimodal interface for 6DOF manipulation techniques. In: Arai, K., Kapoor, S., Bhatia, R. (eds.) IntelliSys 2018. AISC, vol. 868, pp. 309–322. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-01054-6_22

  13. Nor’a, M.N.A., et al.: Fingertips interaction method in handheld augmented reality for 3D manipulation. In: 2020 IEEE 5th International Conference on Computing Communication and Automation (ICCCA), pp. 161–166. IEEE, October 2020

    Google Scholar 

  14. Aladin, M.Y.F., Ismail, A.W., Ismail, N.A., Rahim, M.S.M.: Object selection and scaling using multimodal interaction in mixed reality. In: IOP Conference Series: Materials Science and Engineering, vol. 979, no. 1, p. 012004. IOP Publishing, November 2020

    Google Scholar 

  15. Plopski, A., Hirzle, T., Norouzi, N., Qian, L., Bruder, G., Langlotz, T.: The eye in extended reality: a survey on gaze interaction and eye tracking in head-worn extended reality. ACM Comput. Surv. (CSUR) 55(3), 1–39 (2022)

    Article  Google Scholar 

Download references

Acknowledgement

We appreciate ViCubeLab at Universiti Teknologi Malaysia (UTM) for the equipment and technical assistance. This work has been funded by the Ministry of Higher Education under FRGS, Registration Proposal No: FRGS/1/2020/ICT10/UTM/02/1.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ajune Wanis Ismail .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Ismail, A.W., Aladin, M.Y.F., Halim, N.A.A., Manaf, M.S.A. (2023). Augmented Reality Using Gesture and Speech Accelerates User Interaction. In: Shaw, R.N., Paprzycki, M., Ghosh, A. (eds) Advanced Communication and Intelligent Systems. ICACIS 2022. Communications in Computer and Information Science, vol 1749. Springer, Cham. https://doi.org/10.1007/978-3-031-25088-0_20

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-25088-0_20

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-25087-3

  • Online ISBN: 978-3-031-25088-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics