ABSTRACT
With the explosion of novel technologies, the design of multimodal interfaces has been a topic of importance in HCI. This study aims at understanding users’ response to bi-modal feedback when performing tasks in Augmented Reality applications. Design experiments are conducted in two feedback phases namely Validation and Alerts. During the experiments visual, aural and combined feedbacks are provided and the users’ performances under each uni-, bi-modal feedback conditions are evaluated to establish new recommendations and modality relations.
- Motoyuki Akamatsu, I. Scott Mackenzie, and Thierry Hasbroucq. 1995. A comparison of tactile, auditory, and visual feedback in a pointing task using a mouse-type device. Ergonomics 38, 4: 816–827. http://doi.org/10.1080/00140139508925152Google ScholarCross Ref
- Ronald T. Azuma. 1997. A Survey of Augmented Reality. Presence: Teleoperators and Virtual Environments 6, 4: 355–385. http://doi.org/10.1162/pres.1997.6.4.355Google ScholarDigital Library
- Iris Bakker, Theo Van Der Voordt, Peter Vink, and Jan De Boon. 2014. Pleasure, Arousal, Dominance: Mehrabian and Russell revisited. Current Psychology 33, 3: 405–421. http://doi.org/10.1007/s12144-014-9219-4Google ScholarCross Ref
- Michael D. Basil. 1994. Multiple Resource Theory II. Communication Research 21, 2: 208–231. http://doi.org/10.1177/009365094021002004Google ScholarCross Ref
- Niels Ole Bernsen. 1997. Defining a taxonomy of output modalities from an HCI perspective. Computer Standards & Interfaces 18, 6-7: 537–553. http://doi.org/10.1016/s0920-5489(97)00018-4Google ScholarCross Ref
- Richard A. Bolt. 1980. “Put-that-there.” ACM SIGGRAPH Computer Graphics 14, 3: 262–270. http://doi.org/10.1145/965105.807503Google ScholarDigital Library
- Stephen A. Brewster, Peter C. Wright, and Alistair D. N. Edwards. 1994. The design and evaluation of an auditory-enhanced scrollbar. Conference companion on Human factors in computing systems - CHI '94. http://doi.org/10.1145/259963.260340Google ScholarDigital Library
- Alexander M. Brown, Zoltan R. Kenwell, Brian K.v. Maraj, and David F. Collins. 2008. "Go" Signal Intensity Influences the Sprint Start. Medicine & Science in Sports & Exercise 40, 6: 1142–1148. http://doi.org/10.1249/mss.0b013e31816770e1Google ScholarCross Ref
- Yu-Ning Chang, Raymond Koon Chuan Koh, and Henry Been-Lirn Duh. 2011. Handheld AR games — A triarchic conceptual design framework.2011 IEEE International Symposium on Mixed and Augmented Reality - Arts, Media, and Humanities. http://doi.org/10.1109/ismar-amh.2011.6093653 .Google ScholarDigital Library
- Andy Cockburn and Stephen Brewster. 2005. Multimodal feedback for the acquisition of small targets. Ergonomics 48, 9: 1129–1150. http://doi.org/10.1080/00140130500197260Google ScholarCross Ref
- Philip R. Cohen, Michael Johnston, David Mcgee, 1997. QuickSet. Proceedings of the fifth conference on Applied natural language processing -. http://doi.org/10.3115/974557.974562Google ScholarDigital Library
- Kelly S. Hale and Kay M. Stanney. 2014. Handbook of virtual environments: design, implementation, and applications. CRC Press, Boca Raton.Google ScholarDigital Library
- Granström Björn, David House, and Inger Karlsson. 2014. Multimodality in language and speech systems. Springer, Dordrecht.Google Scholar
- Wolfgang Hürst and Kevin Vriens. 2016. Multimodal feedback for finger-based interaction in mobile augmented reality. Proceedings of the 18th ACM International Conference on Multimodal Interaction - ICMI 2016. http://doi.org/10.1145/2993148.2993163Google ScholarDigital Library
- Sylvia Irawati, Scott Green, Mark Billinghurst, Andreas Duenser, and Heedong Ko. 2006. An Evaluation of an Augmented Reality Multimodal Interface Using Speech and Paddle Gestures. Advances in Artificial Reality and Tele-Existence Lecture Notes in Computer Science: 272–283. http://doi.org/10.1007/11941354_28Google ScholarDigital Library
- Ajune Wanis Ismail, Mark Billinghurst, and Mohd Shahrizal Sunar. 2015. Vision-Based Technique and Issues for Multimodal Interaction in Augmented Reality. Proceedings of the 8th International Symposium on Visual Information Communication and Interaction - VINCI '15.http://doi.org/10.1145/2801040.2801058Google ScholarDigital Library
- Minkyung Lee, Mark Billinghurst, Woonhyuk Baek, Richard Green, and Woontack Woo. 2013. A usability study of multimodal input in an augmented reality environment. Virtual Reality 17, 4: 293–305. http://doi.org/10.1007/s10055-013-0230-0Google ScholarDigital Library
- Sharon Oviatt. 1999. Ten myths of multimodal interaction. Communications of the ACM 42, 11: 74–81. http://doi.org/10.1145/319382.319398Google ScholarDigital Library
- Thammathip Piumsomboon, David Altimira, Hyungon Kim, Adrian Clark, Gun Lee, and Mark Billinghurst. 2014. Grasp-Shell vs gesture-speech: A comparison of direct and indirect natural interaction techniques in augmented reality.2014 IEEE International Symposium on Mixed and Augmented Reality (ISMAR). http://doi.org/10.1109/ismar.2014.6948411Google Scholar
- Michael I. Posner. 1980. Orienting of attention. Quarterly Journal of Experimental Psychology 32, 1: 3–25. http://doi.org/10.1080/00335558008248231Google ScholarCross Ref
- Ivan Poupyrev, Suzanne Weghorst, Mark Billinghurst, and Tadao Ichikawa. 1997. A framework and testbed for studying manipulation techniques for immersive VR. Proceedings of the ACM symposium on Virtual reality software and technology - VRST '97. http://doi.org/10.1145/261135.261141Google ScholarDigital Library
- Karen Renaud, and Richard Cooper. 2000. Feedback in human-computer interaction-characteristics and recommendations. South African Computer Journal 2000, 26: 105–114.Google Scholar
- Barry E. Stein, Terrence R. Stanford, Ramnarayan Ramachandran, Thomas J. Perrault, and Benjamin A. Rowland. 2009. Challenges in quantifying multisensory integration: alternative criteria, models, and inverse effectiveness. Experimental Brain Research 198, 2-3. http://doi.org/10.1007/s00221-009-1880-8Google ScholarCross Ref
- Christopher D Wickens. 1981. Processing Resources in Attention, Dual Task Performance, and Workload Assessment. Defense Technical Information Center.Google Scholar
Recommendations
Analysis of Peripheral Vision and Vibrotactile Feedback During Proximal Search Tasks with Dynamic Virtual Entities in Augmented Reality
SUI '19: Symposium on Spatial User InteractionA primary goal of augmented reality (AR) is to seamlessly embed virtual content into a real environment. There are many factors that can affect the perceived physicality and co-presence of virtual entities, including the hardware capabilities, the ...
Multimodal augmented reality: the norm rather than the exception
MVAR '16: Proceedings of the 2016 workshop on Multimodal Virtual and Augmented RealityAugmented reality (AR) is commonly seen as a technology that overlays virtual imagery onto a participant's view of the world. In line with this, most AR research is focused on what we see. In this paper, we challenge this focus on vision and make a case ...
Multimodal feedback for finger-based interaction in mobile augmented reality
ICMI '16: Proceedings of the 18th ACM International Conference on Multimodal InteractionMobile or handheld augmented reality uses a smartphone's live video stream and enriches it with superimposed graphics. In such scenarios, tracking one's fingers in front of the camera and interpreting these traces as gestures offers interesting ...
Comments