ABSTRACT
Models in HCI describe and provide insights into how humans use interactive technology. They are used by engineers, designers, and developers to understand and formalize the interaction process. At the same time, novel interaction paradigms arise constantly introducing new ways of how interactive technology can support humans. In this work, we look into how these paradigms can be described using the classical HCI model introduced by Schomaker in 1995. We extend this model by presenting new relations that would provide a better understanding of them. For this, we revisit the existing interaction paradigms and try to describe their interaction using this model. The goal of this work is to highlight the need to adapt the models to new interaction paradigms and spark discussion in the HCI community on this topic.
- Jonas Auda, Max Pascher, and Stefan Schneegass. 2019. Around the (Virtual) World: Infinite Walking in Virtual Reality Using Electrical Muscle Stimulation. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. ACM, 431.Google ScholarDigital Library
- Durell Bishop. 1992. Marble answering machine. Royal College of Art, Interaction Design(1992).Google Scholar
- A.W. Cairns, R.R. Bond, D.D. Finlay, C. Breen, D. Guldenring, R. Gaffney, A.G. Gallagher, A.J. Peace, and P. Henn. 2016. A computer-human interaction model to improve the diagnostic accuracy and clinical decision-making during 12-lead electrocardiogram interpretation. Journal of Biomedical Informatics 64 (2016), 93–107. https://doi.org/10.1016/j.jbi.2016.09.016 cited By 13.Google ScholarDigital Library
- Lucas Colusso, Ridley Jones, Sean A. Munson, and Gary Hsieh. 2019. A Translational Science Model for HCI. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems (Glasgow, Scotland Uk) (CHI ’19). Association for Computing Machinery, New York, NY, USA, 1–13. https://doi.org/10.1145/3290605.3300231Google ScholarDigital Library
- C. Despont-Gros, H. Mueller, and C. Lovis. 2005. Evaluating user interactions with clinical information systems: A model based on human-computer interaction models. Journal of Biomedical Informatics 38, 3 (2005), 244–255. https://doi.org/10.1016/j.jbi.2004.12.004 cited By 63.Google ScholarDigital Library
- Alan Dix, Alan John Dix, Janet Finlay, Gregory D Abowd, and Russell Beale. 2004. Human-computer interaction. Pearson Education.Google Scholar
- Ayaka Ebisu, Satoshi Hashizume, Kenta Suzuki, Akira Ishii, Mose Sakashita, and Yoichi Ochiai. 2017. Stimulated percussions: method to control human for learning music by using electrical muscle stimulation. In Proceedings of the 8th Augmented Human International Conference. 1–5.Google ScholarDigital Library
- Sarah Faltaous and Stefan Schneegass. 2020. HCI Model: A Proposed Extension to Human-Actuation Technologies. Association for Computing Machinery, New York, NY, USA, 306–308. https://doi.org/10.1145/3428361.3432081Google Scholar
- Sarah Faltaous, Chris Schönherr, Henrik Detjen, and Stefan Schneegass. 2019. Exploring Proprioceptive Take-over Requests for Highly Automated Vehicles. In Proceedings of the 18th International Conference on Mobile and Ubiquitous Multimedia (Pisa, Italy) (MUM ’19). Association for Computing Machinery, New York, NY, USA, Article 13, 6 pages. https://doi.org/10.1145/3365610.3365644Google ScholarDigital Library
- George W. Fitzmaurice, Hiroshi Ishii, and William A. S. Buxton. 1995. Bricks: Laying the Foundations for Graspable User Interfaces. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (Denver, Colorado, USA) (CHI ’95). ACM Press/Addison-Wesley Publishing Co., USA, 442–449. https://doi.org/10.1145/223904.223964Google ScholarDigital Library
- Sean Follmer, Daniel Leithinger, Alex Olwal, Akimitsu Hogge, and Hiroshi Ishii. 2013. inFORM: dynamic physical affordances and constraints through shape and object actuation.. In Uist, Vol. 13. 2501988–2502032.Google Scholar
- John Hardy, Enrico Rukzio, and Nigel Davies. 2011. Real world responses to interactive gesture based public displays. In Proceedings of the 10th International Conference on Mobile and Ubiquitous Multimedia. 33–39.Google ScholarDigital Library
- Albert Sydney Hornby. 1995. Oxford advanced learner’s dictionary of current English.Google Scholar
- Hiroshi Ishii. 2015. TRANSFORM: Beyond Tangible Bits, Towards Radical Atoms. In Proceedings of the 3rd ACM Symposium on Spatial User Interaction (Los Angeles, California, USA) (SUI ’15). Association for Computing Machinery, New York, NY, USA, 1. https://doi.org/10.1145/2788940.2788958Google ScholarDigital Library
- Hiroshi Ishii, Daniel Leithinger, Sean Follmer, Amit Zoran, Philipp Schoessler, and Jared Counts. 2015. TRANSFORM: Embodiment of ”Radical Atoms” at Milano Design Week. In Proceedings of the 33rd Annual ACM Conference Extended Abstracts on Human Factors in Computing Systems (Seoul, Republic of Korea) (CHI EA ’15). Association for Computing Machinery, New York, NY, USA, 687–694. https://doi.org/10.1145/2702613.2702969Google ScholarDigital Library
- Hiroshi Ishii and Brygg Ullmer. 1997. Tangible Bits: Towards Seamless Interfaces between People, Bits and Atoms. In Proceedings of the ACM SIGCHI Conference on Human Factors in Computing Systems (Atlanta, Georgia, USA) (CHI ’97). Association for Computing Machinery, New York, NY, USA, 234–241. https://doi.org/10.1145/258549.258715Google ScholarDigital Library
- AA Karpov and RM Yusupov. 2018. Multimodal interfaces of human–computer interaction. Herald of the Russian Academy of Sciences 88, 1 (2018), 67–74.Google ScholarCross Ref
- D. Lakatos and H. Ishii. 2012. Towards Radical Atoms — Form-giving to transformable materials. In 2012 IEEE 3rd International Conference on Cognitive Infocommunications (CogInfoCom). 37–40. https://doi.org/10.1109/CogInfoCom.2012.6422023Google ScholarCross Ref
- Mathieu Le Goc, Lawrence H. Kim, Ali Parsaei, Jean-Daniel Fekete, Pierre Dragicevic, and Sean Follmer. 2016. Zooids: Building Blocks for Swarm User Interfaces. In Proceedings of the 29th Annual Symposium on User Interface Software and Technology (Tokyo, Japan) (UIST ’16). Association for Computing Machinery, New York, NY, USA, 97–109. https://doi.org/10.1145/2984511.2984547Google ScholarDigital Library
- T. Lei, L. Xiong, K. Chen, X. Liu, Y. Cao, Q. Zhang, D. Liu, and S. Guo. 2014. Gestures: The reformer of the user’s mental model in mobile HCI. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) 8518 LNCS, PART 2(2014), 586–597. https://doi.org/10.1007/978-3-319-07626-3_55 cited By 0.Google Scholar
- Bonanni Leonardo, Ishii Hiroshi, and Tangible Media Group. [n.d.]. Radical Atoms. https://tangible.media.mit.edu/project/radical-atoms/Google Scholar
- C. J. Lim, Younghwan Pan, and Jane Lee. 2008. Human Factors and Design Issues in Multimodal (Speech/Gesture) Interface. J. Digit. Content Technol. its Appl. 2 (2008), 67–77.Google Scholar
- Hannah Limerick, David Coyle, and James W. Moore. 2014. The experience of agency in human-computer interactions: a review. Frontiers in Human Neuroscience 8 (2014), 643. https://doi.org/10.3389/fnhum.2014.00643Google ScholarCross Ref
- Pedro Lopes, Alexandra Ion, Willi Mueller, Daniel Hoffmann, Patrik Jonell, and Patrick Baudisch. 2015. Proprioceptive Interaction. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems (Seoul, Republic of Korea) (CHI ’15). Association for Computing Machinery, New York, NY, USA, 939–948. https://doi.org/10.1145/2702123.2702461Google ScholarDigital Library
- Pedro Lopes, Sijing You, Lung-Pan Cheng, Sebastian Marwecki, and Patrick Baudisch. 2017. Providing haptics to walls & heavy objects in virtual reality by means of electrical muscle stimulation. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems. ACM, 1471–1482.Google ScholarDigital Library
- Pedro Lopes, Sijing You, Alexandra Ion, and Patrick Baudisch. 2018. Adding force feedback to mixed reality experiences and games using electrical muscle stimulation. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. ACM, 446.Google ScholarDigital Library
- Pedro Lopes, Doăa Yüksel, François Guimbretière, and Patrick Baudisch. 2016. Muscle-plotter: An interactive system based on electrical muscle stimulation that produces spatial output. In Proceedings of the 29th Annual Symposium on User Interface Software and Technology. ACM, 207–217.Google ScholarDigital Library
- Brad A. Myers. 1998. A Brief History of Human-Computer Interaction Technology. Interactions 5, 2 (March 1998), 44–54. https://doi.org/10.1145/274430.274436Google ScholarDigital Library
- Ken Nakagaki, Sean Follmer, and Hiroshi Ishii. 2015. LineFORM: Actuated Curve Interfaces for Display, Interaction, and Constraint. In Proceedings of the 28th Annual ACM Symposium on User Interface Software & Technology(Charlotte, NC, USA) (UIST ’15). Association for Computing Machinery, New York, NY, USA, 333–339. https://doi.org/10.1145/2807442.2807452Google ScholarDigital Library
- Donald A Norman. 1986. Cognitive engineering. User centered system design 31 (1986), 61.Google ScholarCross Ref
- Simon Olberding, Sergio Soto Ortega, Klaus Hildebrandt, and Jürgen Steimle. 2015. Foldio: Digital Fabrication of Interactive and Shape-Changing Objects With Foldable Printed Electronics. In Proceedings of the 28th Annual ACM Symposium on User Interface Software & Technology (Charlotte, NC, USA) (UIST ’15). Association for Computing Machinery, New York, NY, USA, 223–232. https://doi.org/10.1145/2807442.2807494Google ScholarDigital Library
- Antti Oulasvirta. 2019. It’s time to rediscover HCI models. Interactions 26, 4 (2019), 52–56.Google ScholarDigital Library
- Stefano Papetti. 2013. Design and Perceptual Investigations of Audio-Tactile Interactions. Proceedings of AIA DAGA(2013).Google Scholar
- Max Pfeiffer, Tim D’unte, Stefan Schneegass, Florian Alt, and Michael Rohs. 2015. Cruise control for pedestrians: Controlling walking direction using electrical muscle stimulation. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems. 2505–2514.Google ScholarDigital Library
- Lambert Schomaker. 1995. A taxonomy of Multimodal Interaction in the Human Information Processing System.Google Scholar
- Eugénie Shinkle. 2008. Video games, emotion and the six senses. Media, culture & society 30, 6 (2008), 907–915.Google Scholar
- Loretta Staples. 1993. Representation in Virtual Space: Visual Convention in the Graphical User Interface. In Proceedings of the INTERACT ’93 and CHI ’93 Conference on Human Factors in Computing Systems (Amsterdam, The Netherlands) (CHI ’93). Association for Computing Machinery, New York, NY, USA, 348–354. https://doi.org/10.1145/169059.169279Google ScholarDigital Library
- Nicolas Stifani. 2014. Motor neurons and the generation of spinal motor neurons diversity. Frontiers in cellular neuroscience 8 (2014), 293.Google Scholar
- Ivan E Sutherland. 1964. Sketchpad a man-machine graphical communication system. Simulation 2, 5 (1964), R–3.Google Scholar
- Faisal Taher, John Hardy, Abhijit Karnik, Christian Weichel, Yvonne Jansen, Kasper Hornbæk, and Jason Alexander. 2015. Exploring Interactions with Physically Dynamic Bar Charts. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems (Seoul, Republic of Korea) (CHI ’15). Association for Computing Machinery, New York, NY, USA, 3237–3246. https://doi.org/10.1145/2702123.2702604Google ScholarDigital Library
- Emi Tamaki, Takashi Miyaki, and Jun Rekimoto. 2011. PossessedHand: techniques for controlling human hands using electrical muscles stimuli. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. 543–552.Google ScholarDigital Library
- Wondimu W Teka, Khaldoun C Hamade, William H Barnett, Taegyo Kim, Sergey N Markin, Ilya A Rybak, and Yaroslav I Molkov. 2017. From the motor cortex to the movement and back again. PloS one 12, 6 (2017), e0179288.Google ScholarCross Ref
- A.R. Várkonyi-Kóczy and B. Tusor. 2011. Human-Computer interaction for smart environment applications using fuzzy hand posture and gesture models. IEEE Transactions on Instrumentation and Measurement 60, 5(2011), 1505–1514. https://doi.org/10.1109/TIM.2011.2108075 cited By 58.Google ScholarCross Ref
- Robert Walter, Gilles Bailly, Nina Valkanova, and Jörg Müller. 2014. Cuenesics: using mid-air gestures to select items on interactive public displays. In Proceedings of the 16th international conference on Human-computer interaction with mobile devices & services. 299–308.Google ScholarDigital Library
- John Williamson, Roderick Murray-Smith, Benjamin Blankertz, Matthias Krauledat, and K-R Müller. 2009. Designing for uncertain, asymmetric control: Interaction design for brain–computer interfaces. International Journal of Human-Computer Studies 67, 10 (2009), 827–841.Google ScholarDigital Library
- M. Wrzesien, J.-M. Burkhardt, C. Botella, and M. Alcañiz. 2015. Towards a Virtual Reality- and Augmented Reality-Mediated Therapeutic Process model: a theoretical revision of clinical issues and HCI issues. Theoretical Issues in Ergonomics Science 16, 2 (2015), 124–153. https://doi.org/10.1080/1463922X.2014.903307 cited By 9.Google ScholarCross Ref
- Towards a Universal Human-Computer Interaction Model for Multimodal Interactions
Recommendations
Enabling tangible interaction on capacitive touch panels
UIST '10: Adjunct proceedings of the 23nd annual ACM symposium on User interface software and technologyWe propose two approaches to sense tangible objects on capacitive touch screens, which are used in off-the-shelf multi-touch devices such as Apple iPad, iPhone, and 3M's multi-touch displays. We seek for the approaches that do not require modifications ...
Role-Based Human-Computer Interactions
With increased understanding of cognitive informatics and the advance of computer technologies, it is becoming clear that human-computer interaction HCI is an interaction between two kinds of intelligences, i.e., natural intelligence and artificial ...
Investigating Interactions for Text Recognition using a Vibrotactile Wearable Display
IUI '18: Proceedings of the 23rd International Conference on Intelligent User InterfacesVibrotactile skin-reading uses wearable vibrotactile displays to convey dynamically generated textual information. Such wearable displays have potential to be used in a broad range of applications. Nevertheless, the reading process is passive, and users ...
Comments