ABSTRACT
In this paper we discuss the potential of ordinary objects acting as human computer interfaces with an Inertial Measurement Unit, the Twiz, to capture a body's orientation and acceleration. The motivation behind this research is to develop a toolkit that enables end users to quickly prototype custom interfaces for artistic expressions through movement. Through an iterative design process we have enhanced existing technical implementations such as wireless data transfer, battery lifespan, two-way communication and data analysis including machine-learning techniques. We conducted object-making sessions and developed software prototypes for audio and visual feedback. We explored a range of experiments related to visual arts, dance, and music by attaching the Twiz to different types of objects to allow users to carry out impromptu interactions. As a result of this process we have gained a better understand of an object's expressive potential whilst capturing and analyzing its movement.
- Marta Braun. 1994. Picturing time: the work of Etienne-Jules Marey (1830-1904). University of Chicago Press.Google Scholar
- John Cage. 1965. Water walk. Edition Peters Group, Frankfurt/Main, Leipzig, London, New York.Google Scholar
- Marcel Duchamp. 2006. Nude descending a staircase, no. 2. 111847 (2006).Google Scholar
- Rebecca Anne Fiebrink. 2011. Real-time Human Interaction with Supervised Learning Algorithms for Music Composition and Performance. Ph.D. Dissertation. Princeton University, NJ, United States. Advisor(s) Professor Perry R. Cook. http://dl.acm.org/citation.cfm?id=2125776 AAI3445567.Google Scholar
- Rachel Freire, Cedric Honnet, and Paul Strohmeier. 2017. Second Skin: An Exploration of eTextile Stretch Circuits on the Body. In Proceedings of the Eleventh International Conference on Tangible, Embedded, and Embodied Interaction (TEI '17). ACM, NY, USA, 653--658. Google ScholarDigital Library
- Nicholas Gillian and Joseph A. Paradiso. 2014. The Gesture Recognition Toolkit. J. Mach. Learn. Res. (2014). http://dl.acm.org/citation.cfm?id=2627435.2697076 Google ScholarDigital Library
- Frederic Gmeiner. 2007. Table Recorder, instrument for the everyday - Retrieved on February 8th, 2017. (2007). http://www.fregment.com/?cat=0&p=7868Google Scholar
- D. W. Gotshalk. 1954. Aesthetic Expression. The Journal of Aesthetics and Art Criticism 13,1 (1954), 80--85. http://www.jstor.org/stable/427019Google ScholarCross Ref
- Jonathan Hook, Thomas Nappey, Steve Hodges, Peter Wright, and Patrick Olivier. 2014. Making 3D Printed Objects Interactive Using Wireless Accelerometers. In CHI '14 Extended Abstracts on Human Factors in Computing Systems (CHI EA '14). ACM, New York, NY, USA, 1435--1440. Google ScholarDigital Library
- Hanchuan Li, Can Ye, and Alanson P. Sample. 2015. IDSense: A Human Object Interaction Detection System Based on Passive UHF RFID. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems (CHI '15). ACM, NY, USA, 2555--2564. Google ScholarDigital Library
- Sebastian OH Madgwick, Andrew JL Harrison, and Ravi Vaidyanathan. 2011. Estimation of IMU and MARG orientation using a gradient descent algorithm. In Rehabilitation Robotics (ICORR), 2011 IEEE International Conference on. IEEE.Google ScholarCross Ref
- Andrew McPherson and Victor Zappi. 2015. An Environment for Submillisecond-Latency Audio and Sensor Processing on BeagleBone Black. In Audio Engineering Society Convention 138. http://www.aes.org/e-lib/browse.cfm?elib=17755Google Scholar
- David Merrill, Jeevan Kalanithi, and Pattie Maes. 2007. Siftables: Towards Sensor Network User Interfaces. In Proceedings of the 1st International Conference on Tangible and Embedded Interaction (TEI '07). ACM, NY, USA, 75--78. Google ScholarDigital Library
- Thomas Mitchell, Peter Bennett, Sebastian Madgwick, Edward Davies, and Philip Tew. 2016. Tangible Interfaces for Interactive Evolutionary Computation. In Proceedings of the 2016 CHI Conference Extended Abstracts on Human Factors in Computing Systems (CHI EA '16). ACM, New York, NY, USA, 2609--2616. Google ScholarDigital Library
- Dan O'Sullivan and Tom Igoe. 2004. Physical Computing: Sensing and Controlling the Physical World with Computers. Course Technology Press, Boston, MA, USA. Google ScholarDigital Library
- Laura Paolini. 2007. John Cage's Secret - Retrieved on February 8th, 2017. (2007). http://www.johncage.org/blog/paolini-cage-eds-editlp.pdfGoogle Scholar
- Katrin Plaumann, Milos Babic, Tobias Drey, Witali Hepting, Daniel Stooß, and Enrico Rukzio. 2016. Towards Improving Touchscreen Input Speed and Accuracy on Smartphones for Tremor Affected Persons. In Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing: Adjunct (UbiComp '16). ACM, NY, USA. Google ScholarDigital Library
- Ivan Poupyrev, Philipp Schoessler, Jonas Loh, and Munehiko Sato. 2012. Botanicus Interacticus: Interactive Plants Technology. In ACM SIGGRAPH 2012 Emerging Technologies (SIGGRAPH '12). ACM, New York, NY, USA, Article 4,1 pages. Google ScholarDigital Library
- Andreas Schlegel and Cedric Honnet. 2016. Movement of Things Exploring Inertial Motion Sensing When Autonomous, Tiny and Wireless. In Proceedings of the 3rd International Symposium on Movement and Computing (MOCO '16). ACM, NY, USA, 47:1--47:2. Google ScholarDigital Library
- Christa Sommerer and Laurent Mignonneau. 1993. Interactive plant growing. Ars Electronica 93 (1993), 408--414.Google Scholar
- Jacob Davis Babcock Stillman, Eadweard Muybridge, and others. 1882. horse in motion as shown by instantaneous photography. (1882).Google Scholar
- Yuri Suzuki. 2015. Acoustic Pavilion - Retrieved on February 8th, 2017. (2015). http://yurisuzuki.com/news/acoustic-pavilionGoogle Scholar
- Yuri Suzuki. 2016. AR Music Kit - Retrieved on February 8th, 2017. (2016). http://yurisuzuki.com/news/ar-music-kitGoogle Scholar
- Matthew Wright. 2002. Open Sound Control Specification Version 1.0 - Retrieved on February 8th, 2017. (2002). http://opensoundcontrol.org/spec-1_0Google Scholar
Index Terms
Digital Oxymorons: From Ordinary to Expressive Objects Using Tiny Wireless IMUs
Recommendations
Movement of Things Exploring Inertial Motion Sensing When Autonomous, Tiny and Wireless
MOCO '16: Proceedings of the 3rd International Symposium on Movement and ComputingThe Movement of Things project is an exploration into the qualities and properties of movement. Through a range of exercises these movements are captured and translated by custom-built software and the use of an autonomous, tiny and wireless motion ...
The augmentalist: enabling musicians to develop augmented musical instruments
TEI '11: Proceedings of the fifth international conference on Tangible, embedded, and embodied interactionWe present the Augmentalist: a system that allows a musician to augment their existing musical instrument using a variety of sensors. This system is focused on the musician and their musical aims, rather than the sensor technology being used. The ...
Masters' skill explained by visualization of whole-body muscle activity
SIMPAR'12: Proceedings of the Third international conference on Simulation, Modeling, and Programming for Autonomous RobotsIn this paper, we discuss the computation of human motion dynamics and its analysis of experts' motion skills. The computation framework of the wire-driven multi-body dynamics previously developed by the authors is applied to the whole body human ...
Comments