ABSTRACT
This paper presents the Smart Mandolin, an exemplar of the family of the so-called smart instruments. Developed according to the paradigms of autobiographical design, it consists of a conventional acoustic mandolin enhanced with different types of sensors, a microphone, a loudspeaker, wireless connectivity to both local networks and the Internet, and a low-latency audio processing board. Various implemented use cases are presented, which leverage the smart qualities of the instrument. These include the programming of the instrument via applications for smartphones and desktop computer, as well as the wireless control of devices enabling multimodal performances such as screen projecting visuals, smartphones, and tactile devices used by the audience. The paper concludes with an evaluation conducted by the author himself after extensive use, which pinpointed pros and cons of the instrument and provided a comparison with the Hyper-Mandolin, an instance of augmented instruments previously developed by the author.
- L. Atzori, A. Iera, and G. Morabito. 2010. The Internet of Things: a survey. Computer networks 54, 15 (2010), 2787--2805. Google ScholarDigital Library
- P. Brossier. 2006. Automatic annotation of musical audio for interactive systems. Ph.D. Dissertation. Queen Mary University of London.Google Scholar
- Florian Eyben, Martin Wöllmer, and Björn Schuller. 2010. Opensmile: the munich versatile and fast open-source audio feature extractor. In Proceedings of the 18th ACM international conference on Multimedia. ACM, 1459--1462. Google ScholarDigital Library
- E. Farella. 2010. Ambient Intelligence, Smart Objects and Sensor Networks: Practical Experiences. In IEEE/IFIP International Conference on Embedded and Ubiquitous Computing. 416--421. Google ScholarDigital Library
- E. Gunther and S. O'Modhrain. 2003. Cutaneous grooves: composing for the sense of touch. Journal of New Music Research 32, 4 (2003), 369--381.Google ScholarCross Ref
- A.R. Jensenius and M.J. Lyons. 2017. A NIME Reader: Fifteen Years of New Interfaces for Musical Expression. Springer.Google Scholar
- D. Keller and V. Lazzarini. 2017. Ecologically Grounded Creative Practices in Ubiquitous Music. Organised Sound 22, 1 (2017), 61--72.Google ScholarCross Ref
- A. Mazzoni and N. Bryan-Kinns. 2016. Mood Glove: A haptic wearable prototype system to enhance mood music in film. Entertainment Computing 17 (2016), 9--17.Google ScholarCross Ref
- A. McPherson and V. Zappi. 2015. An Environment for Submillisecond-Latency Audio and Sensor Processing on BeagleBone Black. In Audio Engineering Society Convention 138. Audio Engineering Society. http://www.aes.org/elib/browse.cfm?elib=17755Google Scholar
- A. P McPherson, Robert H Jack, and Giulio Moro. 2016. Action-Sound Latency: Are Our Tools Fast Enough?. In Proceedings of the Conference on New Interfaces for Musical Expression.Google Scholar
- E.R. Miranda and M.M. Wanderley. 2006. New digital musical instruments: control and interaction beyond the keyboard. Vol. 21. AR Editions, Inc. Google ScholarDigital Library
- T. Mitchell, S. Madgwick, S. Rankine, G.S. Hilton, A. Freed, and A.R Nix. 2014. Making the Most of Wi-Fi: Optimisations for Robust Wireless Live Music Performance. In Proceedings of the Conference on New Interfaces for Musical Expression. 251--256.Google Scholar
- F. Morreale and A. McPherson. 2017. Design for Longevity: Ongoing Use of Instruments from NIME 2010-14. In Proceedings of the Conference on New Interfaces for Musical Expression. NIME.Google Scholar
- Fabio Morreale, Andrew P McPherson, and Marcelo M Wanderley. 2018. NIME Identity from the Performer's Perspective. In Proceedings of the Conference on New Interfaces for Musical Expression.Google Scholar
- Carman Neustaedter and Phoebe Sengers. 2012. Autobiographical design in HCI research: designing and learning through use-it-yourself. In Proceedings of the Designing Interactive Systems Conference. ACM, 514--523. Google ScholarDigital Library
- S. O'modhrain. 2011. A framework for the evaluation of digital musical instruments. Computer Music Journal 35, 1 (2011), 28--42. Google ScholarDigital Library
- M.S. Puckette, T. Apel, and D.D Ziccarelli. 1998. Real-time audio analysis tools for Pd and MSP. In Proceedings of the International Computer Music Conference.Google Scholar
- C.Rottondi, C. Chafe, C. Allocchio, and A. Sarti. 2016. An Overview on Networked Music Performance Technologies. IEEE Access 4 (2016), 8823--8843.Google ScholarCross Ref
- M. Slaney. 2002. Semantic-audio retrieval. In IEEE International Conference on Acoustics, Speech, and Signal Processing, Vol. 4. 4108--4111.Google Scholar
- P. Sparks. 1995. The classical mandolin. Oxford: Clarendon Press.Google Scholar
- B. Sturm and O. Ben-Tal. 2017. Bringing the models back to music practice: The evaluation of deep learning approaches to music transcription modelling and generation. Journal of Creative Music Systems (2017).Google Scholar
- L. Turchet. 2017. The Hyper-Mandolin. In Proceedings of Audio Mostly Conference. 1:1--1:8. Google ScholarDigital Library
- L. Turchet and M. Barthet. 2017. Envisioning Smart Musical Haptic Wearables to Enhance Performers' Creative Communication. In Proceedings of International Symposium on Computer Music Multidisciplinary Research. 538--549.Google Scholar
- L. Turchet and M. Barthet. 2018. Demo of interactions between a performer playing a Smart Mandolin and audience members using Musical Haptic Wearables. In Proceedings of the Conference on New Interfaces for Musical Expression. 82--83.Google Scholar
- L. Turchet, M. Benincaso, and C. Fischione. 2017. Examples of use cases with Smart Instruments. In Proceedings of Audio Mostly Conference. 47:1--47:5. Google ScholarDigital Library
- L. Turchet, C. Fischione, and M. Barthet. 2017. Towards the Internet of Musical Things. In Proceedings of the Sound and Music Computing Conference. 13--20.Google Scholar
- L. Turchet, A. McPherson, and M. Barthet. 2018. Co-design of a Smart Cajón. Journal of the Audio Engineering Society 66, 4 (2018), 220--230.Google ScholarCross Ref
- L. Turchet, A. McPherson, and M. Barthet. 2018. Real-time hit classification in a Smart Cajón. Frontiers in ICT 5, 16 (2018).Google Scholar
- L. Turchet, A. McPherson, and C. Fischione. 2016. Smart Instruments: Towards an Ecosystem of Interoperable Devices Connecting Performers and Audiences. In Proceedings of the Sound and Music Computing Conference. 498--505.Google Scholar
- J. Tyler and P. Sparks. 2012. The Mandolin: Its Structure and Performance (Sixteenth to Twentieth Centuries). Performance Practice Review 9, 2 (2012), 166--177.Google ScholarCross Ref
Index Terms
- Smart Mandolin: autobiographical design, implementation, use cases, and lessons learned
Recommendations
Examples of use cases with Smart Instruments
AM '17: Proceedings of the 12th International Audio Mostly Conference on Augmented and Participatory Sound and Music ExperiencesThis paper presents some of the possibilities for interaction between performers, audiences, and their smart devices, offered by the novel family of musical instruments, the Smart Instruments. For this purpose, some implemented use cases are described, ...
Jamming with a Smart Mandolin and Freesound-based Accompaniment
FRUCT'23: Proceedings of the 23rd Conference of Open Innovations Association FRUCTThis paper presents an Internet of Musical Things ecosystem involving musicians and audiences interacting with a smart mandolin, smartphones, and the Audio Commons online repository Freesound. The ecosystem has been devised to sup- port performer-...
Some reflections on the relation between augmented and smart musical instruments
AM '18: Proceedings of the Audio Mostly 2018 on Sound in Immersion and EmotionAugmented musical instruments (AMIs) consist of the augmentation of conventional instruments by means of sensor or actuator technologies. Smart musical instruments (SMIs) are instruments embedding not only sensor and actuator technology, but also ...
Comments