ABSTRACT
Any system using voice to communicate becomes personified by that voice. For robots, where the form and non-vocal behaviour also strongly personify the system, we can see a clash between the two technologies. The many challenges in building responsive and interactive robots mean that language systems are often designed in a vacuum and when they are finally brought together can ruin the look, feel and sound of the completed system. This problem is intensified by natural language processing technology which can further add inappropriate behaviours following mythical business use cases, rather than exploring how users really would like to relate and use embodied artificial systems. In this positional paper, we present two studies in robot voice design and a non-vocal use case of the Honda Research Institute robot Haru. Finally, we ask the question what sort of voice should Haru have?
- Matthew P Aylett, Alessandro Vinciarelli, and Mirjam Wester. 2017. Speech synthesis for the generation of artificial personality. IEEE transactions on affective computing (2017).Google Scholar
- Holly P Branigan, Martin J Pickering, Jamie Pearson, Janet F McLean, and Ash Brown. 2011. The role of beliefs in lexical alignment: Evidence from dialogs with humans and computers. Cognition 121, 1 (2011), 41--57.Google ScholarCross Ref
- Michael Braun, Anja Mainz, Ronee Chadowitz, Bastian Pfleging, and Florian Alt. 2019. At your service: Designing voice assistant personalities to improve automotive user interfaces. In CHI'19. ACM, New York, NY, USA. Google ScholarDigital Library
- Joao Paulo Cabral, Benjamin R Cowan, Katja Zibrek, and Rachel McDonnell. 2017. The Influence of Synthetic Voice on the Evaluation of a Virtual Character.. In INTERSPEECH. 229--233.Google Scholar
- Benjamin R Cowan, Holly P Branigan, Habiba Begum, Lucy McKenna, and Eva Szekely. 2017. They Know as Much as We Do: Knowledge Estimationand Partner Modelling of Artificial Partners. In CogSci.Google Scholar
- Paul Foulkes, James M Scobbie, and Dominic Watt. 2010. Sociophonetics. The handbook of phonetic sciences (2010), 703--754.Google Scholar
- Randy Gomez, Deborah Szapiro, Kerl Galindo, and Keisuke Nakamura. 2018. Haru: Hardware design of an experimental tabletop robot assistant. In Proceedings of the 2018 ACM/IEEE international conference on human-robot interaction. ACM, 233--240. Google ScholarDigital Library
- Erico Guizzo. 2014. Cynthia Breazeal Unveils Jibo, a social robot for the home. IEEE Spectrum (2014).Google Scholar
- Ewa Luger and Abigail Sellen. 2016. Like having a really bad PA: the gulf between user expectation and experience of conversational agents. In CHI'16. ACM, 5286--5297. Google ScholarDigital Library
- Roger K Moore. 2017. Appropriate Voices for Artefacts: Some Key Insights. In 1st International Workshop on Vocal Interactivity in-and-between Humans, Animals and Robots.Google Scholar
- Roger K Moore. 2017. Is spoken language all-or-nothing? Implications for future speech-based human-machine interaction. In Dialogues with Social Robots. Springer, 281--291.Google Scholar
- Roger K Moore and Ben Mitchinson. 2017. A biomimetic vocalisation system for MiRo. In Conference on Biomimetic and Biohybrid Systems. Springer, 363--374.Google ScholarCross Ref
- Clifford Nass and Kwan Min Lee. 2001. Does computer-synthesized speech manifest personality? Experimental tests of recognition, similarity-attraction, and consistency-attraction. Journ. of exp. psychology: applied 7, 3 (2001), 171.Google Scholar
- Chris Wiltz. 2018. From Cozmo to Vector: How Anki Designs Robots With Emotional Intelligence. Plastics Today (2018).Google Scholar
Index Terms
- The right kind of unnatural: designing a robot voice
Recommendations
Creating Robot Personality: Effects of Mixing Speech and Semantic Free Utterances
HRI '20: Companion of the 2020 ACM/IEEE International Conference on Human-Robot InteractionPersonality is a vital factor in understanding acceptance, trust or emotional attachment to a robot. From R2D2 to WALL-E, there is a rich history of robots in films using semantic free utterances (SFUs) sounds (such as squeaks, clicks and tones) as ...
Voice Puppetry: Towards Conversational HRI WoZ Experiments with Synthesised Voices
HRI '20: Companion of the 2020 ACM/IEEE International Conference on Human-Robot InteractionIn order to research conversational factors in robot design the use of Wizard of Oz (WoZ) experiments, where an experimenter plays the part of the robot, are common. However, for conversational systems using a synthetic voice, it is extremely difficult ...
Voice Puppetry: Speech Synthesis Adventures in Human Centred AI
IUI '20 Companion: Companion Proceedings of the 25th International Conference on Intelligent User InterfacesState-of-the-art speech synthesis owes much to modern AI machine learning, with recurrent neural networks becoming the new standard. However, how you say something is just as important as what you say. If we draw inspiration from human dramatic ...
Comments