skip to main content
10.1145/3236112.3236171acmconferencesArticle/Chapter ViewAbstractPublication PagesmobilehciConference Proceedingsconference-collections
extended-abstract

Designing speech interaction for the Sony Xperia Ear and Oakley Radar Pace smartglasses

Published: 03 September 2018 Publication History

Abstract

Speech synthesis is a key enabling technology for wearable technology. We discuss the design challenges in customising speech synthesis for the Sony Xperia Ear, and the Oakley Radar Pace smartglasses. In order to support speech interaction designers working on novel interactive eye-free mobile devices, specific functionality is required including: flexibility in terms of performance, memory footprint, disk requirements, server or local configurations, methods for personification and branding, architectures for fast reactive interfaces, and customisation for content, genres and speech styles. We describe implementations of this required functionality and how this functionality can be made available to engineers and designers working on 3rd party devices and the impact they can have on user experience. To conclude we discuss why some customers are reluctant to depend on speech services from well known providers such as Google and Amazon and consider the barrier to entry for custom built personal digital advisors.

References

[1]
M.P. Aylett and C.J. Pidcock. 2008. Adding and Controlling Emotion in Synthesised Speech. (10 September 2008). UK Patent GB2447263A.
[2]
Matthew P Aylett, Per Ola Kristensson, Steve Whittaker, and Yolanda Vazquez-Alvarez. 2014. None of a CHInd: relationship counselling for HCI and speech technology. In CHI'14. ACM, 749--760.
[3]
Matthew P Aylett and Shaun Lawson. 2016. The Smartphone: A Lacanian Stain, A Tech Killer, and an Embodiment of Radical Individualism. In CHI'16. ACM, 501--511.
[4]
Matthew P. Aylett and Christopher J. Pidcock. 2007. The Cere Voice Characterful Speech Synthesiser SDK. In AISB. 174--8.
[5]
Matthew P Aylett, Graham Pullin, David A Braude, Blaise Potard, Shannon Hennig, and Marilia Antunes Ferreira. 2016. Don't Say Yes, Say Yes: Interacting with Synthetic Speech Using Tonetable. In CHI'16. ACM, 3643--3646.
[6]
Matthew P Aylett, Alessandro Vinciarelli, and Mirjam Wester. 2017. Speech Synthesis for the Generation of Artificial Personality. IEEE Transactions on Affective Computing (2017).
[7]
Cosmin Munteanu, Matt Jones, Sharon Oviatt, Stephen Brewster, Gerald Penn, Steve Whittaker, Nitendra Rajput, and Amit Nanavati. 2013. We need to talk: HCI and the delicate topic of spoken language interaction. In CHI'13. ACM, 2459--2464.
[8]
C. Nass and S. Brave. 2005. Wired for speech: How voice activates and advances the Human-Computer relationship. The MIT Press.
[9]
Ben Shneiderman. 2000. The limits of speech recognition. Commun. ACM 43, 9 (2000), 63--65.
[10]
Thad E Starner. 2002. The role of speech input in wearable computing. Pervasive Computing, IEEE 1, 3 (2002), 89--93.

Cited By

View all
  • (2021)Building and Designing Expressive Speech SynthesisThe Handbook on Socially Interactive Agents10.1145/3477322.3477329(173-212)Online publication date: 10-Sep-2021

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
MobileHCI '18: Proceedings of the 20th International Conference on Human-Computer Interaction with Mobile Devices and Services Adjunct
September 2018
445 pages
ISBN:9781450359412
DOI:10.1145/3236112
Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 03 September 2018

Check for updates

Author Tags

  1. audio interaction
  2. mobile
  3. speech synthesis

Qualifiers

  • Extended-abstract

Conference

MobileHCI '18
Sponsor:

Acceptance Rates

Overall Acceptance Rate 202 of 906 submissions, 22%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)6
  • Downloads (Last 6 weeks)0
Reflects downloads up to 25 Feb 2025

Other Metrics

Citations

Cited By

View all
  • (2021)Building and Designing Expressive Speech SynthesisThe Handbook on Socially Interactive Agents10.1145/3477322.3477329(173-212)Online publication date: 10-Sep-2021

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media