ABSTRACT
An in-vehicle speech dialog system (SDS) can support visualhaptic interfaces and reduce eyes-off-road time while driving. This work evaluates two SDS that varied according to the degree of natural language understanding afforded by the speech dialog system. In a Wizard of Oz simulation, two alternative SDSs were tested in a driving simulator. The Lane Change Test was used to compare a command and control system with a system supporting natural language input. This driving simulator study was conducted using the same setup in Germany, USA, and China. 40 participants per country were instructed to perform interaction tasks from contexts like media, telephone, and navigation. The results show that natural language SDS could lead to a faster and more intuitive way of interacting with in-vehicle SDS. US and Chinese users especially preferred the natural language enabled system over the command and control system.
- Branigan, H. P., Pickering, M. J. Pearson, J. and McLean, J.F. 2010. Linguistic alignment between people and computers. Journal of Pragmatics, 42, 2355--2368.Google ScholarCross Ref
- Brooke, J. 1996. SUS: A Quick and Dirty Usability Scale. In Usability Evaluation in Industry, P. W. Jordan, B. Thomas, B. A. Weerdmeester & I. L. McClelland, Eds. London: Taylor & Francis, 189--194.Google Scholar
- Brumby, D. P., Davies, S. C., Janssen, C. P., and Grace, J. J. 2011. Fast or safe? How Performance Objectives Determine Modality Output Choices while Interacting on the Move. Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. CHI'11. ACM, New York, NY, 473--482. Google ScholarDigital Library
- Cameron, H. 2000. Speech at the interface. Proceedings of the COST249 Workshop on Speech in Telephone Networks (Ghent, Belgium, May 2000), 1--7.Google Scholar
- Hackenberg, L. 2013. Maschinen als kollaborative Gesprächspartner. Doctoral Thesis. Technical University Braunschweig.Google Scholar
- Hamerich, S. W. 2009. Sprachbedienung im Automobil: Teilautomatisierte Entwicklung benutzerfreundlicher Dialogsysteme. Springer, Berlin Heidelberg.Google Scholar
- ISO, ISO 26022:2010 - Road vehicles - Ergonomic aspects of transport information and control systems - Simulated lane change test to assess in-vehicle secondary task demand. 2010.Google Scholar
- Leiber, P. 2010. Ergonomische Produktgestaltung am Beispiel mobiler Geräte im interkulturellen Vergleich: China-Deutschland-USA. Doctoral Thesis, Technical University Chemnitz.Google Scholar
- Maciej, J., and Vollrath, M. 2009. Comparison of manual vs. speech-based interaction with in-vehicle information systems. Accident Analysis and Prevention, 41, 924--930.Google ScholarCross Ref
- Pickering, M. J., and Garrod, S. 2004. Toward a mechanistic psychology of dialogue. Behavioral and Brain Sciences, 27, 169--190.Google ScholarCross Ref
- Salvucci, D. D. 2001. Predicting the effects of in-car interface use on driver performance: An integrated model approach. International Journal of Human-Computer Studies, 55, 85--107.Google ScholarDigital Library
- Tsimhoni, O., Smith, D., and Green, P. 2004. Address entry while driving: Speech recognition versus a touch-screen keyboard. Human Factors, 46, 600--610.Google ScholarCross Ref
- Wechsung, I., Schaffer, S., Schleicher, R., Naumann, A., and Möller, S. 2010. Proceedings of the 11th Annual Conference of the ISCA (Makuhari, Chiba, Japan, 2010). Interspeech '10. ISCA, 1930--1933.Google Scholar
- Wickens, C. D. 2002. Multiple resources and performance prediction. Theoretical Issues in Ergonomics Science, 3, 2, 159--177.Google ScholarCross Ref
Index Terms
- International evaluation of NLU benefits in the domain of in-vehicle speech dialog systems
Recommendations
Speech dialogue with facial displays: multimodal human-computer conversation
ACL '94: Proceedings of the 32nd annual meeting on Association for Computational LinguisticsHuman face-to-face conversation is an ideal model for human-computer dialogue. One of the major features of face-to-face communication is its multiplicity of communication channels that act on multiple modalities. To realize a natural multimodal ...
Facilitating multiparty dialog with gaze, gesture, and speech
ICMI-MLMI '10: International Conference on Multimodal Interfaces and the Workshop on Machine Learning for Multimodal InteractionWe study how synchronized gaze, gesture and speech rendered by an embodied conversational agent can influence the flow of conversations in multiparty settings. We begin by reviewing a computational framework for turn-taking that provides the foundation ...
Towards Conversationally Intelligent Dialog Systems
CHI EA '22: Extended Abstracts of the 2022 CHI Conference on Human Factors in Computing SystemsSpoken dialog systems, lacking the means to address the complex phenomena of spontaneous speech and conversational dynamics, force users into a constrained mode of dialog that resembles text-based interaction more closely than spoken conversation. Turn-...
Comments