Abstract
To support smooth conversation between people, we are developing a support system that provides appropriate conversation topics when the speaker cannot immediately think of a topic. The problem with this system is that speech recognition errors occur frequently. The system may not be able to provide appropriate conversation topics due to speech recognition errors. Various methods have been proposed to prevent speech recognition errors, but these methods are still insufficient for domain free daily conversation. Vocalizations become ambiguous across multiple words in human-to-human daily conversation. We believe that it is necessary to understand how acoustic features change due to ambiguous vocalizations, and how changes in acoustic features affect speech recognition results, and to develop countermeasures based on these findings. This paper discusses the speech utterances characteristics that are prone to speech recognition errors to identify speech recognition errors caused by ambiguity in daily speech. In particular, we focus on F0 and mora, and analyze how these factors differ between correctly and incorrectly recognized speech.
We confirmed that the F0 extraction rate tends to be lower for misrecognized speech parts than for correctly recognized speech, and we confirmed that the mora length tends to be shorter for misrecognized speech parts. Regarding the mora, there were a small number of misrecognized utterances in which the mora was extremely long, and the standard deviation of the mor of the misrecognized utterances tended to be larger than that of the misrecognized utterances. By comparing the transcribed sentences with the recognition result sentences, we analyzed the factors contributing to the above tendency, and found that the phenomenon of vowel voicelessness was promoted in daily conversation, with the absence or shortening of vocalized segments including not only vowels but also consonant parts before and after them, and the phenomenon of voiceless nasal sounds at the end of utterances. When the above voicelessness occurred at word-word boundaries, it was found that the words before and after the boundary could be combined and misrecognized as completely different words. In the future, we would like to quantify these features and propose a misrecognized utterances identification method.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Kim, T., Chang, A., Holland, L., Pentland, A.S.: Meeting mediator: enhancing group collaboration using sociometric feedback. In: Proceedings of the 2008 ACM Conference on Computer Supported Cooperative Work, pp. 457–466 (2008)
Otsuka, K., Itoh, J., Munnemori, J.: Communication support system for enabling group management of community units. Procedia Comput. Sci. 60(1), 900–907 (2015)
Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics (NAACL): Human Language Technologies, vol. 1, pp. 4171–4186 (2019)
Xie, Z., et al.: Data noising as smoothing in neural network language model. In: Conference Paper at ICLR (2017)
Asao, Y., Kloetzer, J., Mizuno, J., Saiki, D., Kadowaki, K., Torisawa, K.: Understanding user utterances in a dialog system for caregiving. In: Proceedings of the 12th Language Resources and Evaluation Conference, LREC 2020, pp. 653–661 (2020)
Ko, T., Peddinti, V., Povey, D., Khudanpur, S.: Audio augmentation for speech recognition. In: Proceedings of Interspeech 2015, pp 3586–3589 (2015)
Fujimoto, M.: Effects of consonantal environment and speech rate on vowel devoicing. In: Proceedings of the Phonetic Society, pp.1–13 (2012)
Acknowledgment
This work received additional support from JSPS KAKENHI (grant number 22K04626).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Nishida, H., Iida, Y., Wakita, Y., Nakatoh, Y. (2023). Misrecognized Utterance Identification in Support Systems for Daily Human-to-Human Conversations. In: Degen, H., Ntoa, S. (eds) Artificial Intelligence in HCI. HCII 2023. Lecture Notes in Computer Science(), vol 14051. Springer, Cham. https://doi.org/10.1007/978-3-031-35894-4_9
Download citation
DOI: https://doi.org/10.1007/978-3-031-35894-4_9
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-35893-7
Online ISBN: 978-3-031-35894-4
eBook Packages: Computer ScienceComputer Science (R0)