skip to main content
10.1145/3514197.3549619acmconferencesArticle/Chapter ViewAbstractPublication PagesivaConference Proceedingsconference-collections
extended-abstract

Determining most suitable listener backchannel type for speaker's utterance

Published:06 September 2022Publication History

ABSTRACT

A major hurdle in achieving a dialogue system that enables smooth dialogue is to determine how to generate an appropriate response to a user's utterance. Previous research has focused mainly on estimating whether to make an utterance backchannel in response to the user's utterance. We go one step further by examining, for the first time, the relationship between the type of utterance backchannel to be used and intent and type of the speaker's utterance, known as a dialogue act (DA). Specifically, we propose a new method for classifying utterance backchannels into nine types. We also created a corpus consisting of the DAs of speaker utterances and the backchannel types of listener utterances then used it to analyze the relationship between a speaker's and listener's utterances. Our findings clarify that the occurrence frequencies of a listener's backchannel types significantly depend on the DAs of the speaker's utterances. Since the goal of our research is to construct a dialogue system that generates a more natural backchannel, this classification method, which determines certain types of aids from the speaker's DA, will be beneficial to such a system.

References

  1. Shinya Fujie, Kenta Fukushima, and Tetsunori Kobayashi. 2005. Back-channel feedback generation using linguistic and nonlinguistic information and its application to spoken dialogue system. In INTERSPEECH. 889--892.Google ScholarGoogle Scholar
  2. Kohei Hara, Koji Inoue, Katsuya Takanashi, and Tatsuya Kawahara. 2018. Prediction of Turn-taking Using Multitask Learning with Prediction of Backchannels and Fillers. In INTERSPEECH. 991--995.Google ScholarGoogle Scholar
  3. Ryuichiro Higashinaka, Kenji Imamura, Toyomi Meguro, Chiaki Miyazaki, Nozomi Kobayashi, Hiroaki Sugiyama, Toru Hirano, Toshiro Makino, and Yoshihiro Matsuo. 2014. Towards an open-domain conversational system fully based on natural language processing. In International conference on Computational linguistics. 928--939.Google ScholarGoogle Scholar
  4. Lixing Huang, Louis-Philippe Morency, and Jonathan Gratch. 2010. Parasocial consensus sampling: Combining multiple perspectives to learn virtual human behavior. AAMAS 2, 1265--1272.Google ScholarGoogle Scholar
  5. Ryo Ishii, Ryuichiro Higashinaka, and Junji Tomita. 2018. Predicting Nods by using Dialogue Acts in Dialogue. In Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018).Google ScholarGoogle Scholar
  6. Ryo Ishii, Xutong Ren, Michal Muszynski, and Louis-Philippe Morency. 2021. Multimodal and Multitask Approach to Listener's Backchannel Prediction: Can Prediction of Turn-Changing and Turn-Management Willingness Improve Backchannel Modeling?. In Proceedings of the 21st ACM International Conference on Intelligent Virtual Agents. 131--138.Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Hanae Koiso, Yasuo Horiuchi, Syun Tutiya, Akira Ichikawa, and Yasuharu Den. 1998. An Analysis of Turn-Taking and Backchannels Based on Prosodic and Syntactic Features in Japanese Map Task Dialogs. In Language and Speech, Vol. 41. 295--321.Google ScholarGoogle ScholarCross RefCross Ref
  8. SENKO K. MAYNARD. 1986. On back-channel behavior in Japanese and English casual conversation. 24, 6 (1986), 1079--1108. Google ScholarGoogle ScholarCross RefCross Ref
  9. Toyomi Meguro, Ryuichiro Higashinaka, Yasuhiro Minami, and Kohji Dohsaka. 2010. Controlling listening-oriented dialogue using partially observable Markov decision processes. In International conference on computational linguistics. 761--769.Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Louis-Philippe Morency, Iwan de Kok, and Jonathan Gratch. 2008. Predicting Listener Backchannels: A Probabilistic Multimodal Approach. In IVA. 176--190.Google ScholarGoogle Scholar
  11. Markus Mueller, David Leuschner, Lars Briem, Maria Schmidt, Kevin Kilgour, Sebastian Stueker, and Alex Waibel. 2015. Using Neural Networks for Data-Driven Backchannel Prediction: A Survey on Input Features and Training Techniques. In Human-Computer Interaction: Interaction Technologies. 329--340.Google ScholarGoogle Scholar
  12. Chiharu Mukai. 1999. The Use of Back-channels by Advanced Learners of Japanese : Its Qualitative and Quantitative Aspects.Google ScholarGoogle Scholar
  13. Kazumi Ohira. 1998. Have you changed? Pragmatic transfer of back-channel behavior by Japanese bilingual speakers. University of Illinois at Urbana-Champaign.Google ScholarGoogle Scholar
  14. Robin Ruede, Markus Müller, Sebastian Stüker, and Alex Waibel. 2019. Yeah, Right, Uh-Huh: A Deep Learning Backchannel Predictor. 247--258.Google ScholarGoogle Scholar
  15. Khiet P. Truong, Ronald Poppe, and Dirk Heylen. 2010. A rule-based backchannel prediction model using pitch and pause information.. In INTERSPEECH. ISCA.Google ScholarGoogle Scholar
  16. N. Ward. 1996. Using prosodic clues to decide when to produce back-channel utterances. In Proceeding of Fourth International Conference on Spoken Language Processing. ICSLP '96, Vol. 3. 1728--1731 vol.3.Google ScholarGoogle ScholarCross RefCross Ref
  17. Nigel Ward and Wataru Tsukahara. 2000. Prosodic features which cue backchannel responses in English and Japanese. Journal of Pragmatics 32, 8 (2000), 1177--1207.Google ScholarGoogle ScholarCross RefCross Ref
  18. Takashi Yamaguchi, Koji Inoue, Koichiro Yoshino, Katsuya Takanashi, Nigel G Ward, and Tatsuya Kawahara. 2016. Analysis and prediction of morphological patterns of backchannels for attentive listening agents. In Proc. 7th International Workshop on Spoken Dialogue Systems. 1--12.Google ScholarGoogle Scholar

Index Terms

  1. Determining most suitable listener backchannel type for speaker's utterance

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Conferences
        IVA '22: Proceedings of the 22nd ACM International Conference on Intelligent Virtual Agents
        September 2022
        234 pages
        ISBN:9781450392488
        DOI:10.1145/3514197
        • General Chairs:
        • Carlos Martinho,
        • João Dias,
        • Program Chairs:
        • Joana Campos,
        • Dirk Heylen

        Copyright © 2022 Owner/Author

        Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 6 September 2022

        Check for updates

        Qualifiers

        • extended-abstract

        Acceptance Rates

        IVA '22 Paper Acceptance Rate21of51submissions,41%Overall Acceptance Rate53of196submissions,27%

        Upcoming Conference

        IVA '24
        ACM International Conference on Intelligent Virtual Agents
        September 16 - 19, 2024
        GLASGOW , United Kingdom

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader