Skip to main content

Evaluating Model that Predicts When People Will Speak to a Humanoid Robot and Handling Variations of Individuals and Instructions

  • Chapter
  • First Online:
Situated Dialog in Speech-Based Human-Computer Interaction

Part of the book series: Signals and Communication Technology ((SCT))

  • 726 Accesses

Abstract

We have tackled the problem of predicting when a user is likely to begin speaking to a humanoid robot. The generality of the prediction model should be examined so that it can be applied to various users. We present two empirical evaluations demonstrating that (1) our proposed model does not depend on the specific participants whose data were used in our previous data collection, and (2) the model can handle variations of individuals and instructions. We collect a data set to which 25 human participants in general public gave labels indicating whether or not they would be likely to begin speaking to the robot. We then train a new model with the collected data and determine its performance by cross validation and open tests. We also investigate the relation of how much individual participants feel likely to speak with a model parameter and the instruction given before data collections. Results show that our model can handle the variations.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 109.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    We will investigate more features other than the current ones.

  2. 2.

    http://www.aldebaran-robotics.com/

  3. 3.

    http://voicetext.jp/

References

  1. Chao C, Thomaz A (2010) Turn-taking for human-robot interaction. In: Proceedings of the AAAI fall symposium on dialog with robots, pp 132–134

    Google Scholar 

  2. Duncan S (1972) Some signals and rules for taking speaking turns in conversations. J Pers Soc Psychol 23:283–292

    Article  Google Scholar 

  3. Kanda T, Ishiguro H, Imai M, Ono T (2004) Development and evaluation of interactive humanoid robots. Proc IEEE (Spec Iss Human Interact Rob Psychol Enrich) 92:1839–1850

    Google Scholar 

  4. Kendon A (1967) Some functions of gaze direction in social interaction. Acta Psychol 26:22–63

    Article  Google Scholar 

  5. Kim W, Ko H (2001) Noise variance estimation for Kalman filtering of noisy speech. IEICE Trans Inf Syst E84-D(1):155–160

    Google Scholar 

  6. Komatani K, Ueno S, Kawahara T, Okuno HG (2005) User modeling in spoken dialogue systems to generate flexible guidance. User Model User-Adap Interact 15(1):169–183

    Article  Google Scholar 

  7. Kruijff-Korbayov I, Cuayhuitl H, Kiefer B, Schrder M, Cosi P, Paci G, Sommavilla G, Tesser F, Sahli H, Athanasopoulos G, Wang W, Enescu V, Verhelst W (2012) Spoken language processing in a conversational system for child-robot interaction. In: Proceedings of the interspeech workshop on child-computer interaction, pp 132–134

    Google Scholar 

  8. Lee A, Nakamura K, Nisimura R, Saruwatari H, Shikano K (2004) Noise robust real world spoken dialogue system using GMM based rejection of unintended inputs. In: Proceedings of interspeech, pp 173–176

    Google Scholar 

  9. Mori M, MacDorman KF, Kageki N (2012) The uncanny valley. Rob Autom Mag 19(2):98–100

    Article  Google Scholar 

  10. Sacks H, Schegloff EA, Jefferson G (1974) A simplest systematics for the organization of turn-taking for conversation. Language 50(4):696–735

    Article  Google Scholar 

  11. Skantze G, Gustafson J (2009) Attention and interaction control in a human-human-computer dialogue setting. In: Proceedings of the SIGDIAL 2009 conference, pp 310–313

    Google Scholar 

  12. Sugiyama T, Komatani K, Sato S (2012) Predicting when people will speak to a humanoid robot. In: Proceedings of the international workshop on spoken dialog systems

    Google Scholar 

  13. Vertegaal R, Slagter R, van der Veer GC, Nijholt A (2001) Eye gaze patterns in conversations: there is more to conversational agents than meets the eyes. In: Proceedings of the SIGCHI conference on human factors in computing systems, pp 301–308

    Google Scholar 

  14. Yang Y (1999) An evaluation of statistical approaches to text categorization. Inf Retr 1:69–90

    Article  Google Scholar 

  15. Yoon S, Yoo CD (2002) Speech enhancement based on speech/noise-dominant decision. IEICE Trans Inf Syst E85-D(4):744–750. http://ci.nii.ac.jp/naid/110003219991/

Download references

Acknowledgments

This research has been partly supported by the JST PRESTO Program.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Takaaki Sugiyama .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2016 Springer International Publishing Switzerland

About this chapter

Cite this chapter

Sugiyama, T., Komatani, K., Sato, S. (2016). Evaluating Model that Predicts When People Will Speak to a Humanoid Robot and Handling Variations of Individuals and Instructions. In: Rudnicky, A., Raux, A., Lane, I., Misu, T. (eds) Situated Dialog in Speech-Based Human-Computer Interaction. Signals and Communication Technology. Springer, Cham. https://doi.org/10.1007/978-3-319-21834-2_13

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-21834-2_13

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-21833-5

  • Online ISBN: 978-3-319-21834-2

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics