ABSTRACT
Social robots often have expressive faces. However, it is not always clear how to design expressions that show a certain emotion. We present a method for a social robot to learn the emotional meaning of its own facial expressions, based on which it can automatically generate faces for any emotion. The robot collects data from an imitation game where humans are asked to mimic the robot's facial expression. The interacting person does not need to explicitly input the meaning of the robot's face so the interaction is natural. We show that humans can successfully recognise the emotions from the learned facial expressions.
- Visage Technologies AB. 2018. visage|SDK Technical Specifications 10 ed.). Visage Technologies AB.Google Scholar
- John N. Bassili. 1979. Emotion recognition: The role of facial movement and the relative importance of upper and lower areas of the face. Journal of Personality and Social Psychology , Vol. 37, 11 (1979), 2049--2058. https://doi.org/10.1037/0022--3514.37.11.2049Google ScholarCross Ref
- Tony Belpaeme, James Kennedy, Aditi Ramachandran, Brian Scassellati, and Fumihide Tanaka. 2018. Social robots for education: A review. Science Robotics , Vol. 3, 21 (Aug. 2018), eaat5954. https://doi.org/10.1126/scirobotics.aat5954Google Scholar
- Michael J. Black and Yaser Yacoob. 1997. Recognizing Facial Expressions in Image Sequences Using Local Parameterized Models of Image Motion . International Journal of Computer Vision , Vol. 25, 1 (1997), 23--48. https://doi.org/10.1023/A:1007977618277Google ScholarDigital Library
- Sofiane Boucenna, Philippe Gaussier, Pierre Andry, and Laurence Hafemeister. 2010. Imitation as a communication tool for online facial expression learning and recognition. In 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, 5323--5328.Google ScholarCross Ref
- Cynthia Breazeal. 2003. Toward sociable robots. Robotics and autonomous systems , Vol. 42, 3--4 (2003), 167--175.Google Scholar
- Cynthia Breazeal, Daphna Buchsbaum, Jesse Gray, David Gatenby, and Bruce Blumberg. 2005. Learning From and About Others: Towards Using Imitation to Bootstrap the Social Understanding of Others by Robots . Artificial Life , Vol. 11, 1--2 (Jan. 2005), 31--62. https://doi.org/10.1162/1064546053278955Google ScholarDigital Library
- Nikhil Churamani, Pablo Barros, Erik Strahl, and Stefan Wermter. 2018. Learning empathy-driven emotion expressions using affective modulations. In 2018 International Joint Conference on Neural Networks (IJCNN). IEEE, 1--8.Google ScholarCross Ref
- Nikhil Churamani, Matthias Kerzel, Erik Strahl, Pablo Barros, and Stefan Wermter. 2017. Teaching emotion expressions to a human companion robot using deep neural architectures. In 2017 International Joint Conference on Neural Networks (IJCNN). IEEE, 627--634.Google ScholarCross Ref
- Brian R. Duffy. 2003. Anthropomorphism and the social robot. Robotics and Autonomous Systems , Vol. 42, 3--4 (March 2003), 177--190. https://doi.org/10.1016/S0921--8890(02)00374--3Google ScholarDigital Library
- Paul Ekman. 1992. Facial expressions of emotion: an old controversy and new findings. Philosophical Transactions of the Royal Society of London. Series B: Biological Sciences , Vol. 335, 1273 (Jan. 1992), 63--69. https://doi.org/10.1098/rstb.1992.0008Google Scholar
- Terrence Fong, Illah Nourbakhsh, and Kerstin Dautenhahn. 2003. A survey of socially interactive robots. Robotics and Autonomous Systems , Vol. 42, 3--4 (March 2003), 143--166. https://doi.org/10.1016/S0921--8890(02)00372-XGoogle Scholar
- Trevor Hastie, Robert Tibshirani, and Jerome Friedman. 2009. The elements of statistical learning: data mining, inference and prediction 2 ed.). Springer. http://www-stat.stanford.edu/ tibs/ElemStatLearn/Google Scholar
- Carroll E. Izard. 1994. Innate and universal facial expressions: Evidence from developmental and cross-cultural research. Psychological Bulletin , Vol. 115, 2 (1994), 288--299. https://doi.org/10.1037/0033--2909.115.2.288Google ScholarCross Ref
- David O Johnson, Raymond H Cuijpers, and David van der Pol. 2013. Imitating human emotions with artificial facial expressions. International Journal of Social Robotics , Vol. 5, 4 (2013), 503--513.Google ScholarCross Ref
- Alisa Kalegina, Grace Schroeder, Aidan Allchin, Keara Berlin, and Maya Cakmak. 2018. Characterizing the design space of rendered robot faces. In Proceedings of the 2018 ACM/IEEE International Conference on Human-Robot Interaction. ACM, 96--104.Google ScholarDigital Library
- Kwan Min Lee, Wei Peng, Seung-A Jin, and Chang Yan. 2006. Can Robots Manifest Personality?: An Empirical Test of Personality Recognition, Social Responses, and Social Presence in Human-Robot Interaction . Journal of Communication , Vol. 56, 4 (Dec. 2006), 754--772. https://doi.org/10.1111/j.1460--2466.2006.00318.xGoogle ScholarCross Ref
- Andrew N Meltzoff and M Keith Moore. 1997. Explaining facial imitation: A theoretical model. Infant and child development , Vol. 6, 3--4 (1997), 179--192.Google Scholar
- Jacqueline Nadel, Maud Simon, Pierre Canet, Robert Soussignan, P Blancard, Lola Canamero, and Philippe Gaussier. 2006. Human responses to an expressive robot. In Procs of the Sixth International Workshop on Epigenetic Robotics. Lund University.Google Scholar
- Verena Nitsch and Michael Popp. 2014. Emotions in robot psychology. Biological Cybernetics , Vol. 108, 5 (Oct. 2014), 621--629. https://doi.org/10.1007/s00422-014-0594--6Google ScholarDigital Library
- Tingfan Wu, Nicholas J Butko, Paul Ruvulo, Marian S Bartlett, and Javier R Movellan. 2009. Learning to make facial expressions. In 2009 IEEE 8th International Conference on Development and Learning. IEEE, 1--6.Google ScholarDigital Library
Index Terms
- Calibrate My Smile: Robot Learning Its Facial Expressions through Interactive Play with Humans
Recommendations
Using Facially Expressive Robots to Calibrate Clinical Pain Perception
HRI '17: Proceedings of the 2017 ACM/IEEE International Conference on Human-Robot InteractionIn this paper, we introduce a novel application of social robotics in healthcare: high fidelity, facially expressive, robotic patient simulators (RPSs), and explore their usage within a clinical experimental context. Current commercially-available RPSs, ...
Child's recognition of emotions in robot's face and body
HRI '11: Proceedings of the 6th international conference on Human-robot interactionSocial robots can comfort and support children who have to cope with chronic diseases. In previous studies, a "facial robot", the iCat, proved to show well-recognized emotional expressions that are important in social interactions. The question is if a ...
Mood contagion of robot body language in human robot interaction
The aim of our work is to design bodily mood expressions of humanoid robots for interactive settings that can be recognized by users and have (positive) effects on people who interact with the robots. To this end, we develop a parameterized behavior ...
Comments