skip to main content
10.1145/3371382.3378253acmconferencesArticle/Chapter ViewAbstractPublication PageshriConference Proceedingsconference-collections
abstract

Your Eyes Never Lie: A Robot Magician Can Tell if You Are Lying

Authors Info & Claims
Published:01 April 2020Publication History

ABSTRACT

Detecting lies in a real-world scenario is an important skill for a humanoid robot that aims to act as a teacher, a therapist, or a caregiver. In these contexts, it is essential to detect lies while preserving the pleasantness of the social interaction and the informality of the relation. This study investigates whether pupil dilation related to an increase in cognitive load can be used to swiftly identify a lie in an entertaining scenario. The iCub humanoid robot plays the role of a magician in a card game, telling which card the human partner is lying about. The results show a greater pupil dilation in presence of a false statement even if in front of a robot and without the need of a strictly controlled scenario. We developed a heuristic method (accuracy of 71.4% against 16.6% chance level) and a random forest classifier (precision and recall of 83.3%) to detect the false statement. Additionally, the current work suggests a potential method to assess the lying strategy of the partner.

References

  1. A. Westbrook and T. S. Braver, "Cognitive effort: A neuroeconomic approach," Cognitive, Affective and Behavioral Neuroscience, vol. 15, no. 2. Springer New York LLC, pp. 395--415, 22-Jun-2015.Google ScholarGoogle Scholar
  2. J. Beatty, "Task-Evoked Pupillary Responses, Processing Load, and the Structure of Processing Resources," 1982.Google ScholarGoogle Scholar
  3. A. Koenig et al., "Real-time closed-loop control of cognitive load in neurological patients during robot-assisted gait training.," IEEE Trans. Neural Syst. Rehabil. Eng., vol. 19, no. 4, pp. 453--64, Aug. 2011.Google ScholarGoogle ScholarCross RefCross Ref
  4. J. A. Stern, L. C. Walrath, and R. Goldstein, "The endogenous eyeblink.," Psy-chophysiology, vol. 21, no. 1, pp. 22--33, Jan. 1984.Google ScholarGoogle ScholarCross RefCross Ref
  5. J. Fallon, "Box of Lies." [Online]. Available: https://www.youtube.com/watch?v=Md4QnipNYqM.Google ScholarGoogle Scholar
  6. C. R. Honts, D. C. Raskin, and J. C. Kircher, "Mental and physical countermeas-ures reduce the accuracy of polygraph tests.," J. Appl. Psychol., vol. 79, no. 2, pp. 252--9, Apr. 1994.Google ScholarGoogle ScholarCross RefCross Ref
  7. B. M. DePaulo, J. J. Lindsay, B. E. Malone, L. Muhlenbruck, K. Charlton, and H. Cooper, "Cues to deception.," Psychol. Bull., vol. 129, no. 1, pp. 74--118, 2003.Google ScholarGoogle ScholarCross RefCross Ref
  8. D. P. Dionisio, E. Granholm, W. A. Hillix, and W. F. Perrine, "Differentiation of deception using pupillary responses as an index of cognitive processing.," Psy-chophysiology, vol. 38, no. 2, pp. 205--11, Mar. 2001.Google ScholarGoogle ScholarCross RefCross Ref
  9. S. M. Kassin, "On the Psychology of Confessions: Does Innocence Put Innocents at Risk?," Am. Psychol., vol. 60, no. 3, pp. 215--228, Apr. 2005.Google ScholarGoogle ScholarCross RefCross Ref
  10. J. G. May, R. S. Kennedy, M. C. Williams, W. P. Dunlap, and J. R. Brannan, "Eye movement indices of mental workload," Acta Psychol. (Amst)., vol. 75, no. 1, pp. 75--89, 1990.Google ScholarGoogle ScholarCross RefCross Ref
  11. M. Nakayama and Y. Shimizu, "Frequency analysis of task evoked pupillary response and eye-movement," in Proceedings of the Eye tracking research & applications symposium on Eye tracking research & applications - ETRA'2004, 2004, pp. 71--76.Google ScholarGoogle Scholar
  12. K. F. Van Orden, W. Limbert, S. Makeig, and T. P. Jung, "Eye activity correlates of workload during a visuospatial memory task.," Hum. Factors, vol. 43, no. 1, pp. 111--21, 2001.Google ScholarGoogle ScholarCross RefCross Ref
  13. B. C. Goldwater, "Psychological significance of pupillary movements," Psychol. Bull., vol. 77, no. 5, pp. 340--355, May 1972.Google ScholarGoogle ScholarCross RefCross Ref
  14. A. Szulewski, N. Roth, and D. Howes, "The Use of Task-Evoked Pupillary Re-sponse as an Objective Measure of Cognitive Load in Novices and Trained Physi-cians: A New Tool for the Assessment of Expertise," Acad. Med., vol. 90, no. 7, pp. 981--987, Jul. 2015.Google ScholarGoogle ScholarCross RefCross Ref
  15. M. Ahmad, B. Jasmin, L. Katrin, and F. Eyssel, "Trust and Cognitive Load During Human-Robot Interaction," 2019.Google ScholarGoogle Scholar
  16. J. Klingner, "Measuring Cognitive Load During Visual Task by Combining Pupil-lometry and Eye Tracking," Ph.D. Diss., no. May, 2010.Google ScholarGoogle Scholar
  17. G. Hossain and M. Yeasin, "Understanding effects of cognitive load from pupil-lary responses using hilbert analytic phase," IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit. Work., pp. 381--386, 2014.Google ScholarGoogle Scholar
  18. A. Gaggioli, "Beyond the Truth Machine: Emerging Technologies for Lie Detec-tion," Cyberpsychology, Behav. Soc. Netw., vol. 21, no. 2, pp. 144--144, Feb. 2018.Google ScholarGoogle ScholarCross RefCross Ref
  19. M. Gamer, "Detecting of deception and concealed information using neuroimag-ing techniques," in Memory Detection, B. Verschuere, G. Ben-Shakhar, and E. Meijer, Eds. Cambridge: Cambridge University Press, 2011, pp. 90--113.Google ScholarGoogle Scholar
  20. J. Gonzalez-Billandon et al., "Can a Robot Catch You Lying? A Machine Learn-ing System to Detect Lies During Interactions," Front. Robot. AI, vol. 6, Jul. 2019.Google ScholarGoogle Scholar
  21. K. Kobayashi and S. Yamada, "Human-robot interaction design for low cognitive load in cooperative work," in RO-MAN 2004. 13th IEEE International Work-shop on Robot and Human Interactive Communication (IEEE Catalog No.04TH8759), pp. 569--574.Google ScholarGoogle Scholar
  22. S. M. Al Mahi, M. Atkins, and C. Crick, "Learning to assess the cognitive capacity of human partners," ACM/IEEE Int. Conf. Human-Robot Interact., pp. 63--64, 2017.Google ScholarGoogle Scholar
  23. A. M. Aroyo et al., "Can a Humanoid Robot Spot a Liar?," IEEE-RAS 18th Int. Conf. Humanoid Robot., 2018.Google ScholarGoogle Scholar
  24. C. Wangwiwattana, X. Ding, and E. C. Larson, "PupilNet, Measuring Task Evoked Pupillary Response using Commodity RGB Tablet Cameras," Proc. ACM Interactive, Mobile, Wearable Ubiquitous Technol., vol. 1, no. 4, pp. 1--26, Jan. 2018.Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. G. Metta, G. Sandini, D. Vernon, L. Natale, and F. Nori, "The iCub humanoid robot," in Proceedings of the 8th Workshop on Performance Metrics for Intelli-gent Systems - PerMIS '08, 2008, p. 50.Google ScholarGoogle Scholar
  26. L. Riek, "Wizard of Oz Studies in HRI: A Systematic Review and New Reporting Guidelines," J. Human-Robot Interact., vol. 1, no. 1, pp. 119--136, 2012.Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. S. Mathôt, J. Fabius, E. Van Heusden, and S. Van der Stigchel, "Safe and sensible preprocessing and baseline correction of pupil-size data," Behav. Res. Methods, vol. 50, no. 1, pp. 94--106, 2018.Google ScholarGoogle ScholarCross RefCross Ref
  28. S. Shapiro and M. Wilk, "An analysis of variance test for normality (complete samples)," Z. Bibliothekswes. Bibliogr., vol. 52, no. 6, pp. 311--319, 2005.Google ScholarGoogle Scholar
  29. R. B. D'Agostino, A. Belanger, and R. B. D'Agostino, "A Suggestion for Using Powerful and Informative Tests of Normality," Am. Stat., vol. 44, no. 4, p. 316, Nov. 1990.Google ScholarGoogle Scholar
  30. F. Wilcoxon, "Individual Comparisons by Ranking Methods," Biometrics Bull., vol. 1, no. 6, p. 80, Dec. 1945.Google ScholarGoogle ScholarCross RefCross Ref
  31. S. G. K. Patro and K. K. sahu, "Normalization: A Preprocessing Stage," Iarjset, no. April, pp. 20--22, 2015.Google ScholarGoogle Scholar
  32. Nitesh V. Chawla, K. W. Bowyer, and L. O. Hall, "SMOTE: Synthetic Minority Over-sampling Technique Nitesh," J. Artif. Intell. Res., 2006.Google ScholarGoogle Scholar

Index Terms

  1. Your Eyes Never Lie: A Robot Magician Can Tell if You Are Lying

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      HRI '20: Companion of the 2020 ACM/IEEE International Conference on Human-Robot Interaction
      March 2020
      702 pages
      ISBN:9781450370578
      DOI:10.1145/3371382

      Copyright © 2020 Owner/Author

      Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 1 April 2020

      Check for updates

      Qualifiers

      • abstract

      Acceptance Rates

      Overall Acceptance Rate192of519submissions,37%

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader