Abstract
The quality of scientific oral presentations is often poor, owing to a number of factors, including public speaking anxiety. We present DynamicDuo, a system that uses an automated, life-sized, animated agent to help inexperienced scientists deliver their presentations in front of an audience. The design of the system was informed by an analysis of TED talks given by pairs of human presenters to identify the most common dual-presentation formats and transition behaviors used. We explore the usability and acceptability of DynamicDuo in both controlled laboratory-based studies and real-world environments, and its ability to decrease public speaking anxiety and improve presentation quality. In a within-subjects study (N = 12) comparing co-presenting with DynamicDuo against solo-presenting with conventional presentation software, we demonstrated that our system led to significant improvements in public speaking anxiety and speaking confidence for non-native English speakers. Judges who viewed videotapes of these presentations rated those with DynamicDuo significantly higher on speech quality and overall presentation quality for all presenters. We also explore the affordances of the virtual co-presenter through empirical evaluation of novel roles the agent can play in scientific presentations and novel ways it can interact with the speaker in front of the audience.
Similar content being viewed by others
References
Goodman A (2006) Why bad presentations happen to good causes. Andy Goodman & Cause Communication
Bishop J, Bauer K, Becker E (1998) A survey of counseling needs of male and female college students. J College Stud Dev 39:205–210
Weingart P, Joubert M (2019) The conflation of motives of science communication: Causes, consequences, remedies. J Sci Commun 18(3):1–13
National Academies of Sciences Engineering, and Medicine (2017) Communicating science effectively: a research agenda. The National Academies Press, Washington, DC
Niemann P, Bittner L, Schrögel P, Hauser C (2020) Science slams as edutainment: a reception study. Media Commun 8(1):177–190
Jackson J, Latané B (1981) All alone in front of all those people: stage fright as a function of number and type of co-performers and audience. J Pers Soc Psychol 40(1):73–81
Cassell J, Sullivan J, Prevost S, Churchill E (eds) (2000) Embodied conversational agents. The MIT Press, Cambridge, MA
Edge D, Savage J, Yatani K (2013) HyperSlides: dynamic presentation prototyping. Paper presented at CHI’13
Trinh H, Yatani K, Edge D (2014) PitchPerfect: integrated rehearsal environment for structured presentation preparation. Paper presented at the CHI’14
Saket B, Yang S, Tan H, Yatani K, Edge D (2014) TalkZones: section-based time support for presentations. Paper presented at the MobileHCI 2014 conference on human computer interaction with mobile devices and services
McVeigh-Schultz J, Kolesnichenko A, Isbister K (2019) Shaping pro-social interaction in VR: an emerging design framework. Paper presented at CHI’19
André E, Rist T, Müller J (1998) WebPersona: a lifelike presentation agent for the World-Wide Web. Knowl Based Syst 11(1):25–36
Fourati N, Richard A, Caillou S, Sabouret N, Martin J-C, Chanoni E, Clavel C (2016) Facial expressions of appraisals displayed by a virtual storyteller for children. Paper presented at Intelligent Virtual Agents
Noma T, Zhao L, Badler N (2000) Design of a virtual human presenter. IEEE Comput Graphics Appl 20(4):79–85
Nijholt A, van Welbergen H, Zwiers J (2005) Introducing an embodied virtual presenter agent in a virtual meeting room. Paper presented at the IASTED International Conference on Artificial Intelligence and Applications
Kang N, Brinkman W, van Riemsdijk M, Neerincx M (2013) An expressive virtual audience with flexible behavioral styles. IEEE Trans Affect Comput 4(4):326–340
Pertaub D, Slater M, Barker C (2002) An experiment on public speaking anxiety in response to three different types of virtual audience. Presence Teleoper Virt Environ 11(1):68–78
Chollet M, Scherer S (2017) Perception of virtual audiences. IEEE Comput Graphics Appl 37(4):50–59. https://doi.org/10.1109/MCG.2017.3271465
Pertaub D-P, Slater M, Barker C (2002) An experiment on public speaking anxiety in response to three different types of virtual audience. Pres Teleoper Virt Environ 11(1):68–78
Barreda-Ángeles M, Aleix-Guillaume S, Pereda-Baños A (2019) Users’ psychophysiological, vocal, and self-reported responses to the apparent attitude of a virtual audience in stereoscopic 360°-video. Virt Real 24:289–302
Trinh H, Asadi R, Edge D, Bickmore T (2017) RoboCOP: a robotic coach for oral presentations. Paper presented at the ACM international joint conference on pervasive and ubiquitous computing (Ubicomp)
Hoque M, Courgeon M, Mutlu B, Martin J-C, Picard R (2013) MACH: my automated conversation coach. Paper presented at the 15th international conference on ubiquitous computing (Ubicomp)
Anderson K, André E, Baur T, Bernardini S, Chollet M, Chryssafidou E, Damian I, Ennis C, Egges A, Gebhard P, Jones H, Ochs M, Pelachaud C, Porayska-Pomsta K, Rizzo P, Sabouret N (2013) The TARDIS framework: intelligent virtual agents for social coaching in job interviews. In: Cham, 2013. Advances in computer entertainment. Springer International Publishing, pp 476–491
Duncan S (1974) On the structure of speaker-auditor interaction during speaking turns. Lang Soc 3:161–180
Goodwin C (1981) Achieving mutual orientation at turn beginning. Conversational organization: interaction between speakers and hearers. Academic Press, New York, pp 55–89
Kendon A (1967) Some functions of gaze direction in social interaction. Acta Physiol (Oxf) 26:22–63
Sacks H, Schegloff EA, Jefferson G (1974) A simplest systematics for the organization of turn-taking for conversation. Language 50:696–735
Duncan S (1972) Some signals and rules for taking speaking turns in conversations. J Pers Soc Psychol 23:283–292
Raux A, Eskenazi M (2012) Optimizing the turn-taking behavior of task-oriented spoken dialog systems. ACM Trans Speech Lang Process 9(1):1–23
Ward N, Fuentes O, Vega A (2010) Dialog prediction for a general model of turn-taking. Paper presented at Interspeech
Koiso H, Horiuchi Y, Tutiya S, Ichikawa A, Den Y (1998) An analysis of turn-taking and backchannels based on prosodic and syntactic features in Japanese Map Task dialogs. Lang Speech 41:295–321
Gravano A, Hirschberg J (2011) Turn-taking cues in task-oriented dialogue. Comput Speech Lang 25(3):601–634
Hjalmarsson A (2011) The additive effect of turn-taking cues in human and synthetic voice. Speech Commun 53(1):23–35
Thorisson KR (1997) Gandalf: an embodied humanoid capable of real-time multimodal dialogue with people. Paper presented at autonomous agents ‘97
Cassell J, Bickmore T, Billinghurst M, Campbell L, Chang K, Vilhjalmsson H, Yan H Embodiment in conversational interfaces: rea. Paper presented at CHI’99
Huang L, Morency L, Gratch J (2011) Virtual Rapport 2.0. Paper presented at intelligent virtual agents (IVA)
Jonsdottir G, Thorisson K, Nivel E (2008) Learning smooth, human-like turntaking in realtime dialogue. Paper presented at intelligent virtual agents (IVA)
Chao C, Thomaz A (2016) Timed Petri nets for fluent turn-taking over multimodal interactoin resources in human-robot collaboration. Int J Robot Res 35(11):1330–1353
Jegou M, Lefebvre L, Chevaillier P (2015) A continuous model for the management of turn-taking in user-agent spoken interactions based on the variations of prosodic signals. Paper presented at intelligent virtual agents (IVA)
Trinh H, Ring L, Bickmore T (2015) DynamicDuo: co-presenting with virtual agents. Paper presented at CHI’15
Murali P, Ring L, Trinh H, Asadi R, Bickmore T (2018) Speaker hand-offs in collaborative human-agent oral presentations. Paper presented at the intelligent virtual agents (IVA)
Cassell J, Vilhjálmsson H, Bickmore T (2001) BEAT: the behavior expression animation toolkit. Paper presented at SIGGRAPH ‘01
Bickmore T, Trinh H, Hoppmann M, Asadi R (2016) Virtual agents in the classroom: experience fielding a co-presenter agent in university courses. Paper presented at intelligent virtual agents (IVA)
Spielberger C (1989) State-trait anxiety inventory: bibliography. Consulting Psychologists Press, Palo Alto, CA
Paul G (1966) Insight and desensitization in psychotherapy: an experiment in anxiety reduction. Stanford University Press, Stanford, CA
Goffman E (1981) Forms of talk. University of Pennsylvania Press, Philadelphia, pp 124–159
Kimani E, Shamekhi A, Parmar D, Bickmore T (2019) Stagecraft for scientists: exploring novel interaction formats for virtual co-presenter agents. Paper presented at intelligent virtual agents (IVA)
Holtzblatt K, Wendell J, Wood S (2004) Rapid contextual design: a how-to guide to key techniques for user-centered design. Elsevier, New York
Bakhtin M (1986) The Problem of Speech Genres. In: Holquist M, Emerson C (eds) Speech genres and other essays. University of Texas Press, Austin, pp 60–102
Trinh H, Bickmore T, Edge D, Ring L (2016) Thinking outside the box: co-planning scientific presentations with virtual agents. Paper presented at intelligent virtual agents (IVA)
Asadi R, Fell H, Bickmore T, Trinh H (2016) Real-time presentation tracking using semantic keyword spotting. Paper presented at Interspeech
Asadi R, Trinh A, Fell H, Bickmore T (2018) Quester: a speech-based question-answering support system for oral presentations. Paper presented at the international conference on intelligent user interfaces (IUI)
Asadi R, Trinh H, Bickmore T (2017) IntelliPrompter: speech-based dynamic note display interface for oral presentations. Paper presented at the international conference on multimodal interaction (ICMI)
Kimani E, Trinh H, Pedrelli P, Bickmore T (2019) You’ll be great: virtual agent-based cognitive restructuring to reduce public speaking anxiety. Paper presented at the 8th international conference on affective computing & intelligent interaction (ACII)
Kimani E (2019) A sensor-based framework for real-time detection and alleviation of public speaking anxiety. Paper presented at the international conference on affective computing & intelligent interaction (ACII) doctoral consortium
Acknowledgements
This material is based upon work supported by the National Science Foundation under Grant No. IIS-1514490. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Bickmore, T., Kimani, E., Shamekhi, A. et al. Virtual agents as supporting media for scientific presentations. J Multimodal User Interfaces 15, 131–146 (2021). https://doi.org/10.1007/s12193-020-00350-y
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s12193-020-00350-y