Abstract
Autonomous agents, both software and robotic, are becoming increasingly common. They are being used to supplement human operators in accomplishing complex tasks, often acting as collaborators or teammates. Agents can be designed to keep their human operators ‘in the loop’ by reporting information concerning their internal decision making process. This transparency can be expressed in a number of ways, including the communication of the human and agent’s respective responsibilities. Agents can communicate information supporting transparency to human operators using visual, auditory, or a combination of both modalities. Based on this information, we suggest an approach to exploring the utility of the teamwork model of transparency. We propose some considerations for future research into feedback supporting teamwork transparency, including multimodal communication methods, human-like feedback, and the use of multiple forms of automation transparency.
You have full access to this open access chapter, Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
There is an increasing reliance on autonomous agents to perform functions previously done by human actors. Agents are being used in business to maintain interactivity between businesses and their customers [1]; in the U.S. military to conduct dangerous activities such as explosive disposal and firefighting [2]; and in Homeland Security to help analysts process vast quantities of intelligence information [3]. As technology improves, so too does the need to understand how agents can be implemented in a way that yields the most effective partnership between humans and technology [4].
1.1 Agents and Their Roles in Human-Agent Teams
The progression of technology has led to circumstances where complex tasks can be delegated to a machine, automated, and thus be done with fewer errors and by fewer people [5, 6]. One means of implementing automation is through an agent. Agents are hardware or software-based computer system that are characterized by four properties: autonomy, (the ability to operate without direct human intervention for a significant length of time); social ability, (the capacity to interact with humans or other agents using language); reactivity, (the ability to perceive the environment and react to it); and proactiveness, (ability to exhibit goal directed behavior in anticipation of future events) [6–9]. In this manuscript, references to ‘agents’ will use this definition. Robots, by definition, are a kind of agent that are physically embodied [2, 10]. In systems where humans work with agents as team members rather than operators, agents usually play one of three roles: individual support, team support, or team member [9, 10]. When agents support individual team members, they can either be person-specific or task-specific [9]. When agents support a team as a whole, they act to facilitate the group’s teamwork [10]. When agents assume the role of an equal team member, they are expected to act similarly to their human teammates [9, 10]. Thus, similar to a human team member, an agent in a human-agent team must also communicate relevant information to their teammate to maintain shared knowledge and shared intent [9].
In many circumstances, the introduction of automation has changed the human operator’s task to that of monitor and backup [5, 11]. Autonomous agents and their ability to choose goals and act independently also require coordination and cooperation [12]. In systems with flexible automation, agents act in tandem with humans to make decisions, mirroring the relationship between a human and a subordinate [2]. The flexibility gained by this joint decision-making requires continuous collaboration and communication [2]. Human teams have the advantage of flexible communication strategies. Team members elicit relevant information from other team members, which in turn supports development of effective shared mental models [13]. However, in human-agent teams, the agent team members often cannot effectively share information without some means of translating their ‘understanding’ [3, 14]. By establishing a common understanding of the situation, the task, the team members and their respective duties, human teams are able to coordinate effectively [13, 15]. Similarly, effective human-agent teams also maintain a shared understanding of the situation and of their teammates [9]. When both parties maintain a shared understanding of the team environment, they can give, seek, and receive clarifying feedback, which are critical actions for teamwork.
To support the shared awareness and intent needed to perform as an effective human-agent team member, the agent must share information pertaining to its historical and current operation, how the underlying algorithms of the agent govern its behavior (the agent’s “reasoning”), and the extent to which it acts in accordance with the designer’s intent and the operator’s goals [2, 16, 17]. Agents that communicate their performance abilities, intent, reasoning process, and future plans in a way that facilitates operators’ comprehension of such provide transparency [18, 19].
2 Transparency
2.1 Transparency Overview
In the context of human-agent interaction, transparency has been described as a method by which a human and a machine can gain shared intent and awareness [19]. A transparent system facilitates this understanding by explaining its choices and behaviors, allowing its human operators to understand the way it works [20]. One approach, the Belief, Desire, and Intention (BDI) view describes the agent as having mental attitudes [11, 21]. The BDI approach to transparency, communicating the information, motivational, and deliberative states of the agent, helped inspire the Situation awareness-based Agent Transparency (SAT) model [18, 21]. The SAT model informs the design of transparent systems by supporting the human operator’s situation awareness [18]. In the SAT model, a transparent system communicates three levels of information [18]:
-
Level 1 describes the agent’s current actions and plans and its knowledge of the environment,
-
Level 2 describes the agent’s underlying reasoning behind its actions and plans, and
-
Level 3 describes the agent’s predictions about its future state or outcomes of its planned actions
Given the complexities of the transparency construct and the information needed to support it, Lyons divides transparency into two categories, Robot-to-Human transparency and Robot-of-Human transparency, a useful demarcation when describing humans and autonomous agents as teammates [19].
2.2 Models of Transparency
Robot-to-Human Transparency describes models of transparency which focus on the agent’s information about the world [19]. The following models are in this category: the intentional model, the task model, the analytical model, and the environment model [19]. The intentional model focuses on communicating the purpose of the system through the use of social intent cues; the task model focuses on communicating information pertaining to the agent’s task, expressing its goals and its progress towards meeting those goals, its capabilities, and its performance while pursuing those goals; the analytical model focuses on the underlying principles the agent uses to make decisions; and the environment model focuses on communicating variance in terrain, weather, and temporal constraints to humans [17, 19].
Robot-of-Human Transparency describes two models of transparency focusing on the communication of the agent’s awareness of the state of human teammates [19]. The teamwork model focuses on the division of labor between the agent and the human and the human state model focuses on the agent’s communicating their understanding of the human’s cognitive, emotional, and physical state [19]. This delineation of models is particularly important to keep in mind, given the parallels between human teams and human-agent teams. Robot-to-Human agent transparency models describe information relevant to task performance, while Robot-of-Human transparency models focus on the members of the human-agent team.
2.3 Effects of Transparency
The implementation of features that support agent transparency can have a positive influence on the relationship between the human and the agent working together. Depending on the kind of transparency explored and the amount of information presented to the user, information supporting transparency can influence operator trust, situation awareness, and workload.
Trust.
In the context of a human-agent team, trust can be defined as “the attitude that an agent will help achieve an individual’s goals in a situation characterized by uncertainty and vulnerability” [16]. In a transparent system, human observation of the agent’s actions—such as its history of action, capability, and reliability—should be supplemented by information concerning the goals, reasoning, and situational information that led to these actions [5, 16, 22]. By providing this information, which supports transparency, the human’s trust in the agent can be reasonably informed and more solidly held [7, 16]. However, trust is continuously updated [7]. If the agent communicates uncertainty or reports incorrect information, the human operator can use that information to calibrate their trust, matching their trust with the system’s capabilities, leading to appropriate use [7, 16].
With the incorporation of agents into teams as equal team members, however, humans and agent system designers must account for both trust in the agent as an automated system as well as trust in a teammate to act in the best interests of the team [9, 16]. In pursuit of transparency, system designers attempt to make sure that the agent communicates knowledge about itself, its goals, its underlying reasoning, and its awareness of environmental factors, but this communication of information is most frequently one-way [2, 17, 18].
Situation Awareness.
When humans work with robots, the human needs to maintain situation awareness (SA) in order to make appropriate decisions [23]. Situation awareness refers to the process by which a human attempts to understand their environment and use that understanding to perform competently in a situation as it occurs [24]. When working with agents, SA may include awareness of the environment in which agents may be located or what the agent is doing [23, 25]. Agent transparency can be used to keep the human operator from being pulled ‘out of the loop’ by allowing human operators to focus only on information relevant to the mission [2, 25]. Contextual awareness of an agent is a key factor in the success of human operators, and the communication of the agent’s intent can facilitate overall SA [25–27]. Global SA requires awareness of not only the immediate working environment, but the relationship between the agent and the human within that environment, so transparency specific to the human-agent relationship can support that awareness.
Workload.
A perennial concern is the impact of adding more information to an interface. Supporting transparency could require additional information, which may affect the human’s workload [18]. Although a unitary definition of workload does not exist, it can be conceptualized as the perceived impact of task demand imposed on the human operator and the associated physiological responses [28, 29]. Additional information supporting transparency may cognitively overload the operator, leading to performance decrements [30]. However, if additional information mitigates workload that would have been caused by having to recall or estimate information, the added information may lead to similar, if not lower, levels of workload overall [18, 30]. If overload is a concern, then features that reduce workload to a more manageable level are preferred. In two studies on joint human-agent decision-making, increased amounts of information supporting transparency were added to an interface without a significant increase to workload [31, 32]. Thus, transparency can possibly lead to benefits without a noticeable increase to the human teammates’ workload.
2.4 Communication and the Teamwork Model of Transparency
In the pursuit of transparency, agent interfaces have been designed to communicate the agent’s behaviors, goals, reasoning, and environmental constraints to the human operator in order to facilitate shared intent and shared awareness [17, 18]. While information encompassed by the Robot-to-Human transparency models has been explored as a means of supporting humans’ understanding of their agent teammates, less work has been done exploring the influence of Robot-of-Human transparency. While research concerning automated systems responding to human physiological states and the dynamic division of labor between humans and agents exist, fewer studies focus on the transparency of these systems [2, 19].
A particular area of interest is how agents can express the human’s and agent’s fulfillment of their responsibilities, and how the communication of this information could influence the human’s relationship with their agent team member. In human teams, team members can engage in mutual performance monitoring, a behavior where team members keep track of all team members’ performance, which can be coupled with feedback [33, 34]. This kind of feedback reflects the agent’s model of the human operator back to that selfsame operator, allowing the human and agent to establish a greater shared awareness of both the human’s and agent’s roles in the team and how they fulfill those roles [16, 34].
One study found that humans who worked with a highly autonomous robot attributed more blame to the robot than those who worked with a low autonomy robot [35]. While increased transparency led to a marginally significant reduction in the attribution of credit to other group members working with the highly autonomous robot, the robot’s feedback did not influence credit or blame to the robot or the self [35]. While an explanation of the robot’s reasoning did not influence the aforementioned factors, feedback concerning the robot’s and human’s performance of their roles may do so. While human and robot role responsibility feedback may lead to different credit and blame attribution, feedback concerning human operators’ shortfalls may influence the human’s perception of their own trustworthiness. Similarly to how humans can determine the trustworthiness of an agent through its history of action, capability, and reliability, they may also determine their own trustworthiness by seeing this same information about themselves [5, 16]. However, if the human feels that their own trustworthiness is low, they may feel the need to delegate tasks to the agent [16]. The impact of this kind of information, especially with robots and other embodied agents, is underexplored, despite the implications it may have towards the relationship between humans and agents. The means by which this information is communicated, however, may emphasize benefits or mitigate the possible negative repercussions of this communication. Hence, an exploration of different modes of communication is of interest.
2.5 Multimodal Feedback
Research in the field of agent transparency focuses on the communication of different kinds of information in order to maintain shared intent and awareness [19]. While the content of this communication is important, the means by which it is communicated is notable as well. Content is most frequently communicated through auditory or visual methods [36]. Humans can only process a limited amount of visual or auditory information, and the available mental effort used for each is distinct, so communicating information across multiple modes can extend mental limitations by splitting the burden across multiple channels [37]. Multimodal communication, communication across more than one sensory channel, allows for an increased complexity in communication through the use of redundant or non-redundant signals. The transmission of redundant signals can ensure that the message is received; the transmission of non-redundant signals can communicate two separate messages simultaneously, the modulation of a message, or the communication of an entirely new message [38]. Multimodal communication, can influence workload, which in turn can affect error rate and operator effectiveness [36, 38]. There are several methods in which agents can effectively integrate visual and auditory feedback.
Visual feedback.
The most common form of feedback is visual, with multimodal research investigating the effects of supplementing the visual modality [36]. In learning environments, information is usually presented to learners visually, either through text or graphics [39]. Text feedback has the advantage of facilitating understanding of complex, semantically-rich content [40]. In addition to the content of the message, the social presentation can influence humans’ responses to what was written. Increased etiquette, expressed by automated systems warning operators before giving them support and avoiding interruptions during requested actions, has been shown to lead to better performance and improved trust, though it has hampered situation awareness [7, 41]. The ubiquity of text means that it is frequently supplemented by other methods of communication [42]. Text-based feedback has been paired with graphics and speech, yielding more comprehensible output and more creative solutions [43, 44]. While agents can use disembodied text to communicate information, an agent with a human-like avatar can potentially provide an emotional connection that can create a more positive relationship between the agent and their human operator [42].
Software agents that are depicted using virtual characters are capable of communicating using human-like verbal and nonverbal signals [45]. A virtual character can provide feedback through the use of facial expressions or gaze [7, 45]. Positive facial expressions on pedagogical agents’ characters have been used to facilitate learning, motivate learners, and aid in attitudinal learning [45]. Gestures as social cues can be used to draw attention to an important feature in an interface or can be coupled with speech to improve the recall of verbal information and captivate the human operator [42, 46]. When robot gestures were combined with synthetic speech, it was more positively evaluated than when it communicated using speech alone [47].
The desired type of agent feedback—gestures, facial expressions, and so on—informs the design of a robot. A robot’s shape can potentially influence how it is seen; in one study, spider-legged robots were seen as more aggressive than wheeled robots, while robots with arms were seen as more intelligent than those without [48]. An agent with a human face or avatar may be more likely to engender human-like treatment from their human teammates, but human-like treatment may not be the desired response [49]. These different shapes are conducive to different gestures. A robot with arms and legs can make different gestures than a robot with no arms and wheels. The addition of expressive lights can add an entirely new dimension in communication, with speed, regularity, and color providing an avenue through which messages can be communicated [50]. The major limit in visual communication seems to lie in physical feasibility and human understanding.
Auditory feedback.
Sound has often been used as a social cue to focus people’s attention and provide feedback [36]. One study found that when earcons (i.e., abstract musical tones) were used to indicate movement of an interface element, it led to slower completion of a highlighting task than visual highlighting alone [36]. When abstract tones were used as robot feedback, participants did not extend assistance to it as much as when it had a voice, either synthetic or human [51]. When working with a mix of synthetic and human speech, users’ perception of their own performance was higher than users who only received a synthetic voice, but their actual performance in a series of office tasks was worse than their counterparts who only received synthetic speech [52]. Synthetic speech, unfortunately, is judged based on its intelligibility, naturalness, and acceptability to the human [53]. Participants exhibited faster response latencies when listening to natural voice compared to those who were listening to a synthetic voice [54]. Overall, speech is suited for presenting short, semantically simple content which carries only essential information [40, 42].
3 Experimental Approach
The exploration of human-agent teams, agent transparency, and multimodal communication has set the stage for a proposed approach to the investigation of transparency in human-agent teams. Specifically, this investigation seeks to make the case that the influence of an agent’s multimodal feedback in response to human operators’ meeting and not meeting their responsibilities within the division of labor is an area of research that has not yet been plumbed.
3.1 Experimental Considerations
A central question in the exploration of information concerning division of labor is how that information is expressed to the human operator. In tasks where humans and agents must work together to complete a task, the teamwork model of transparency suggests that the agent inform the human about their respective responsibilities and if they are being fulfilled. If an agent communicates to its human counterpart that they are fulfilling or failing to fulfill their responsibilities, how will that communication influence the human’s performance and their relationship with that agent? Establishing how feedback concerning division of labor can influence the relationship between human and agent may be useful. Additionally, a human-appearing agent leads to human-like expectations from it, which may or may not be desired [17, 49]. Will human-like social cues and non-human-like social cues yield equal benefits of transparency? Would teamwork transparency alone provide the same benefits that robot-to-human transparency does, or would they work better together? Scales pertaining to Workload, Trust, and Situation Awareness are useful indicators to determine the extent to which the information supporting teamwork transparency facilitates a beneficial relationship between the agent and the human.
3.2 Research Questions
Information pertaining to division of labor can be communicated multimodally, with human-like interfaces, using a human-like avatar and natural voice feedback, or non-human interfaces, using flashing lights and beeps. One aim of this approach is to determine the extent to which a human-like presentation of information supporting transparency influences operator behaviors. Human-like robots have resulted in a specific behavioral pattern from human operators [49].
Additionally, information supporting the teammate transparency model may require precise communication, so the use of non-redundant multimodal messaging should be explored as well. Non-redundant multimodal signals can be used to communicate modulated messages, which allows for a finer-tuned message concerning division of labor [38].
Furthermore, another area that bears exploring is the coordination of different models of transparency. Would a model of transparency, such as the SAT model, benefit from the addition of information supporting teamwork transparency? Would the operator attend to features of the interface supporting their situation awareness if periodically reminded of their responsibilities? Given the parallels between human teamwork and human-agent teamwork, determining the utility of information supporting teamwork transparency on its own and combined with a task-oriented transparency is a key area of exploration.
4 Evaluation
Dependent variables pertaining to the utility of human-agent transparency include performance, how well the agent supports the operator, and the human operator’s relationship with the agent. Performance, a key dependent variable, can include the successful completion of decision-making tasks, comprehension of presented information, and search tasks. The human’s performance and perceptions of their performance should be measured. An agent can support the operator by assisting in the maintenance of situation awareness and workload in a way that serves performance, so these two factors should be evaluated. The human operator’s relationship with the agent includes trust, which influences automation use, disuse, and overreliance [16]. Given the large impact that trust can have, it should be evaluated as well. Additionally, it is also important to evaluate the human’s perception of the agent, including technology acceptance and perceived usability.
5 Conclusion
To maintain the benefits of keeping the human operator ‘in the loop,’ autonomous agents must maintain transparency. As agents are increasingly tasked to act as teammates, however, they need to also communicate information to support teamwork, rather than just information to support operators’ tasks. Like mutual performance monitoring can aid human teams, the presentation of information supporting the teamwork model of transparency should benefit human-agent teams. The communication of role responsibility informs operators of the agent’s understanding of the humans’ and agents’ responsibilities in the system and how those responsibilities are being fulfilled. Providing feedback about the human’s actions as a team member may allow for greater teamwork between humans and agents. Presentation of this information by an agent may have unintended consequences, though, so research should look at presenting information using both visual and auditory communication methods, both human-like and non, and with other forms of automation transparency. This research will inform the design of agents for human-agent teams where an authentic artificial teammate is desired. As agents become more complex and are able to do more, our understanding of how humans treat their teammates, human or not, becomes even more necessary to facilitate effective performance from human-agent teams.
References
Saade, R., Vahidov, R., Yu, B.: Agents and E-commerce: beyond automation. In: Americas Conference on Information Systems. Puerto Rico (2015)
Chen, J.Y., Barnes, M.J.: Human–agent teaming for multirobot control: a review of human factors issues. IEEE Trans. Hum.-Mach. Syst. 44(1), 13–29 (2014)
Yen, J., et al.: Agents with shared mental models for enhancing team decision makings. Decis. Support Syst. 41(3), 634–653 (2006)
Sheridan, T.B., Parasuraman, R.: Human-automation interaction. Rev. Hum. Factors Ergon. 1(1), 89–129 (2005)
Parasuraman, R., Sheridan, T.B., Wickens, C.D.: A model for types and levels of human interaction with automation. IEEE Trans. Syst. Man Cybern. Part A Syst. Hum. 30(3), 286–297 (2000)
Zhu, H., Hou, M.: A Literature Review on Operator Interface Technologies for Network Enabled Operational Environments Using Complex System Analysis. W7711-083931/001/TOR: Defence R & D Canada, Toronto (2009)
de Visser, E.J., Cohen, M., Freedy, A., Parasuraman, R.: A design methodology for trust cue calibration in cognitive agents. In: Shumaker, R., Lackey, S. (eds.) VAMR 2014, Part I. LNCS, vol. 8525, pp. 251–262. Springer, Heidelberg (2014)
Wooldridge, M., Jennings, N.R.: Intelligent agents: theory and practice. Knowl. Eng. Rev. 10(2), 115–152 (1995)
Sycara, K., Sukthankar, G.: Literature review of teamwork models, Carnegie Mellon University CMU-RI-TR-06-50 (2006)
Sukthankar, G., Shumaker, R., Lewis, M.: Intelligent agents as teammates. In: Theories of Team Cognition: Cross-Disciplinary Perspectives, pp. 313–343 (2012)
Urlings, P., et al.: A future framework for interfacing BDI agents in a real-time teaming environment. J. Netw. Comput. Appl. 29(2), 105–123 (2006)
Atkinson, D.J., Clancey, W.J., Clark, M.H.: Shared awareness, autonomy and trust in human-robot teamwork. In: Papers from the 2014 AAAI Spring Symposium on Artificial Intelligence and Human-Computer Interaction (2014)
Shah, J., Breazeal, C.: An empirical analysis of team coordination behaviors and action planning with application to human–robot teaming. Hum. Factors J. Hum. Factors Ergon. Soc. 52(2), 234–245 (2010)
Cannon-Bowers, J.A., Bowers, C.A., Sanchez, A.: Using synthetic learning environments to train teams. Work group learning: Understanding, improving and assessing how groups learn in organizations, pp. 315–346 (2008)
Cannon-Bowers, J.A., Salas, E.: Reflections on shared cognition. J. Organ. Behav. 22(2), 195–202 (2001)
Lee, J.D., See, K.A.: Trust in automation: designing for appropriate reliance. Hum. Factors J. Hum. Factors Ergon. Soc. 46(1), 50–80 (2004)
Lyons, J.B., Havig, P.R.: Transparency in a human-machine context: approaches for fostering shared awareness/intent. In: Shumaker, R., Lackey, S. (eds.) VAMR 2014, Part I. LNCS, vol. 8525, pp. 181–190. Springer, Heidelberg (2014)
Chen, J.Y., et al.: Situation Awareness-Based Agent Transparency. Army Research Laboratory (ARL): ARL-TR-6905, Aberdeen Proving Grounds, MD (2014)
Lyons, J.B.: Being transparent about transparency: a model for human-robot interaction. In: 2013 AAAI Spring Symposium Series (2013)
Cramer, H., et al.: The effects of transparency on trust in and acceptance of a content-based art recommender. User Model. User-Adap. Inter. 18(5), 455–496 (2008)
Rao, A.S., Georgeff, M.P.: BDI agents: from theory to practice. In: ICMAS (1995)
Hoffman, R.: An integrated model of macrocognitive work and trust in automation. In: AAAI Spring Symposium: Trust and Autonomous Systems (2013)
Adams, J.A.: Human-robot interaction design: understanding user needs and requirements. In: Proceedings of the Human Factors and Ergonomics Society Annual Meeting. SAGE Publications (2005)
Smith, K., Hancock, P.: Situation awareness is adaptive, externally directed consciousness. Hum. Factors J. Hum. Factors Ergon. Soc. 37(1), 137–148 (1995)
Miller, C.A.: Delegation and transparency: coordinating interactions so information exchange is no surprise. In: Shumaker, R., Lackey, S. (eds.) VAMR 2014, Part I. LNCS, vol. 8525, pp. 191–202. Springer, Heidelberg (2014)
Mercado, J.E., et al.: Effects of agent transparency on multi-robot management effectiveness. Army Research Laboratory (ARL): ARL-TR-7466, Aberdeen Proving Grounds, MD (2015)
Stubbs, K., Wettergreen, D., Hinds, P.H.: Autonomy and common ground in human-robot interaction: a field study. IEEE Intell. Syst. 22(2), 42–50 (2007)
Abich, J.: Investigating the universality and comprehensive ability of measures to assess the state of workload. Doctoral Dissertation, University of Central Florida (2013)
Hart, S.G.: NASA-task load index (NASA-TLX); 20 years later. In: Proceedings of the Human Factors and Ergonomics Society Annual Meeting. Sage Publications (2006)
Hancock, P., Warm, J.: A dynamic model of stress and sustained attention. Hum. Factors 31(5), 519–537 (1989)
Mercado, J.E., et al.: Intelligent agent transparency in human-agent teaming for multi-UxV management. Human Factors. In Press
Wright, J., et al.: Agent Reasoning Transparency’s Effect on Operator Workload. Manuscript Submitted for Publication (2016)
Mathieu, J., et al.: Team effectiveness 1997-2007: a review of recent advancements and a glimpse into the future. J. Manage. 34(3), 410–476 (2008)
Salas, E., et al.: Does team training work? principles for health care. Acad. Emerg. Med. 15(11), 1002–1009 (2008)
Kim, T., Hinds, P.: Who should I blame? effects of autonomy and transparency on attributions in human-robot interaction. In: The 15th IEEE International Symposium on Robot and Human Interactive Communication, 2006. IEEE (2006)
Vitense, H.S., Jacko, J.A., Emery, V.K.: Multimodal feedback: an assessment of performance and mental workload. Ergonomics 46(1–3), 68–87 (2003)
Wickens, C.D.: Multiple resources and mental workload. Hum. Factors J. Hum. Factors Ergon. Soc. 50(3), 449–455 (2008)
Partan, S.R., Marler, P.: Issues in the classification of multimodal communication signals. Am. Nat. 166(2), 231–245 (2005)
Moreno, R., Mayer, R.: Interactive multimodal learning environments. Educ. Psychol. Rev. 19(3), 309–326 (2007)
Merkt, M., et al.: Learning with videos vs. learning with print: The role of interactive features. Learn. Instruction 21(6), 687–704 (2011)
Parasuraman, R., Miller, C.A.: Trust and etiquette in high-criticality automated systems. Commun. ACM 47(4), 51–55 (2004)
Woo, H.L.: Designing multimedia learning environments using animated pedagogical agents: factors and issues. J. Comput. Assist. Learn. 25(3), 203–218 (2009)
Beskow, J.: Animation of talking agents. In: Audio-Visual Speech Processing: Computational & Cognitive Science Approaches (1997)
Mayer, R.E.: The promise of multimedia learning: using the same instructional design methods across different media. Learn. Instruction 13(2), 125–139 (2003)
Krämer, N.C.: Psychological research on embodied conversational agents: the case of pedagogical agents. J. Media Psychol. 22, 47–51 (2010)
Moreno, R., Reislein, M., Ozogul, G.: Using virtual peers to guide visual attention during learning. J. Media Psychol. 22(2), 52–60 (2010)
Salem, M., et al.: Generation and evaluation of communicative robot gesture. Int. J. Soc. Rob. 4(2), 201–217 (2012)
Ososky, S., et al.: Building appropriate trust in human-robot teams. In: 2013 AAAI Spring Symposium Series (2013)
Perzanowski, D., et al.: Building a multimodal human-robot interface. IEEE Intell. Syst. 16(1), 16–21 (2001)
Baraka, K., Paiva, A., Veloso, M.: Expressive lights for revealing mobile service robot state. In: Reis, L.P., Moreira, A.P., Lima, P.U., Montano, L., Muñoz-Martinez, V. (eds.) Robot 2015: Second Iberian Robotics Conference. Advances in Intelligent Systems and Computing, vol. 417, pp. 107–119. Springer, Heidelberg (2016)
Sims, V.K., et al.: Robots’ auditory cues are subject to anthropomorphism. In: Proceedings of the Human Factors and Ergonomics Society Annual Meeting. SAGE Publications (2009)
Gong, L., Lai, J.: Shall we mix synthetic speech and human speech?: impact on users’ performance, perception, and attitude. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. ACM (2001)
Delogu, C., Paoloni, A., Pocci, P.: New directions in the evaluation of voice input/output systems. IEEE J. Sel. Areas Commun. 9(4), 566–573 (1991)
Reynolds, M.E., Isaacs-Duvall, C., Haddox, M.L.: A comparison of learning curves in natural and synthesized speech comprehension. J. Speech Lang. Hear. Res. 45(4), 802–810 (2002)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing Switzerland
About this paper
Cite this paper
Lakhmani, S., Abich, J., Barber, D., Chen, J. (2016). A Proposed Approach for Determining the Influence of Multimodal Robot-of-Human Transparency Information on Human-Agent Teams. In: Schmorrow, D., Fidopiastis, C. (eds) Foundations of Augmented Cognition: Neuroergonomics and Operational Neuroscience. AC 2016. Lecture Notes in Computer Science(), vol 9744. Springer, Cham. https://doi.org/10.1007/978-3-319-39952-2_29
Download citation
DOI: https://doi.org/10.1007/978-3-319-39952-2_29
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-39951-5
Online ISBN: 978-3-319-39952-2
eBook Packages: Computer ScienceComputer Science (R0)