Skip to main content
Log in

Leveraging mobile eye-trackers to capture joint visual attention in co-located collaborative learning groups

  • Published:
International Journal of Computer-Supported Collaborative Learning Aims and scope Submit manuscript

Abstract

This paper describes a promising methodology for studying co-located groups: mobile eye-trackers. We provide a comprehensive description of our data collection and analysis processes so that other researchers can take advantage of this cutting-edge technology. Data were collected in a controlled experiment where 27 student dyads (N = 54) interacted with a Tangible User Interface. They first had to define some design principles for optimizing a warehouse layout by analyzing a set of Contrasting Cases, and build a small-scale layout based on those principles. The contributions of this paper are that: 1) we replicated prior research showing that levels of Joint Visual Attention (JVA) are correlated with collaboration quality across all groups; 2) we then qualitatively analyzed two dyads with high levels of JVA and show that it can hide a free-rider effect (Salomon and Globerson 1989); 3) in conducting this analysis, we additionally developed a new visualization (augmented cross-recurrence graphs) that allows researchers to distinguish between high JVA groups that have balanced and unbalanced levels of participations; 4) finally, we generalized this effect to the entire sample and found a significant negative correlation between dyads’ learning gains and unbalanced levels of participation (as computed from the eye-tracking data). We conclude by discussing implications for automatically analyzing students’ interactions using dual eye-trackers.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1.
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10

Similar content being viewed by others

References

  • Baker, M., Hansen, T., Joiner, R., & Traum, D. (1999). The role of grounding in collaborative learning tasks. In P. Dillenbourg (Ed.), Collaborative learning: Cognitive and computational approaches (pp. 31–63). Amsterdam: Elsevier.

  • Barron, B. (2000). Achieving coordination in collaborative problem-solving groups. The Journal of the Learning Sciences, 9(4), 403–436.

    Article  Google Scholar 

  • Barron, B., & Roschelle, J. (2009). Shared cognition. In E. Anderman (Ed.), Psychology of classroom learning: An encyclopedia (pp. 819–823). Detroit: Macmillan Reference USA.

    Google Scholar 

  • Bonnard, Q., Lemaignan, S., Zufferey, G., Mazzei, A., Cuendet, S., Li, N., … Dillenbourg, P. (2013). Chilitags 2: Robust fiducial markers for augmented reality and robotics. CHILI, EPFL, Switzerland. Retrieved from http://chili.epfl.ch/software.

  • Bransford, J. D., & Schwartz, D. L. (1999). Rethinking transfer: A simple proposal with multiple implications. Review of Research in Education, 24, 61–100.

    Google Scholar 

  • Clark, H. H. (1985). Language use and language users. In G. Lindzey & E. Aronson (Eds.), Handbook of social psychology (3rd ed., pp. 179–231). New York: Harper and Row.

    Google Scholar 

  • Clark, H. H., & Wilkes-Gibbs, D. (1986). Referring as a collaborative process. Cognition, 22(1), 1–39.

    Article  Google Scholar 

  • D’Angelo, S., & Begel, A. (2017, May). Improving communication between pair programmers using shared gaze awareness. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems (pp. 6245–6290). New York: ACM.

  • Dillenbourg, P., Lemaignan, S., Sangin, M., Nova, N., & Molinari, G. (2016). The symmetry of partner modelling. International Journal of Computer-Supported Collaborative Learning, 11(2), 227–253.

    Article  Google Scholar 

  • Gergle, D., & Clark, A. T. (2011). See what i'm saying?: Using dyadic mobile eye tracking to study collaborative reference. In Proceedings of the ACM 2011 conference on Computer supported cooperative work (pp. 435–444). ACM.

  • Hayes, A. F., & Krippendorff, K. (2007). Answering the call for a standard reliability measure for coding data. Communication methods and measures, 1(1), 77–89.

    Article  Google Scholar 

  • Jermann, P., Mullins, D., Nüssli, M.-A., & Dillenbourg, P. (2011). Collaborative gaze footprints: Correlates of interaction quality. CSCL2011 Conference Proceedings. Volume I - Long Papers, 184–191.

  • Mason, L., Pluchino, P., & Tornatora, M. C. (2015). Eye-movement modeling of text and picture integration during reading: Effects on processing and learning. Contemporary Educational Psychology, 14, 172–187.

    Article  Google Scholar 

  • Meier, A., Spada, H., & Rummel, N. (2007). A rating scheme for assessing the quality of computer-supported collaboration processes. International Journal of Computer-Supported Collaborative Learning, 2(1), 63–86.

    Article  Google Scholar 

  • Mundy, P., Sigman, M., & Kasari, C. (1990). A longitudinal study of joint attention and language development in autistic children. Journal of Autism and Developmental Disorders, 20(1), 115–128.

    Article  Google Scholar 

  • Papavlasopoulou, S., Sharma, K., Giannakos, M., & Jaccheri, L. (2017). Using eye-tracking to unveil differences between kids and teens in coding activities. In Proceedings of the 2017 Conference on Interaction Design and Children (pp. 171–181). ACM.

  • Prieto, L. P., Wen, Y., Caballero, D., Sharma, K., & Dillenbourg, P. (2014). Studying teacher cognitive load in multi-tabletop classrooms using mobile eye-tracking. In Proceedings of the Ninth ACM International Conference on Interactive Tabletops and Surfaces (pp. 339–344). ACM.

  • Richardson, D. C., & Dale, R. (2005). Looking to understand: The coupling between speakers’ and listeners’ eye movements and its relationship to discourse comprehension. Cognitive Science, 29(6), 1045–1060.

    Article  Google Scholar 

  • Roschelle, J. (1992). Learning by collaborating: Convergent conceptual change. The Journal of the Learning Sciences, 2(3), 235–276.

    Article  Google Scholar 

  • Salomon, G., & Globerson, T. (1989). When teams do not function the way they ought to. International Journal of Educational Research, 13(1), 89–99.

    Article  Google Scholar 

  • Schlösser, C., Schlieker-Steens, P., Kienle, A., & Harrer, A. (2015). Using real-time gaze based awareness methods to enhance collaboration. In CYTED-RITOS International Workshop on Groupware (pp. 19–27). Cham: Springer.

    Google Scholar 

  • Schneider, B., & Pea, R. (2013). Real-time mutual gaze perception enhances collaborative learning and collaboration quality. International Journal of Computer-Supported Collaborative Learning, 8(4), 375–397.

    Article  Google Scholar 

  • Schneider, B., Sharma, K., Cuendet, S., Zufferey, G., Dillenbourg, P., & Pea, R. (2016). Unpacking the perceptual benefits of a tangible Interface. ACM Transactions on Computer-Human Interactions (TOCHI), 23(6), 39.

    Google Scholar 

  • Schwartz, D. L. (1995). The emergence of abstract representations in dyad problem solving. The Journal of the Learning Sciences, 4(3), 321–354.

    Article  Google Scholar 

  • Stahl, G. (2007). Meaning making in CSCL: Conditions and preconditions for cognitive processes by groups. In Proceedings of the 8th international conference on Computer Supported Collaborative Learning (pp. 652–661).

  • Tomasello, M. (1995). Joint attention as social cognition. In C. Moore & P. J. Dunham (Eds.), Joint attention: Its origins and role in development (pp. 103–130). Hillsdale: Lawrence Erlbaum.

    Google Scholar 

  • Zufferey, G., Jermann, P., Do-Lenh, S., & Dillenbourg, P. (2009, September). Using augmentations as bridges from concrete to abstract representations. In Proceedings of the 23rd British HCI Group Annual Conference on People and Computers: Celebrating People and Technology (pp. 130–139). British Computer Society.

Download references

Acknowledgments

We gratefully acknowledge support from the National Science Foundation for this work from the LIFE Center (NSF #0835854) as well as the Leading House Technologies for Vocation Education, funded by the Swiss State Secretariat for Education, Research and Innovation. Finally, we would like to thank SMI (SensoMotoric Instruments) for their eye-tracking technology and Jacques Kurzo for his deep involvement in this project.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Bertrand Schneider.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Schneider, B., Sharma, K., Cuendet, S. et al. Leveraging mobile eye-trackers to capture joint visual attention in co-located collaborative learning groups. Intern. J. Comput.-Support. Collab. Learn 13, 241–261 (2018). https://doi.org/10.1007/s11412-018-9281-2

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11412-018-9281-2

Keywords

Navigation