skip to main content
10.1145/3636555.3636931acmotherconferencesArticle/Chapter ViewAbstractPublication PageslakConference Proceedingsconference-collections
short-paper

Navigating (Dis)agreement: AI Assistance to Uncover Peer Feedback Discrepancies

Published:18 March 2024Publication History

ABSTRACT

Engaging students in the peer review process has been recognized as a valuable educational tool. It not only nurtures a collaborative learning environment where reviewees receive timely and rich feedback but also enhances the reviewer’s critical thinking skills and encourages reflective self-evaluation. However, a common concern arises when students encounter misaligned or conflicting feedback. Not only can such feedback confuse students; but it can also make it difficult for the instructor to rely on the reviews when assigning a score to the work. Addressing this pressing issue, our paper introduces an innovative, AI-assisted approach that is designed to detect and highlight disagreements within formative feedback. We’ve harnessed extensive data from 170 students, analyzing 15,500 instances of peer feedback from a software development course. By utilizing clustering techniques coupled with sophisticated natural language processing (NLP) models, we transform feedback into distinct feature vectors to pinpoint disagreements. The findings from our study underscore the effectiveness of our approach in enhancing text representations to significantly boost the capability of clustering algorithms in discerning disagreements in feedback. These insights bear implications for educators and software development courses, offering a promising route to streamline and refine the peer review process for the betterment of student learning outcomes.

References

  1. Ali Darvishi, Hassan Khosravi, Solmaz Abdi, Shazia Sadiq, and Dragan Gašević. 2022. Incorporating training, self-monitoring and AI-assistance to improve peer feedback quality. In Proceedings of the Ninth ACM Conference on Learning@ Scale. 35–47.Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Ali Darvishi, Hassan Khosravi, Afshin Rahimi, Shazia Sadiq, and Dragan Gašević. 2022. Assessing the Quality of Student-Generated Content at Scale: A Comparative Analysis of Peer-Review Models. IEEE Transactions on Learning Technologies 16, 1 (2022), 106–120.Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Ali Darvishi, Hassan Khosravi, and Shazia Sadiq. 2021. Employing peer review to evaluate the quality of student generated content at scale: A trust propagation approach. In Proceedings of the eighth ACM conference on learning@ scale. 139–150.Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Ali Darvishi, Hassan Khosravi, Shazia Sadiq, and Dragan Gašević. 2022. Incorporating AI and learning analytics to build trustworthy peer assessment systems. British Journal of Educational Technology 53, 4 (2022), 844–875.Google ScholarGoogle ScholarCross RefCross Ref
  5. Michel Galley, Kathleen McKeown, Julia Bell Hirschberg, and Elizabeth Shriberg. 2004. Identifying agreement and disagreement in conversational speech: Use of bayesian networks to model pragmatic dependencies. (2004).Google ScholarGoogle Scholar
  6. Edward F Gehringer. 2014. A survey of methods for improving review quality. In International Conference on Web-Based Learning. Springer, 92–97.Google ScholarGoogle ScholarCross RefCross Ref
  7. Renchu Guan, Hao Zhang, Yanchun Liang, Fausto Giunchiglia, Lan Huang, and Xiaoyue Feng. 2020. Deep feature-based text clustering and its explanation. IEEE Transactions on Knowledge and Data Engineering (2020).Google ScholarGoogle ScholarCross RefCross Ref
  8. Dustin Hillard, Mari Ostendorf, and Elizabeth Shriberg. 2003. Detection of agreement vs. disagreement in meetings: Training with unlabeled data. In Companion Volume of the Proceedings of HLT-NAACL 2003-Short Papers. 34–36.Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Sushant Hiray and Venkatesh Duppada. 2017. Agree to disagree: Improving disagreement detection with dual GRUs. In 2017 Seventh International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW). IEEE, 147–152.Google ScholarGoogle ScholarCross RefCross Ref
  10. Adam Janin, Don Baron, Jane Edwards, Dan Ellis, David Gelbart, Nelson Morgan, Barbara Peskin, Thilo Pfau, Elizabeth Shriberg, Andreas Stolcke, 2003. The ICSI meeting corpus. In 2003 IEEE International Conference on Acoustics, Speech, and Signal Processing, 2003. Proceedings.(ICASSP’03)., Vol. 1. IEEE, I–I.Google ScholarGoogle Scholar
  11. Supakpong Jinarat, Bundit Manaskasemsak, and Arnon Rungsawang. 2018. Short text clustering based on word semantic graph with word embedding model. In 2018 Joint 10th International Conference on Soft Computing and Intelligent Systems (SCIS) and 19th International Symposium on Advanced Intelligent Systems (ISIS). IEEE, 1427–1432.Google ScholarGoogle ScholarCross RefCross Ref
  12. Hassan Khosravi, Simon Buckingham Shum, Guanliang Chen, Cristina Conati, Yi-Shan Tsai, Judy Kay, Simon Knight, Roberto Martinez-Maldonado, Shazia Sadiq, and Dragan Gašević. 2022. Explainable artificial intelligence in education. Computers and Education: Artificial Intelligence 3 (2022), 100074.Google ScholarGoogle ScholarCross RefCross Ref
  13. Klaus Krippendorff. 2018. Content analysis: An introduction to its methodology. Sage publications.Google ScholarGoogle Scholar
  14. Heng Luo, Anthony Robinson, and Jae-Young Park. 2014. Peer grading in a MOOC: Reliability, validity, and perceived effects. Online Learning Journal 18, 2 (2014).Google ScholarGoogle Scholar
  15. Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. 2013. Distributed representations of words and phrases and their compositionality. Advances in neural information processing systems 26 (2013).Google ScholarGoogle Scholar
  16. Marie-Francine Moens, Erik Boiy, Raquel Mochales Palau, and Chris Reed. 2007. Automatic detection of arguments in legal texts. In Proceedings of the 11th international conference on Artificial intelligence and law. 225–230.Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Saif M Mohammad, Parinaz Sobhani, and Svetlana Kiritchenko. 2017. Stance and sentiment in tweets. ACM Transactions on Internet Technology (TOIT) 17, 3 (2017), 1–23.Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Raquel Mochales Palau and Marie-Francine Moens. 2009. Argumentation mining: the detection, classification and structure of arguments in text. In Proceedings of the 12th international conference on artificial intelligence and law. 98–107.Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Ernesto Panadero and Maryam Alqassab. 2019. An empirical review of anonymity effects in peer assessment, peer feedback, peer review, peer evaluation and peer grading. Assessment & Evaluation in Higher Education 44, 8 (2019), 1253–1278.Google ScholarGoogle ScholarCross RefCross Ref
  20. M Parvez Rashid, Edward F Gehringer, Mitchell Young, Divyang Doshi, Qinjin Jia, and Yunkai Xiao. 2021. Peer Assessment Rubric Analyzer: An NLP approach to analyzing rubric items for better peer-review. In 2021 19th International Conference on Information Technology Based Higher Education and Training (ITHET). IEEE, 1–9.Google ScholarGoogle Scholar
  21. M Parvez Rashid, Yunkai Xiao, and Edward F Gehringer. 2022. Going beyond" Good Job": Analyzing Helpful Feedback from the Student’s Perspective.International Educational Data Mining Society (2022).Google ScholarGoogle Scholar
  22. Nils Reimers and Iryna Gurevych. 2019. Sentence-bert: Sentence embeddings using siamese bert-networks. arXiv preprint arXiv:1908.10084 (2019).Google ScholarGoogle Scholar
  23. Sara Rosenthal and Kathleen McKeown. 2015. I couldn’t agree more: The role of conversational structure in agreement and disagreement detection in online discussions. In Proceedings of the 16th Annual Meeting of the Special Interest Group on Discourse and Dialogue. 168–177.Google ScholarGoogle ScholarCross RefCross Ref
  24. Burr Settles. 2012. Active learning. Synthesis lectures on artificial intelligence and machine learning 6, 1 (2012), 1–114.Google ScholarGoogle Scholar
  25. Nan Sun, Dongsheng Liu, Anding Zhu, Yahui Chen, and Yufei Yuan. 2019. Do Airbnb’s “Superhosts” deserve the badge? An empirical study from China. Asia Pacific Journal of Tourism Research 24, 4 (2019), 296–313.Google ScholarGoogle ScholarCross RefCross Ref
  26. Keith Topping. 2017. Peer assessment: Learning by judging and discussing the work of other learners. Interdisciplinary Education and Psychology 1, 1 (2017), 1–17.Google ScholarGoogle ScholarCross RefCross Ref
  27. Jie Yin, Nalin Narang, Paul Thomas, and Cecile Paris. 2012. Unifying local and global agreement and disagreement classification in online debates. In Proceedings of the 3rd Workshop in Computational Approaches to Subjectivity and Sentiment Analysis. 61–69.Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Navigating (Dis)agreement: AI Assistance to Uncover Peer Feedback Discrepancies
          Index terms have been assigned to the content through auto-classification.

          Recommendations

          Comments

          Login options

          Check if you have access through your login credentials or your institution to get full access on this article.

          Sign in
          • Published in

            cover image ACM Other conferences
            LAK '24: Proceedings of the 14th Learning Analytics and Knowledge Conference
            March 2024
            962 pages
            ISBN:9798400716188
            DOI:10.1145/3636555

            Copyright © 2024 ACM

            Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

            Publisher

            Association for Computing Machinery

            New York, NY, United States

            Publication History

            • Published: 18 March 2024

            Permissions

            Request permissions about this article.

            Request Permissions

            Check for updates

            Qualifiers

            • short-paper
            • Research
            • Refereed limited

            Acceptance Rates

            Overall Acceptance Rate236of782submissions,30%
          • Article Metrics

            • Downloads (Last 12 months)47
            • Downloads (Last 6 weeks)47

            Other Metrics

          PDF Format

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader

          HTML Format

          View this article in HTML Format .

          View HTML Format