Abstract
Automatic short answer scoring (ASAS) has received considerable attention in the field of education. However, existing methods typically treat ASAS as a standard text classification problem, following conventional pre-training or fine-tuning procedures. These approaches often generate embedding spaces that lack clear boundaries, resulting in overlapping representations for answers of different scores. To address this issue, we introduce a novel metric learning (MeL)-based pre-training method for answer representation optimization. This strategy encourages the clustering of similar representations while pushing dissimilar ones apart, thereby facilitating the formation of a more coherent same-score and distinct different-score answer embedding space. To fully exploit the potential of MeL, we define two types of answer similarities based on scores and rubrics, providing accurate supervised signals for improved training. Extensive experiments on thirteen short answer questions show that our method, even when paired with a simple linear model for downstream scoring, significantly outperforms prior ASAS methods in both scoring accuracy and efficiency.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Alikaniotis, D., Yannakoudakis, H., Rei, M.: Automatic text scoring using neural networks. In: ACL, pp. 715–725 (2016)
Chen, H., He, B.: Automated essay scoring by maximizing human-machine agreement. In: EMNLP 2013, pp. 1741–1752 (2013)
Condor, A., Litster, M., Pardos, Z.: Automatic short answer grading with SBERT on out-of-sample questions. In: EDM, pp. 345–352 (2021)
Condor, A., Pardos, Z., Linn, M.: Representing scoring rubrics as graphs for automatic short answer grading. In: AIED, pp. 354–365 (2022)
Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. arXiv:1810.04805 (2018)
Dong, F., Zhang, Y.: Automatic features for essay scoring-an empirical study. In: EMNLP 2016, pp. 1072–1077 (2016)
Dong, F., Zhang, Y., Yang, J.: Attention-based recurrent convolutional neural network for automatic essay scoring. In: CoNLL, pp. 153–162 (2017)
Hoffer, E., Ailon, N.: Deep metric learning using triplet network. In: International Workshop on Similarity-Based Pattern Recognition, pp. 84–92 (2015)
Khosla, P., et al.: Supervised contrastive learning. arXiv:2004.11362 (2020)
Larkey, L.S.: Automatic essay grading using text categorization techniques. In: ACM SIGIR, pp. 90–95 (1998)
Li, X., Yang, H., Hu, S., Geng, J., Lin, K., Li, Y.: Enhanced hybrid neural network for automated essay scoring. Expert. Syst. 39(10), e13068 (2022)
Lun, J., Zhu, J., Tang, Y., Yang, M.: Multiple data augmentation strategies for improving performance on automatic short answer scoring. In: AAAI, pp. 13389–13396 (2020)
Luo, D., Su, J., Yu, S.: A BERT-based approach with relation-aware attention for knowledge base question answering. In: IJCNN, pp. 1–8. IEEE (2020)
Mayfield, E., Black, A.W.: Should you fine-tune BERT for automated essay scoring? In: Proceedings of the Fifteenth Workshop on Innovative Use of NLP for Building Educational Applications, pp. 151–162 (2020)
Oord, A.V.D., Li, Y., Vinyals, O.: Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748 (2018)
Phandi, P., Chai, K.M.A., Ng, H.T.: Flexible domain adaptation for automated essay scoring using correlated linear regression. In: EMNLP, pp. 431–439 (2015)
Reimers, N., Gurevych, I.: Sentence-BERT: sentence embeddings using Siamese BERT-networks. arXiv:1908.10084 (2019)
Sung, C., Dhamecha, T., Saha, S., Ma, T., Reddy, V., Arora, R.: Pre-training BERT on domain resources for short answer grading. In: EMNLP 2019, pp. 6071–6075 (2019)
Viji, D., Revathy, S.: A hybrid approach of weighted fine-tuned BERT extraction with deep Siamese Bi-LSTM model for semantic text similarity identification. Multimedia Tools Appl. 81(5), 6131–6157 (2022)
Wang, T., Funayama, H., Ouchi, H., Inui, K.: Data augmentation by rubrics for short answer grading. J. Nat. Lang. Process. 28, 183–205 (2021)
Wang, Z., Lan, A.S., Waters, A.E., Grimaldi, P., Baraniuk, R.G.: A meta-learning augmented bidirectional transformer model for automatic short answer grading. In: EDM, pp. 667–670 (2019)
Weinberger, K.Q., Saul, L.K.: Distance metric learning for large margin nearest neighbor classification. J. Mach. Learn. Res. 10(2), 207–244 (2009)
Yang, R., Cao, J., Wen, Z., Wu, Y., He, X.: Enhancing automated essay scoring performance via fine-tuning pre-trained language models with combination of regression and ranking. In: Findings of EMNLP 2020, pp. 1560–1569 (2020)
Zhu, X., Wu, H., Zhang, L.: Automatic short-answer grading via BERT-based deep neural networks. IEEE Trans. Learn. Technol. 15(3), 364–375 (2022)
Acknowledgment
This work was supported in part by JST SPRING No. JPMJSP2136 and JSPS KAKENHI No. JP21H00907 and JP23H03511.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Wang, B., Dawton, B., Ishioka, T., Mine, T. (2024). Optimizing Answer Representation Using Metric Learning for Efficient Short Answer Scoring. In: Liu, F., Sadanandan, A.A., Pham, D.N., Mursanto, P., Lukose, D. (eds) PRICAI 2023: Trends in Artificial Intelligence. PRICAI 2023. Lecture Notes in Computer Science(), vol 14326. Springer, Singapore. https://doi.org/10.1007/978-981-99-7022-3_21
Download citation
DOI: https://doi.org/10.1007/978-981-99-7022-3_21
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-7021-6
Online ISBN: 978-981-99-7022-3
eBook Packages: Computer ScienceComputer Science (R0)