Skip to main content

Reducing the Error Mapping of the Students’ Performance Using Feature Selection

  • Conference paper
  • First Online:
Proceedings of the 12th International Conference on Soft Computing and Pattern Recognition (SoCPaR 2020) (SoCPaR 2020)

Abstract

In an educational environment, classifying the cognitive aspect of students is critical. It is because an accurate classification is needed by a lecturer to take the right decision for enhancing a better educational environment. To the best of our knowledge, there is no previous research that focuses on this classification process. In this paper, we propose discretization and feature selection methods before the classification. For this purpose, we adopt the equal frequency for the discretization whose result is evaluated by using logistic regression with two regularizations: lasso and ridge. The experimental result shows that four-intervals on the ridge achieve the highest accuracy. It is to be the base to determine the level of the student’s performance: excellent, good, fair, and poor. Next, we remove unnecessary features, by using the Gain Ratio and Gini Index. Also, we build classifiers to evaluate our proposed methods by using k-Nearest Neighbors (k-NN), Neural Network (NN), and CN2 Rule Induction. The experimental result indicates that both discretization and feature selection can enhance the performance of the classification process. Concerning the accuracy level, there is an increase of about 35%, 2.14%, and 3.8% on average of k-NN, NN, and CN2 Rule Induction respectively, from those with original features.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Wanarti, P., Ismayanti, E., Peni, H., Yamasari, Y.: The enhancement of teaching-learning process effectiveness through the development of instructional media based on e-learning of Surabaya’s vocational student. In: Proceedings of the 6th International Conference on Educational, Management, Administration and Leadership, pp. 342–346 (2016). https://doi.org/10.2991/icemal-16.2016.71

  2. Tan, P.-N., Steinbach, M., Kumar, V.: Introduction to Data Mining. Pearson Addison Wesley, Boston (2005)

    Google Scholar 

  3. Dillenbourg, P.: The evolution of research on digital education. Int. J. Artif. Intell. Educ. 26(2), 544–560 (2016). https://doi.org/10.1007/s40593-016-0106-z

    Article  Google Scholar 

  4. Liñán, L.C., Pérez, Á.A.J.: Educational data mining and learning analytics: differences, similarities, and time evolution. RUSC. Univ. Knowl. Soc. J. 12(3), 98 (2015). https://doi.org/10.7238/rusc.v12i3.2515

    Article  Google Scholar 

  5. Koza, J.R., Bennett, F.H., Andre, D., Keane, M.A.: Automated design of both the topology and sizing of analog electrical circuits using genetic programming. In: Gero, J.S., Sudweeks, F. (eds.) Artificial Intelligence in Design 1996, pp. 151–170. Springer, Dordrecht (1996). https://doi.org/10.1007/978-94-009-0279-4_9

    Chapter  Google Scholar 

  6. Samuel, A.L.: Some studies in machine learning using the game of checkers. IBM J. Res. Dev. 3(3), 210–229 (1959). https://doi.org/10.1147/rd.33.0210

    Article  MathSciNet  Google Scholar 

  7. Troussas, C., Virvou, M., Mesaretzidis, S.: Comparative analysis of algorithms for student characteristics classification using a methodological framework (2015)

    Google Scholar 

  8. Sukajaya, N., Purnama, K.E., Purnomo, M.H.: Intelligent classification of learner’s cognitive domain using Bayes Net, Naïve Bayes, and J48 utilizing bloom’s taxonomy-based serious game. Int. J. Emerg. Technol. Learn. 10(2), 46–52 (2015). https://doi.org/10.3991/ijet.v10i1.4451

    Article  Google Scholar 

  9. Gobert, J.D., Kim, Y.J., Sao, M.A., Pedro, M.K., Betts, C.G.: Using educational data mining to assess students’ skills at designing and conducting experiments within a complex systems microworld. Think. Skills Creativity 18, 81–90 (2015). https://doi.org/10.1016/j.tsc.2015.04.008

    Article  Google Scholar 

  10. Ko, C.-Y., Leu, F.-Y.: Applying data mining to explore students’ self-regulation in learning contexts. In: 2016 IEEE 30th International Conference on Advanced Information Networking and Applications (AINA), March 2016, pp. 74–78 (2016). https://doi.org/10.1109/AINA.2016.123.

  11. Casey, K., Azcona, D.: Utilizing student activity patterns to predict performance. Int. J. Educ. Technol. High. Educ. 14(1), 4 (2017). https://doi.org/10.1186/s41239-017-0044-3

    Article  Google Scholar 

  12. Ramaswami, M., Bhaskaran, R.: A study on feature selection techniques in educational data mining. J. Comput. 1(1), 2151–9617 (2009). Accessed 16 Aug 2017. https://pdfs.semanticscholar.org/d11c/46515632f3e462d1a952e67fd4657a5f009e.pdf

  13. Punlumjeak, W., Rachburee, N.: A comparative study of feature selection techniques for classify student performance. In: 2015 7th International Conference on Information Technology and Electrical Engineering (ICITEE), October 2015, pp. 425–429 (2015). https://doi.org/10.1109/ICITEED.2015.7408984

  14. Sasi Regha, R., Uma Rani, R.: Optimization feature selection for classifying student in educational data mining. Int. J. Innov. Eng. Technol. 490(4) (2016). https://ijiet.com/wp-content/uploads/2017/01/65.pdf. Accessed 16 Aug 2017

  15. Han, J., Kamber, M., Pei, J.: Data Mining Concepts and Techniques. Elsevier, USA (2012)

    Google Scholar 

  16. Yamasari, Y., Nugroho, S.M.S., Sukajaya, I.N., Purnomo, M.H.: Features extraction to improve performance of clustering process on student achievement. In: 2016 International Computer Science and Engineering Conference (ICSEC), December 2016, pp. 1–5 (2016). https://doi.org/10.1109/ICSEC.2016.7859946

  17. Quinlan, J.R.: Induction of decision trees. Mach. Learn. 1(1), 81–106 (1986). https://doi.org/10.1023/A:1022643204877

    Article  Google Scholar 

  18. Altman, N.S.: An introduction to Kernel and nearest-neighbor nonparametric regression. Am. Stat. 46(3), 175–185 (1992). https://doi.org/10.1080/00031305.1992.10475879

    Article  MathSciNet  Google Scholar 

  19. Rumelhart, D.E., Hinton, G.E., Williams, R.J.: Learning representations by back-propagating errors. Nature 323 (1986). https://www.iro.umontreal.ca/~pift6266/A06/refs/backprop_old.pdf

  20. LeCun, Y., Bottou, L., Orr, G.B., Muller, K.-R.: Efficient BackProp. Springer, New York (1998)

    Book  Google Scholar 

  21. Xu, W.: Towards optimal one pass large scale learning with averaged stochastic gradient descent, July 2011. https://arxiv.org/abs/1107.2490. Accessed 4 April 2018

  22. Kingma, D.P., Lei Ba, J.: ADAM: a method for stochastic optimization (2015)

    Google Scholar 

  23. Clark, P., Niblett, T.: The CN2 induction algorithm. Mach. Learn. 3(4), 261–283 (1989). https://doi.org/10.1023/A:1022641700528

    Article  Google Scholar 

  24. Clark, P., Boswell, R.: ‘Rule induction with CN2: some recent improvements. In: Clark, P., Boswell, R. (eds.) Machine Learning - Proceedings of the 5th European Conference, EWSL-91, pp. 151–163. Springer, Heidelberg (1991)https://www.cs.utexas.edu/users/pclark/papers/newcn.pdf

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yuni Yamasari .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2021 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Yamasari, Y., Rochmawati, N., Qoiriah, A., Suyatno, D.F., Ahmad, T. (2021). Reducing the Error Mapping of the Students’ Performance Using Feature Selection. In: Abraham, A., et al. Proceedings of the 12th International Conference on Soft Computing and Pattern Recognition (SoCPaR 2020). SoCPaR 2020. Advances in Intelligent Systems and Computing, vol 1383. Springer, Cham. https://doi.org/10.1007/978-3-030-73689-7_18

Download citation

Publish with us

Policies and ethics