skip to main content
10.1145/2498328.2500061acmconferencesArticle/Chapter ViewAbstractPublication Pagesacm-seConference Proceedingsconference-collections
research-article

Using artificial neural networks to predict first-year traditional students second year retention rates

Published:04 April 2013Publication History

ABSTRACT

This research investigates the use of Artificial Neural Networks (ANNs) to predict first year student retention rates. Based on a significant body of previous research, this work expands on previous attempts to predict student outcomes using machine-learning techniques. Using a large data set provided by Columbus State University's Information Technology department, ANNs were used to analyze incoming first-year traditional freshmen students' data over a period from 2005--2011. Using several different network designs, the students' data was analyzed, and a basic predictive network was devised. While the overall accuracy was high when including the first and second semesters worth of data, once the data set was reduced to a single semester, the overall accuracy dropped significantly. Using different network designs, more complex learning algorithms, and better training strategies, the prediction accuracy rate for a student's return to the second year approached 75% overall. Since the rate is still low, there is room for improvements, and several techniques that might increase the reliability of these networks are discussed.

References

  1. Barker, K. et al. 2004. Learning From Student Data. Proceedings of the 2004 IEEE Systems and Information Engineering Design Symposium (2004), 79--86.Google ScholarGoogle Scholar
  2. Bogard, M. et al. 2011. A Comparison of Empirical Models for Predicting Student Retention. White Paper. Office of Institutional Research, Western Kentucky University.Google ScholarGoogle Scholar
  3. Delen, D. 2010. A comparative analysis of machine learning techniques for student retention management. Decision Support Systems. 49, 4 (Nov. 2010), 498--506. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Hagen, M. T. and Menhaj, M.-B.B. 1994. Training feedforward networks with the Marquardt algorithm. IEEE Transactions on Neural Networks. 5, 6 (1994), 989--993. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Herzog, S. 2006. Estimating student retention and degree-completion time: Decision trees and neural networks vis-à-vis regression. New Directions for Institutional Research. 2006, 131 (2006), 17--33.Google ScholarGoogle ScholarCross RefCross Ref
  6. Imbrie, P. K. et al. 2008. Artificial Intelligence Methods to Forecast Engineering Students' Retention Based on Cognitive and Non-cognitive Factors. 2008 Annual Conference & Exposition of the American Society for Engineering Education (Pittsburg, PA, 2008).Google ScholarGoogle Scholar
  7. Karamouzis, S. T. and Vrettos, A. 2008. An Artificial Neural Network for Predicting Student Graduation Outcomes. Proceedings of the World Congress on Engineering and Computer Science (San Francisco, 2008), 22--25.Google ScholarGoogle Scholar
  8. The Condition of Education - Postsecondary Education - Completions - Postsecondary Graduation Rates: 2012. http://nces.ed.gov/programs/coe/indicator_pgr.asp.Google ScholarGoogle Scholar
  9. Wongkhamdi, T. and Seresangtakul, P. 2010. A Comparison of Classical Discriminant Analysis and Artificial Neural Networks in Predicting Student Graduation Outcomes. Proceedings of the Second International Conference on Knowledge and Smart Technologies (Chonburi, 2010), 29--34.Google ScholarGoogle Scholar

Index Terms

  1. Using artificial neural networks to predict first-year traditional students second year retention rates

    Recommendations

    Reviews

    Jonathan P. E. Hodgson

    This paper reports the results of a project using neural networks and data supplied by the IT department of Columbus State University in Georgia to predict which students will drop out after their first year of college. The author began with two-layer networks, that is, networks without any hidden layers. When data from the first two semesters were provided, these networks were able to predict whether a given student would drop out with an accuracy of 75 percent. However, accuracy fell to 10 percent when the second semester data was not used. To get better results without second semester data, the author experimented with networks with a single hidden layer. Using a feed-forward cascade three-layer network increased accuracy to 69 percent. (The confusion matrix for the most accurate of these networks is not reproduced in the paper, even though the author says it is. Rather, the matrix for a less accurate network is given. However, the results are described in the body of the paper.) The neural network was then used to extract weights for the input data categories, thus removing data that was not found relevant to the outcome. The items that were assigned the highest weight included number of fall semester courses, fall grade point average, selected minor, aggregate score on the Columbus State entrance test, and highest education level attained by both parents. Although the author does not draw any conclusions as to how an advisor might make use of the results, readers can draw their own conclusions. It would be interesting to compare the results with those obtained from a decision tree generating system. Online Computing Reviews Service

    Access critical reviews of Computing literature here

    Become a reviewer for Computing Reviews.

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      ACMSE '13: Proceedings of the 51st ACM Southeast Conference
      April 2013
      224 pages
      ISBN:9781450319010
      DOI:10.1145/2498328
      • General Chair:
      • Ashraf Saad

      Copyright © 2013 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 4 April 2013

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article

      Acceptance Rates

      Overall Acceptance Rate178of377submissions,47%

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader