Research Article
BibTex RIS Cite

Year 2025, Volume: 21 Issue: 4, 79 - 91, 29.12.2025
https://doi.org/10.18466/cbayarfbe.1644296

Abstract

References

  • [1]. Mienye, I. D. and Sun, Y. (2022). A survey of ensemble learning: concepts, algorithms, applications, and prospects. IEEE Access; 10, 99129-99149. https://doi.org/10.1109/access.2022.3207287
  • [2]. Sagi, O. and Rokach, L. (2018). Ensemble learning: a survey. WIREs Data Mining and Knowledge Discovery; 8(4). https://doi.org/10.1002/widm.1249
  • [3]. Ataş, P.K. (2024). Evaluate student achievement by classifying brain structure and its functionality with novel hybrid method. Neural Computing and Applications; 36, 3357–3368. https://doi.org/10.1007/s00521-023-09031-9
  • [4]. Bako, H. 2023. Predicting timely graduation of postgraduate students using random forests ensemble method. Fudma Journal of Sciences; 7(3): 177-185. https://doi.org/10.33003/fjs-2023-0703-1773
  • [5]. Rismayati, R., Ismarmiaty, I., Hidayat, S. (2022). Ensemble Implementation for Predicting Student Graduation with Classification Algorithm. International Journal of Engineering and Computer Science Applications (IJECSA); 1(1), 35–42. https://doi.org/10.30812/ijecsa.v1i1.1805
  • [6]. Pandey, M. and Taruna, S. 2014. A comparative study of ensemble methods for students’ performance modeling. International Journal of Computer Applications; 103(8): 26-32. https://doi.org/10.5120/18095-9151
  • [7]. Kang, Z. 2019. Using machine learning algorithms to predict first-generation college students’ six-year graduation: a case study. International Journal of Information Technology and Computer Science;11(9): 1-8. https://doi.org/10.5815/ijitcs.2019.09.01
  • [8]. Desfiandi, A. and Soewito, B. 2023. Student graduation time prediction using logistic regression, decision tree, support vector machine, and adaboost ensemble learning. IJISCS (International Journal of Information System and Computer Science); 7(3): 195. https://doi.org/10.56327/ijiscs.v7i2.1579
  • [9]. Law, T.-J., Ting, C.-Y., Ng, H., Goh, H.-N., Quek, A. (2024). Ensemble-SMOTE: Mitigating Class Imbalance in Graduate on Time Detection. Journal of Informatics and Web Engineering; 3(2), 229–250. https://doi.org/10.33093/jiwe.2024.3.2.17
  • [10]. Ananto, N. (2024). Leveraging ensemble learning for predicting student graduation: a data mining approach. Prosiding Seminar & Conference FMI; 2: 353-365. https://doi.org/10.47747/snfmi.v2i1.2320
  • [11]. Lagman, A., Alfonso, L., Goh, M., Lalata, J., Magcuyao, J., & Vicente, H. 2020. Classification algorithm accuracy improvement for student graduation prediction using ensemble model. International Journal of Information and Education Technology; 10(10): 723-727. https://doi.org/10.18178/ijiet.2020.10.10.1449
  • [12]. Rachmawati, D. A., Ibadurrahman, N. A., Zeniarja, J., Hendriyanto, N. (2023). Implementation of the Random Forest Algorithm in Classifying the Accuracy of Graduation Time for Computer Engineering Students at Dian Nuswantoro University. Jurnal Teknik Informatika (Jutif); 4(3): 565–572. https://doi.org/10.52436/1.jutif.2023.4.3.920
  • [13]. Freund, Y., Schapire, R. E. (1997). A Decision-Theoretic Generalization of On-Line Learning and an Application to Boosting. Journal of Computer and System Sciences; 55(1), 119–139.
  • [14]. Prokhorenkova, L., Gusev, G., Vorobev, A., Dorogush, A. V., Gulin, A. (2017). CatBoost: unbiased boosting with categorical features. https://doi.org/10.48550/ARXIV.1706.09516
  • [15]. Friedman, J. H. (2001). Greedy function approximation: A gradient boosting machine. The Annals of Statistics, 29(5), 1189–1232.
  • [16]. Chen, T., Guestrin, C. (2016). XGBoost: A Scalable Tree Boosting System. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ACM; San Francisco California USA, pp 785–794
  • [17]. Breiman, L. (2001). Random Forests. Machine Learning, 45(1), 5–32.
  • [18]. Hosmer, D. W., Lemeshow, S. (2000). Applied Logistic Regression, 1st ed, Wiley, New York
  • [19]. Goodfellow, I., Bengio, Y., Courville, A. (2016). Deep learning, The MIT press, Cambridge
  • [20]. Cortes, C., Vapnik, V. (1995). Support-vector networks. Machine Learning, 20(3), 273–297.
  • [21]. Breiman, L., Friedman, J. H., Olshen, R. A., Stone, C. J. (2017). Classification And Regression Trees, 1st ed, Routledge, New York

A Comparative Study of Ensemble Learning Models for Predicting Student On-Time Graduation

Year 2025, Volume: 21 Issue: 4, 79 - 91, 29.12.2025
https://doi.org/10.18466/cbayarfbe.1644296

Abstract

On-time graduation is one of the primary goals of undergraduate programs, aiming to equip students with the necessary knowledge and skills in a specific field to ensure their rapid integration into the workforce. However, delays in graduation can postpone career entry, increase financial burdens, and create psychological stress. Therefore, accurately predicting students' graduation outcomes at early stages is critically important for higher education institutions to develop effective educational strategies and provide timely interventions for at-risk individuals. This study aimed to determine the most effective approach for predicting on-time graduation by comparing ensemble learning methods with traditional machine learning models using demographic data, high school performance, and progressively accumulated university academic data over six academic periods. The models used include ensemble methods such as Random Forest, CatBoost, and XGBoost, as well as traditional models like Logistic Regression, Decision Trees, and Support Vector Machines. Across seven prediction checkpoints (T0–T6), CatBoost consistently outperformed all other models, achieving 83.0% accuracy, 84.8% F1-score, 84.4% precision, and 85.1% recall in the final prediction stage (T6). Among traditional models, Logistic Regression performed the best, with 83.5% accuracy and 84.7% F1-score at T6. Statistical analyses (paired t-test and Wilcoxon test) confirmed that ensemble models significantly outperformed traditional ones in terms of both accuracy and F1-score (p < 0.05).

References

  • [1]. Mienye, I. D. and Sun, Y. (2022). A survey of ensemble learning: concepts, algorithms, applications, and prospects. IEEE Access; 10, 99129-99149. https://doi.org/10.1109/access.2022.3207287
  • [2]. Sagi, O. and Rokach, L. (2018). Ensemble learning: a survey. WIREs Data Mining and Knowledge Discovery; 8(4). https://doi.org/10.1002/widm.1249
  • [3]. Ataş, P.K. (2024). Evaluate student achievement by classifying brain structure and its functionality with novel hybrid method. Neural Computing and Applications; 36, 3357–3368. https://doi.org/10.1007/s00521-023-09031-9
  • [4]. Bako, H. 2023. Predicting timely graduation of postgraduate students using random forests ensemble method. Fudma Journal of Sciences; 7(3): 177-185. https://doi.org/10.33003/fjs-2023-0703-1773
  • [5]. Rismayati, R., Ismarmiaty, I., Hidayat, S. (2022). Ensemble Implementation for Predicting Student Graduation with Classification Algorithm. International Journal of Engineering and Computer Science Applications (IJECSA); 1(1), 35–42. https://doi.org/10.30812/ijecsa.v1i1.1805
  • [6]. Pandey, M. and Taruna, S. 2014. A comparative study of ensemble methods for students’ performance modeling. International Journal of Computer Applications; 103(8): 26-32. https://doi.org/10.5120/18095-9151
  • [7]. Kang, Z. 2019. Using machine learning algorithms to predict first-generation college students’ six-year graduation: a case study. International Journal of Information Technology and Computer Science;11(9): 1-8. https://doi.org/10.5815/ijitcs.2019.09.01
  • [8]. Desfiandi, A. and Soewito, B. 2023. Student graduation time prediction using logistic regression, decision tree, support vector machine, and adaboost ensemble learning. IJISCS (International Journal of Information System and Computer Science); 7(3): 195. https://doi.org/10.56327/ijiscs.v7i2.1579
  • [9]. Law, T.-J., Ting, C.-Y., Ng, H., Goh, H.-N., Quek, A. (2024). Ensemble-SMOTE: Mitigating Class Imbalance in Graduate on Time Detection. Journal of Informatics and Web Engineering; 3(2), 229–250. https://doi.org/10.33093/jiwe.2024.3.2.17
  • [10]. Ananto, N. (2024). Leveraging ensemble learning for predicting student graduation: a data mining approach. Prosiding Seminar & Conference FMI; 2: 353-365. https://doi.org/10.47747/snfmi.v2i1.2320
  • [11]. Lagman, A., Alfonso, L., Goh, M., Lalata, J., Magcuyao, J., & Vicente, H. 2020. Classification algorithm accuracy improvement for student graduation prediction using ensemble model. International Journal of Information and Education Technology; 10(10): 723-727. https://doi.org/10.18178/ijiet.2020.10.10.1449
  • [12]. Rachmawati, D. A., Ibadurrahman, N. A., Zeniarja, J., Hendriyanto, N. (2023). Implementation of the Random Forest Algorithm in Classifying the Accuracy of Graduation Time for Computer Engineering Students at Dian Nuswantoro University. Jurnal Teknik Informatika (Jutif); 4(3): 565–572. https://doi.org/10.52436/1.jutif.2023.4.3.920
  • [13]. Freund, Y., Schapire, R. E. (1997). A Decision-Theoretic Generalization of On-Line Learning and an Application to Boosting. Journal of Computer and System Sciences; 55(1), 119–139.
  • [14]. Prokhorenkova, L., Gusev, G., Vorobev, A., Dorogush, A. V., Gulin, A. (2017). CatBoost: unbiased boosting with categorical features. https://doi.org/10.48550/ARXIV.1706.09516
  • [15]. Friedman, J. H. (2001). Greedy function approximation: A gradient boosting machine. The Annals of Statistics, 29(5), 1189–1232.
  • [16]. Chen, T., Guestrin, C. (2016). XGBoost: A Scalable Tree Boosting System. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ACM; San Francisco California USA, pp 785–794
  • [17]. Breiman, L. (2001). Random Forests. Machine Learning, 45(1), 5–32.
  • [18]. Hosmer, D. W., Lemeshow, S. (2000). Applied Logistic Regression, 1st ed, Wiley, New York
  • [19]. Goodfellow, I., Bengio, Y., Courville, A. (2016). Deep learning, The MIT press, Cambridge
  • [20]. Cortes, C., Vapnik, V. (1995). Support-vector networks. Machine Learning, 20(3), 273–297.
  • [21]. Breiman, L., Friedman, J. H., Olshen, R. A., Stone, C. J. (2017). Classification And Regression Trees, 1st ed, Routledge, New York
There are 21 citations in total.

Details

Primary Language English
Subjects Computer Software
Journal Section Research Article
Authors

Ahmet Kala 0000-0002-0598-1181

Cem Özkurt 0000-0002-1251-7715

Hasan Yaşar 0009-0007-6676-3122

Submission Date February 21, 2025
Acceptance Date June 18, 2025
Publication Date December 29, 2025
Published in Issue Year 2025 Volume: 21 Issue: 4

Cite

APA Kala, A., Özkurt, C., & Yaşar, H. (2025). A Comparative Study of Ensemble Learning Models for Predicting Student On-Time Graduation. Celal Bayar University Journal of Science, 21(4), 79-91. https://doi.org/10.18466/cbayarfbe.1644296
AMA Kala A, Özkurt C, Yaşar H. A Comparative Study of Ensemble Learning Models for Predicting Student On-Time Graduation. CBUJOS. December 2025;21(4):79-91. doi:10.18466/cbayarfbe.1644296
Chicago Kala, Ahmet, Cem Özkurt, and Hasan Yaşar. “A Comparative Study of Ensemble Learning Models for Predicting Student On-Time Graduation”. Celal Bayar University Journal of Science 21, no. 4 (December 2025): 79-91. https://doi.org/10.18466/cbayarfbe.1644296.
EndNote Kala A, Özkurt C, Yaşar H (December 1, 2025) A Comparative Study of Ensemble Learning Models for Predicting Student On-Time Graduation. Celal Bayar University Journal of Science 21 4 79–91.
IEEE A. Kala, C. Özkurt, and H. Yaşar, “A Comparative Study of Ensemble Learning Models for Predicting Student On-Time Graduation”, CBUJOS, vol. 21, no. 4, pp. 79–91, 2025, doi: 10.18466/cbayarfbe.1644296.
ISNAD Kala, Ahmet et al. “A Comparative Study of Ensemble Learning Models for Predicting Student On-Time Graduation”. Celal Bayar University Journal of Science 21/4 (December2025), 79-91. https://doi.org/10.18466/cbayarfbe.1644296.
JAMA Kala A, Özkurt C, Yaşar H. A Comparative Study of Ensemble Learning Models for Predicting Student On-Time Graduation. CBUJOS. 2025;21:79–91.
MLA Kala, Ahmet et al. “A Comparative Study of Ensemble Learning Models for Predicting Student On-Time Graduation”. Celal Bayar University Journal of Science, vol. 21, no. 4, 2025, pp. 79-91, doi:10.18466/cbayarfbe.1644296.
Vancouver Kala A, Özkurt C, Yaşar H. A Comparative Study of Ensemble Learning Models for Predicting Student On-Time Graduation. CBUJOS. 2025;21(4):79-91.