Machine Learning for Student Performance Prediction in Online Learning, MOOCS, and Learning Management Systems: A Systematic Literature Review

Authors

M. Z. A. Chek

Actuarial Science Department, UiTM Perak Branch (Malaysia)

I. L. Ismail

Department of Statistics and Decision Science, UiTM Perak Branch (Malaysia)

N. Jamal

Department of Statistics and Decision Science, UiTM Perak Branch (Malaysia)

Z. H. Zulkifli

Actuarial Partners Consulting (Malaysia)

Rinda Nariswari

Department of Computer Science, BINUS (Malaysia)

Article Information

DOI: 10.47772/IJRISS.2026.100300191

Subject Category:

Volume/Issue: 10/3 | Page No: 2655-2668

Publication Timeline

Submitted: 2026-03-15

Accepted: 2026-03-20

Published: 2026-03-31

Abstract

The rapid expansion of online learning in higher education has generated large volumes of learner interaction data through Learning Management Systems (LMSs), Massive Open Online Courses (MOOCs), and related digital platforms. These data provide new opportunities for machine learning to predict academic performance, identify at-risk learners, and support timely intervention.
This study presents a systematic literature review of machine learning approaches used for student performance prediction in online learning environments, with specific focus on MOOCs, LMS data, and digital learning traces. Guided by the PRISMA 2020 framework, the review synthesizes evidence from peer-reviewed studies and addresses five questions: the most common machine learning algorithms, the types of online learning data and predictive features employed, the major prediction targets, the evaluation methods used, and the main research gaps in the field.
The literature indicates that classification-based models dominate the field, with Random Forest, Support Vector Machine, Decision Tree, Artificial Neural Network, and Naïve Bayes among the most frequently used approaches. LMS logs, MOOC clickstreams, assessment records, historical grades, and demographic variables are the most common predictive inputs, while final grades, pass/fail outcomes, dropout, and retention are the main targets.
The review also identifies persistent weaknesses, including limited explainability, weak cross-institutional validation, inconsistent reporting of feature importance, and relatively few studies that evaluate the effect of interventions after prediction. The manuscript concludes with a conceptual framework and a future research agenda to support robust, ethical, and actionable machine learning in online higher education.

Keywords

Online learning; MOOCs; learning management systems; machine learning; student performance prediction; educational data mining; learning analytics; higher education

Downloads

References

1. A. Palancı, R. M. Yılmaz, and Z. Turan, “Learning analytics in distance education: A systematic review study,” Educ. Inf. Technol., vol. 29, pp. 22629–22650, 2024. [Google Scholar] [Crossref]

2. S. R. A. Abdullah and A. Al-Azawei, “Predicting online learners’ performance through ontologies: A systematic literature review,” Int. Rev. Res. Open Distrib. Learn., 2025. [Google Scholar] [Crossref]

3. S. Azimi, C.-G. Popa, and T. Cucić, “Improving students performance in small-scale online courses: A machine learning-based intervention,” Int. J. Learn. Anal. Artif. Intell. Educ., vol. 2, no. 2, pp. 80–95, 2020. [Google Scholar] [Crossref]

4. P. Muljana and T. Luo, “Factors contributing to student retention in online learning and recommended strategies for improvement: A systematic literature review,” J. Inf. Technol. Educ. Res., vol. 18, pp. 19–57, 2019. [Google Scholar] [Crossref]

5. L. R. Pelima, Y. Sukmana, and Y. Rosmansyah, “Predicting university student graduation using academic performance and machine learning: A systematic literature review,” IEEE Access, vol. 12, pp. 23451–23465, 2024. [Google Scholar] [Crossref]

6. S. Kharis et al., “Design of student success prediction application in online learning using fuzzy-KNN,” BAREKENG J. Ilmu Mat. dan Terap., 2023. [Google Scholar] [Crossref]

7. M. Z. A. Chek and I. L. Ismail, “Retirement Planning Issues, Problems, and Opportunities in Malaysia,” Retire. Plan. Issues, Probl. Oppor. Malaysia, vol. VII, no. 2454, pp. 1926–1932, 2023, doi: 10.47772/IJRISS. [Google Scholar] [Crossref]

8. G. Siemens, “Massive open online courses: Innovation in education?,” Open Learn. J. Open, Distance e-Learning, vol. 28, no. 3, pp. 207–214, 2013. [Google Scholar] [Crossref]

9. R. Koper and C. Tattersall, “Preface to Learning Design: A Handbook on Modelling and Delivering Networked Education and Training,” J. Interact. Media Educ., vol. 18, pp. 1–7, 2005. [Google Scholar] [Crossref]

10. D. Laurillard, “The educational problem that MOOCs could solve: Professional development for teachers of disadvantaged students,” Res. Learn. Technol., vol. 24, 2016. [Google Scholar] [Crossref]

11. H. Munir, B. Vogel, and A. Jacobsson, “Artificial intelligence and machine learning approaches in digital education: A systematic revision,” Information, vol. 13, no. 4, p. 203, 2022. [Google Scholar] [Crossref]

12. G. G. Dongre, “Predicting Student Dropout Rates in Higher Education: A Comparative Study of Machine Learning Algorithms,” Int. J. Sci. Res. Eng. Manag., vol. 8, no. 2, pp. 1–10, 2024. [Google Scholar] [Crossref]

13. M. Yağcı, “Educational Data Mining: Prediction of Students’ Academic Performance Using Machine Learning Algorithms,” Smart Learn. Environ., vol. 9, p. 11, 2022, doi: 10.1186/s40561-022-00192-z. [Google Scholar] [Crossref]

14. H. Abdul Halim, W. C. Yap, and K. H. Chua, “Application of Artificial Neural Networks in Predicting Return-to-Work Outcomes: A Case Study of SOCSO’s RTW Initiatives,” Int. J. Occup. Med. Environ. Health, vol. 36, no. 4, pp. 345–358, 2023. [Google Scholar] [Crossref]

15. H. K. Teoh, A. B. Abdullah, and W. C. Yap, “Performance Evaluation of Artificial Neural Networks in Predicting RTW Outcomes: A Case Study of SOCSO,” J. Occup. Rehabil., vol. 40, no. 4, pp. 456–469, 2023. [Google Scholar] [Crossref]

16. C. Romero and S. Ventura, “Educational Data Mining: A Review of the State of the Art,” IEEE Trans. Syst. Man. Cybern., vol. 40, no. 6, pp. 601–618, 2010. [Google Scholar] [Crossref]

17. Y. Zhang, Y. Yun, R. An, J. Cui, H. Dai, and X. Shang, “Educational data mining techniques for student performance prediction: Method review and comparison analysis,” Front. Psychol., vol. 12, p. 698490, 2021. [Google Scholar] [Crossref]

18. L. Yu, J. Zhang, and B. Zhao, “Proximity effect in disciplinary and regional MOOC performance: Evidence from global data,” Educ. Technol. \& Soc., vol. 26, no. 1, pp. 55–70, 2023. [Google Scholar] [Crossref]

19. B. A. Weyori, P. S. Bañeres, and D. Rodríguez-González, “Using LMS log data to identify at-risk students: A systematic review of machine learning approaches and bibliographic analysis,” African J. Appl. Res., 2025. [Google Scholar] [Crossref]

20. R. S. Baker and K. Yacef, “The State of Educational Data Mining in 2009: A Review and Future Visions,” J. Educ. Data Min., vol. 1, no. 1, pp. 3–17, 2009. [Google Scholar] [Crossref]

21. K. Alalawi, R. Athauda, and R. Chiong, “Contextualizing the current state of research on the use of machine learning for student performance prediction: A systematic literature review,” Eng. Reports, vol. 5, p. e12699, 2023. [Google Scholar] [Crossref]

22. G. Siemens and R. Baker, “Learning Analytics and Educational Data Mining: Towards Communication and Collaboration,” in Proceedings of the 2nd International Conference on Learning Analytics and Knowledge, 2012, pp. 252–254. [Google Scholar] [Crossref]

23. K. F. Hew and W. S. Cheung, “Students’ and instructors’ use of massive open online courses (MOOCs): Motivations and challenges,” Educ. Res. Rev., vol. 12, pp. 45–58, 2014. [Google Scholar] [Crossref]

24. D. Jansen and R. Schuwer, “Institutional MOOC strategies in Europe: Status report based on a mapping survey conducted in October–December 2014,” 2015. [Google Scholar] [Crossref]

25. L. Luo, L. Zhang, and Y. Wang, “Research on MOOC learning effectiveness based on learner behavioral classification,” China Educ. Technol., vol. 10, no. 2, pp. 34–42, 2018. [Google Scholar] [Crossref]

26. F. Tahiru, “Predicting at-risk students in a higher educational institution in Ghana for early intervention using machine learning.” 2023. [Google Scholar] [Crossref]

27. R. Trakunphutthirak and V. C. Lee, “Application of educational data mining approach for student academic performance prediction using progressive temporal data,” J. Educ. Comput. Res., vol. 60, pp. 742–776, 2021. [Google Scholar] [Crossref]

28. C. Romero and S. Ventura, “Data Mining in Education,” Wiley Interdiscip. Rev. Data Min. Knowl. Discov., vol. 3, no. 1, pp. 12–27, 2013. [Google Scholar] [Crossref]

29. I. L. Ismail, N. F. Jamal, M. Z. Awang Chek, and M. S. Baharuddin, “Learning Basic Statistics and Probability Through MOOC,” Int. J. Mod. Trends Soc. Sci., vol. 2, no. 8, pp. 99–107, 2019, doi: 10.35631/ijmtss.280010. [Google Scholar] [Crossref]

30. A. N. A. Ahmad Ridzuan, M. Z. Awang Chek, N. M. Abdul Ghafar, and A. B. Ahmad, “Developing an Introduction to Actuarial Science MOOC,” Int. J. Acad. Res. Bus. Soc. Sci., vol. 8, no. 1, pp. 600–605, 2018, doi: 10.6007/ijarbss/v8-i1/3833. [Google Scholar] [Crossref]

31. E. Ahmed, “Student performance prediction using machine learning algorithms,” Appl. Comput. Intell. Soft Comput., 2024. [Google Scholar] [Crossref]

32. M. Chek, I. L. Ismail, and N. F. Jamal, “Learning actuarial mathematics through UiTM future MOOC,” in MOOC for University and Beyond, Springer, 2020, pp. 75–90. doi: 10.1007/978-3-030-21836-2_6. [Google Scholar] [Crossref]

33. M. Z. A. Chek, I. L. Ismail, M. Syakir, and A. F. Mansor, “Development Micro-Credentials Overview Social Security,” vol. 13, no. 11, pp. 2201–2211, 2023, doi: 10.6007/IJARBSS/v13-i11/19628. [Google Scholar] [Crossref]

Metrics

Views & Downloads

Similar Articles