Learning Analytics for Early Identification of At-Risk Students and Feedback Intervention

Authors

DOI:

https://doi.org/10.18608/jla.2025.8735

Keywords:

learning analytics, predictive modelling, model generalization ability, early feedback intervention, effective feedback design, research paper

Abstract

Supporting academically at-risk students has attracted much attention in the field of learning analytics. However, much of the research in this area has focused on developing advanced machine learning models to predict students' academic performance, which alone is insufficient to improve student learning without the implementation of timely interventions. Among the studies that attempted to mitigate this limitation by deploying intervention feedback to enhance learning, few created their feedback based on established theories of effective feedback. This theoretical oversight may limit students' uptake of the provided intervention. In response to these gaps, we conducted a study that aimed at supporting at-risk students at the early stage of an undergraduate-level course. Specifically, we developed predictive machine learning models using trace and academic data from the previous offering of a course, and applied these models to identify at-risk students in the subsequent semester's offering of the same course. For the identified at-risk students, we sent intervention emails designed by feedback experts based on a relational feedback framework designed to enhance feedback effectiveness by strengthening student-instructor relationships. We evaluated the effectiveness of the proposed approach by assessing the performance of the predictive models in terms of generalisability, and measuring the impact of the feedback intervention on students' learning engagement. Results showed that i) our predictive models demonstrated a high prediction accuracy (with AUC scores above 0.8) when applied to a new cohort of students; ii) more than 30% of the identified at-risk students visited previously unengaged learning activities within two weeks following the intervention; and iii) survey responses from 9.27% of at-risk students indicated general satisfaction with the provided feedback intervention, and 60\% of the respondents expressed a preference for receiving the intervention more frequently than the twice-per-semester frequency implemented in the present study.

References

Adnan, M., Habib, A., Ashraf, J., Mussadiq, S., Raza, A. A., Abid, M., . . . Khan, S. U. (2021). Predicting at-risk students at different percentages of course length for early intervention using machine learning models. Ieee Access, 9, 7519–7539. https://doi.org/10.1109/ACCESS.2021.3049446

Akcapınar, G., Altun, A., & Askar, P. (2019). Using learning analytics to develop early-warning system for at-risk students. International Journal of Educational Technology in Higher Education, 16(1), 1–20. https://doi.org/10.1186/s41239-019-0172-z

Arnold, K. E., & Pistilli, M. D. (2012). Course signals at purdue: Using learning analytics to increase student success. In Proceedings of the 2nd international conference on learning analytics and knowledge (LAK 2012), 29 April–2 May 2012, Vancouver, British Columbia, Canada (pp. 267–270). ACM.. https://doi.org/10.1145/2330601.2330666

Bainbridge, J., Melitski, J., Zahradnik, A., Laurıa, E. J., Jayaprakash, S., & Baron, J. (2015). Using learning analytics to predict at-risk students in online graduate public affairs and administration education. Journal of Public Affairs Education, 21(2), 247–262. https://doi.org/10.1080/15236803.2015.12001831

Baker, R., Evans, B., & Dee, T. (2016). A randomized experiment testing the efficacy of a scheduling nudge in a massive open online course (mooc). Aera Open, 2(4), 2332858416674007. https://doi.org/10.1177/2332858416674007

Baker, R. S., Lindrum, D., Lindrum, M. J., & Perkowski, D. (2015). Analyzing early at-risk factors in higher education e learning courses. International Educational Data Mining Society. vitz, S. Ventura, & M. Desmarais (Eds.), Proceedings of the Eighth International Conference on Educational Data Mining (EDM 2015), 26–29 June 2015, Madrid, Spain (pp. 150–155). International Educational Data Mining Society. https://files.eric.ed.gov/fulltext/ED560503.pdf

Baneres, D., Rodríguez, M. E., Guerrero-Roldán, A. E., & Karadeniz, A. (2020). An early warning system to detect at-risk students in online higher education. Applied Sciences, 10(13), 4427. https://doi.org/10.3390/app10134427

Berka, P., & Marek, L. (2021). Bachelor’s degree student dropouts: Who tend to stay and who tend to leave? Studies in Educational Evaluation, 70, 100999. https://doi.org/10.1016/j.stueduc.2021.100999

Borrella, I., Caballero-Caballero, S., & Ponce-Cueto, E. (2019). Predict and intervene: Addressing the dropout problem in a mooc-based program. In Proceedings of the sixth (2019) acm conference on learning@ scale (pp. 1–9). ACM. https://doi.org/10.1145/3330430.3333634

Borrella, I., Caballero-Caballero, S., & Ponce-Cueto, E. (2022). Taking action to reduce dropout in moocs: Tested interventions. Computers & Education, 179, 104412. https://doi.org/10.1016/j.compedu.2021.104412

Brooks, C., Thompson, C., & Teasley, S. (2015). A time series interaction analysis method for building predictive models of learners using log data. In Proceedings of the Fifth International Conference on Learning Analytics and Knowledge (LAK 2015), 16–20 April 2015, Poughkeepsie, New York, USA (pp. 126–135). ACM. https://doi.org/10.1145/2723576.2723581

Campbell, J. P., DeBlois, P. B., & Oblinger, D. G. (2007). Academic analytics: A new tool for a new era. EDUCAUSE review, 42(4), 40. https://er.educause.edu/articles/2007/7/academic-analytics-a-new-tool-for-a-new-era

Chipchase, L., Davidson, M., Blackstock, F., Bye, R., Clothier, P., Klupp, N., Nickson, W., Turner, D., & Williams, M. (2017). Conceptualising and measuring student disengagement in higher education: A synthesis of the literature. International Journal of Higher Education, 6(2), 31–42. https://doi.org/10.5430/ijhe.v6n2p31

Christou, V., Tsoulos, I., Loupas, V., Tzallas, A. T., Gogos, C., Karvelis, P. S., Antoniadis, N., Glavas, E., & Giannakeas, N. (2023). Performance and early drop prediction for higher education students using machine learning. Expert Systems with Applications, 225, 120079. https://doi.org/10.1016/j.eswa.2023.120079

Clow, D. (2012). The learning analytics cycle: Closing the loop effectively. In Proceedings of the Second International Conference on Learning Analytics and Knowledge (LAK 2012), 29 April–2 May 2012, Vancouver, British Columbia, Canada (pp. 134–138). ACM. https://doi.org/10.1145/2330601.2330636

Dai, W., Tsai, Y.-S., Gašević, D., & Chen, G. (2025). Designing relational feedback: A rapid review and qualitative synthesis. Assessment & Evaluation in Higher Education, 50(1), 16–30. https://doi.org/10.1080/02602938.2024.2361166

Dawson, S., Jovanovic, J., Gašević, D., & Pardo, A. (2017). From prediction to impact: Evaluation of a learning analytics retention program. In Proceedings of the Seventh International Conference on Learning Analytics and Knowledge (LAK 2017), 13–17 March 2017, Vancouver, British Columbia, Canada (pp. 474–478). ACM. https://doi.org/10.1145/3027385.3027405

Dix, N., Lail, A., Birnbaum, M., & Paris, J. (2020). Exploring the “at-risk” student label through the perspectives of higher education professionals. The Qualitative Report, 25(11). https://doi.org/10.46743/2160-3715/2020.3371

Er, E. (2012). Identifying at-risk students using machine learning techniques: A case study with IS 100. International Journal of Machine Learning and Computing, 2(4), 476. https://www.ijml.org/show-32-132-1.html

Faas, C., Benson, M. J., Kaestle, C. E., & Savla, J. (2018). Socioeconomic success and mental health profiles of young adults who drop out of college. Journal of Youth Studies, 21(5), 669–686. https://doi.org/10.1080/13676261.2017.1406598

Falkner, N. J., & Falkner, K. E. (2012). A fast measure for identifying at-risk students in computer science. In Proceedings of the Ninth Annual International Conference on International Computing Education Research (ICER 2012), 9–11 September 2012, Auckland, New Zealand (pp. 55–62). ACM. https://doi.org/10.1145/2361276.2361288

Figueroa-Cañas, J., & Sancho-Vinuesa, T. (2019). Predicting early dropout students is a matter of checking completed quizzes: The case of an online statistics module. In M. Caeiro-Rodrıguez, A. Hernandez-Garcıa, & P. J. Munoz-Merino (Eds.), Learning Analytics Summer Institute Spain 2019: Learning Analytics in Higher Education (LASI 2019), 27–28 June 2019, Vigo, Spain (pp. 100–111). CEUR Workshop Proceedings. https://ceur-ws.org/Vol-2415/paper09.pdf

Foster, E., & Siddle, R. (2020). The effectiveness of learning analytics for identifying at-risk students in higher education. Assessment & Evaluation in Higher Education, 45(6), 842–854. https://doi.org/10.1080/02602938.2019.1682118

Gardner, J., Yu, R., Nguyen, Q., Brooks, C., & Kizilcec, R. (2023). Cross-institutional transfer learning for educational models: Implications for model performance, fairness, and equity. In Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency (FAccT 2023), 12–15 June 2023, Chicago, Illinois, USA (pp. 1664–1684). ACM. https://doi.org/10.1145/3593013.3594107

Gašević, D., Dawson, S., Rogers, T., & Gašević, D. (2016). Learning analytics should not promote one size fits all: The effects of instructional conditions in predicting academic success. The Internet and Higher Education, 28, 68–84. https://doi.org/10.1016/j.iheduc.2015.10.002

Gupta, S. K., Antony, J., Lacher, F., & Douglas, J. (2020). Lean Six Sigma for reducing student dropouts in higher education—An exploratory study. Total Quality Management & Business Excellence, 31(1-2), 178–193. https://doi.org/10.1080/14783363.2017.1422710

Hattie, J., & Timperley, H. (2007). The power of feedback. Review of Educational Research, 77(1), 81–112. https://doi.org/10.3102/003465430298487

He, J., Bailey, J., Rubinstein, B., & Zhang, R. (2015). Identifying at-risk students in massive open online courses. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI 2015), 25–30 June 2023, Austin, Texas, USA (Vol. 29). PKP|PS. https://doi.org/10.1609/aaai.v29i1.9471

Heron, M., Medland, E., Winstone, N., & Pitt, E. (2023). Developing the relational in teacher feedback literacy: Exploring feedback talk. Assessment & Evaluation in Higher Education, 48(2), 172–185. https://doi.org/10.1080/02602938.2021.1932735

Hu, Y. - H., Lo, C.-L., & Shih, S.- P. (2014). Developing early warning systems to predict students’ online learning performance. Computers in Human Behavior, 36, 469–478. https://doi.org/10.1016/j.chb.2014.04.002

Jayaprakash, S. M., Moody, E. W., Laur´ıa, E. J., Regan, J. R., & Baron, J. D. (2014). Early alert of academically at-risk students: An open source analytics initiative. Journal of Learning Analytics, 1(1), 6–47. https://doi.org/10.18608/jla.2014.11.3

Jovanovic, J., Saqr, M., Joksimovic, S., & Gasevic, D. (2021). Students matter the most in learning analytics: The effects of internal and instructional conditions in predicting academic success. Computers & Education, 172, 104251. https://doi.org/10.1016/j.compedu.2021.104251

Kastberg, S. E., Lischka, A. E., & Hillman, S. L. (2020). Written feedback as a relational practice: Revealing mediating factors. Studying Teacher Education, 16(3), 324–344. https://doi.org/10.1080/17425964.2020.1834152

Khalil, M., Slade, S., & Prinsloo, P. (2024). Learning analytics in support of inclusiveness and disabled students: A systematic review. Journal of Computing in Higher Education, 36(1), 202–219. https://doi.org/10.1007/s12528-023-09363-4

Khan, I. A., Subhani, A., Rasheed, Z., Ahmad, U., & Brohi, M. N. (2023). Comprehensive assessment of risk assessment tools and academic performance in higher education: A meta-analytic perspective. Journal of Applied Engineering & Technology (JAET), 7(2), 10–24. https://doi.org/10.55447/jaet.07.02.116

Latif, A., Choudhary, A., & Hammayun, A. (2015). Economic effects of student dropouts: A comparative study. Journal of Global Economics, 3(2), 1–4. https://www.hilarispublisher.com/open-access/economic-effects-of-student-dropouts-a-comparative-study-2375-4389-1000137.pdf

Liang, Z., Sha, L., Tsai, Y. - S., Gašević, D., & Chen, G. (2024). Towards the automated generation of readily applicable personalised feedback in education. In A. Olney, I. Chounta, Z. Liu, O. Santos, & I. Bittencourt (Eds.), Artificial intelligence in education. AIED 2024. Lecture notes in computer science (pp. 75–88, Vol. 14830). Springer. https://doi.org/10.1007/978-3-031-64299-9_6

Lin, J., Dai, W., Lim, L.-A., Tsai, Y.-S., Mello, R. F., Khosravi, H., Gašević, D., & Chen, G. (2023). Learner-centred analytics of feedback content in higher education. In Proceedings of the 13th International Conference on Learning Analytics and Knowledge (LAK 2023), 13–17 March 2023, Arlington, Texas, USA (pp. 100–110). ACM. https://doi.org/10.1145/3576050.3576064

Liz Domínguez, M., Caeiro Rodríguez, M., Llamas Nistal, M., & Mikic Fonte, F. A. (2019). Predictors and early warning systems in higher education: A systematic literature review. In M. Caeiro-Rodrıguez, A. Hernandez-Garcıa, & P. J. Munoz-Merino (Eds.), Learning Analytics Summer Institute Spain 2019: Learning Analytics in Higher Education (LASI 2019), 27–28 June 2019, Vigo, Spain (pp. 84–99). CEUR Workshop Proceedings. https://ceur-ws.org/Vol-2415/paper08.pdf

López-García, A., Blasco-Blasco, O., Liern-García, M., & Parada-Rico, S. E. (2023). Early detection of students’ failure using machine learning techniques. Operations Research Perspectives, 11, 100292. https://doi.org/10.1016/j.orp.2023.100292

López-Zambrano, J., Lara, J. A., & Romero, C. (2020). Towards portability of models for predicting students’ final performance in university courses starting from Moodle logs. Applied Sciences, 10(1), 354. https://doi.org/10.3390/app10010354

Lu, O. H., Huang, J. C., Huang, A. Y., & Yang, S. J. (2017). Applying learning analytics for improving students engagement and learning outcomes in an MOOCs enabled collaborative programming course. Interactive Learning Environments, 25(2), 220–234. https://doi.org/10.1080/10494820.2016.1278391

Lundberg, S. M., & Lee, S.-I. (2017). A unified approach to interpreting model predictions. arXiv:1705.07874. https://doi.org/10.48550/arXiv.1705.07874

Masabo, E., Nzabanita, J., Ngaruye, I., Ruranga, C., Nizeyimana, J. P., Uwonkunda, J., & Ndanguza, D. (2023). Early detection of students at risk of poor performance in Rwanda higher education using machine learning techniques. International Journal of Information Technology, 15(6), 3201–3210. https://doi.org/10.1007/s41870-023-01334-3

Middleton, T., ahmed Shafi, A., Millican, R., & Templeton, S. (2023). Developing effective assessment feedback: Academic buoyancy and the relational dimensions of feedback. Teaching in Higher Education, 28(1), 118–135. https://doi.org/10.1080/13562517.2020.1777397

Na, K. S., & Tasir, Z. (2017). Identifying at-risk students in online learning by analysing learning behaviour: A systematic review. In 2017 IEEE Conference on Big Data and Analytics (ICBDA 2017), 16–17 November 2017, Kuching, Malaysia (pp. 118–123). IEEE. https://doi.org/10.1109/ICBDAA.2017.8284117

Nimy, E., Mosia, M., & Chibaya, C. (2023). Identifying at-risk students for early intervention—A probabilistic machine learning approach. Applied Sciences, 13(6), 3869. https://doi.org/10.3390/app13063869

Osborne, J. B., & Lang, A. S. (2023). Predictive identification of at-risk students: Using learning management system data. Journal of Postsecondary Student Success, 2(4), 108–126. https://doi.org/10.33009/fsop_jpss132082

Poulos, A., & Mahony, M. J. (2008). Effectiveness of feedback: The students’ perspective. Assessment & Evaluation in Higher Education, 33(2), 143–154. https://doi.org/10.1080/02602930601127869

Ryan, T., Henderson, M., Ryan, K., & Kennedy, G. (2021). Designing learner-centred text-based feedback: A rapid review and qualitative synthesis. Assessment & Evaluation in Higher Education, 46(6), 894–912. https://doi.org/10.1080/02602938.2020.1828819

Schmidt, A., Cechinel, C., Queiroga, E. M., Primo, T., Ramos, V., Bordin, A. S., Mello, R. F., & Muñoz, R. (2025). Analyzing intervention strategies employed in response to automated academic-risk identification: A systematic review. IEEE Revista Iberoamericana de Tecnologias del Aprendizaje, 20, 77–85. https://doi.org/10.1109/RITA.2025.3540161

Singell, L. D., & Waddell, G. R. (2010). Modeling retention at a large public university: Can at-risk students be identified early enough to treat? Research in Higher Education, 51, 546–572. https://doi.org/10.1007/s11162-010-9170-7

Sonnleitner, B., Madou, T., Deceuninck, M., Theodosiou, F., & Sagaert, Y. R. (2025). Evaluation of early student performance prediction given concept drift. Computers and Education: Artificial Intelligence, 8, 100369. https://doi.org/10.1016/j.caeai.2025.100369

Veerasamy, A. K., D’Souza, D., Apiola, M.-V., Laakso, M.-J., & Salakoski, T. (2020). Using early assessment performance as early warning signs to identify at-risk students in programming courses. In Proceedings of the 2020 IEEE Frontiers in Education Conference (FIE), 21–24 October 2020, Uppsala, Sweden (pp. 1–9). IEEE. https://doi.org/10.1109/FIE44824.2020.9274277

Waheed, H., Hassan, S. - U., Nawaz, R., Aljohani, N. R., Chen, G., & Gasevic, D. (2023). Early prediction of learners at risk in self-paced education: A neural network approach. Expert Systems with Applications, 213, 118868. https://doi.org/10.1016/j.eswa.2022.118868

Wang, Q., & Mousavi, A. (2023). Which log variables significantly predict academic achievement? A systematic review and meta-analysis. British Journal of Educational Technology, 54(1), 142–191. https://doi.org/10.1111/bjet.13282

Whitelock-Wainwright, A., Gašević, D., Tsai, Y.- S., Drachsler, H., Scheffel, M., Muñoz-Merino, P. J., Tammets, K., & Delgado Kloos, C. (2020). Assessing the validity of a learning analytics expectation instrument: A multinational study. Journal of Computer Assisted Learning, 36(2), 209–240. https://doi.org/10.1111/jcal.12401

Winston, K. A., van der Vleuten, C. P., & Scherpbier, A. J. (2014). Prediction and prevention of failure: An early intervention to assist at-risk medical students. Medical Teacher, 36(1), 25–31. https://doi.org/10.3109/0142159X.2013.836270

Wong, B. T.-M., & Li, K. C. (2020). A review of learning analytics intervention in higher education (2011–2018). Journal of Computers in Education, 7(1), 7–28. https://doi.org/10.1007/s40692-019-00143-7

Zacharis, N. Z. (2015). A multivariate approach to predicting student outcomes in web-enabled blended learning courses. The Internet and Higher Education, 27, 44–53. https://doi.org/10.1016/j.iheduc.2015.05.002

Zhang, L., & Rangwala, H. (2018). Early identification of at-risk students using iterative logistic regression. In C. P. Rose, R. Martınez-Maldonado, H. U. Hoppe, R. Luckin, M. Mavrikis, K. Porayska-Pomsta, B. McLaren, & B. du Boulay (Eds.), Artificial intelligence in education. AIED 2018. Lecture notes in computer science (pp. 613–626, Vol. 10947). Springer. https://doi.org/10.1007/978-3-319-93843-1_45

Zhang, Y., Fei, Q., Quddus, M., & Davis, C. (2014). An examination of the impact of early intervention on learning outcomes of at-risk students. Research in Higher Education Journal, 26. https://files.eric.ed.gov/fulltext/EJ1055303.pdf

Downloads

Published

2025-11-30

How to Cite

Dai, W., Lin, J., Jin, F. J.-Y., Tsai, Y.-S., Srivastava, N., Le Bodic, P., Gašević, D., & Chen, G. (2025). Learning Analytics for Early Identification of At-Risk Students and Feedback Intervention. Journal of Learning Analytics, 12(3), 102-125. https://doi.org/10.18608/jla.2025.8735

Issue

Section

Research Papers