Diabetic Patient Readmission Prediction Using BiLSTM with XGBoost

Authors

  • Sneha Nain Undergraduate Student, School of Computer Science, Engineering and Information Systems, Vellore Institute of Technology, Vellore, Tamil Nadu, India
  • Dhruv Sonar Undergraduate Student, School of Computer Science, Engineering and Information Systems, Vellore Institute of Technology, Vellore, Tamil Nadu, India
  • Prathmesh Gahukar Undergraduate Student, School of Computer Science, Engineering and Information Systems, Vellore Institute of Technology, Vellore, Tamil Nadu, India
  • B. K. Tripathy Professor, School of Computer Science, Engineering and Information Systems, Vellore Institute of Technology, Vellore, Tamil Nadu, India

DOI:

https://doi.org/10.46610/JoANNLS.2025.v02i03.001

Keywords:

BiLSTM, Deep learning, Hybrid model, Machine learning, XGBoost

Abstract

Predicting diabetic patient hospital readmissions is a crucial healthcare task, as unplanned readmissions result in increased medical expenses and may lead to treatment gaps. Proper prediction of readmissions enables better patient care and resource allocation. Whereas conventional machine learning algorithms tend to fail at retaining sequential dependencies in patient data, deep learning algorithms such as LSTM suffer when handling imbalanced datasets. A hybrid BiLSTM-XGBoost model is presented in this study to enhance readmission prediction in patients with diabetes. While XGBoost improves feature importance-based classification and gradient boosting, the BiLSTM component records temporal trends in patient histories. A highly unbalanced dataset of 50,000 diabetes patient histories—which includes class imbalance and missing value handling from preprocessing operations—is used to train and evaluate the model. The proposed BiLSTM-XGBoost model is also superior, yielding an AUROC of 0.9973, a recall of 0.9983, and an F1-score of 0.9900. The suggested model surpasses conventional machine learning algorithms like random forest, XGBoost, LightGBM, naïve Bayes, and CatBoost, as well as deep learning algorithms like LSTM, in forecasting patient readmission hazards. The findings emphasize the usefulness of integrating deep learning with boosting methods for more stable decision-making in healthcare analysis.

References

D. Kansagara, H. Englander, A. Salanitro, D. Kagen, C. Theobald, M. Freeman, and S. Kripalani. Risk prediction models for hospital readmission: A systematic review. JAMA, vol. 306, no. 15, pp. 1688–1698, Oct. 2011. doi: https://doi.org/10.1001/jama.2011.1515

S. F. Jencks, M. V. Williams, and E. A. Coleman, “Rehospitalizations among Patients in the Medicare Fee-for-Service Program,” New England Journal of Medicine, vol. 360, no. 14, pp. 1418–1428, Apr. 2009, doi: https://doi.org/10.1056/nejmsa0803563.

R. B. Zuckerman, S. H. Sheingold, E. J. Orav, J. Ruhter, and A. M. Epstein. Readmissions, observation, and the hospital readmissions reduction program. New England Journal of Medicine. vol. 374, no. 16, pp. 1543–1551, Apr. 2016. doi: https://doi.org/10.1056/NEJMsa1513024

D. Kaul, H. Raju, and B. K. Tripathy. Deep learning in healthcare. InDeep learning in data analytics: Recent techniques, practices and applications 2021 Aug 12 (pp. 97-115). Cham: Springer International Publishing. doi: https://doi.org/10.1007/978-3-030-75855-4_6

D. J. Rubin. Hospital readmission of patients with diabetes. Current Diabetes Reports. vol. 15, no. 4, p. 17, Apr. 2015.doi: https://doi.org/10.1007/s11892-015-0584-7

Y. Huang, A. Talwar, S. Chatterjee, and R. R. Aparasu, “Application of machine learning in predicting hospital readmissions: a scoping review of the literature,” BMC Medical Research Methodology, vol. 21, no. 1, May 2021, doi: https://doi.org/10.1186/s12874-021-01284-z.

A. Rajkomar, J. Dean, and I. Kohane, “Machine Learning in Medicine,” New England Journal of Medicine, vol. 380, no. 14, pp. 1347–1358, 2019, doi: https://doi.org/10.1056/nejmra1814259.

T. Chen and C. Guestrin. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2016 Aug 13 (pp. 785-794). doi: https://doi.org/10.1145/2939672.2939785

S. Hochreiter and J. Schmidhuber, “Long Short-Term Memory,” Neural Computation, vol. 9, no. 8, pp. 1735–1780, Nov. 1997, doi: https://doi.org/10.1162/neco.1997.9.8.1735

R. Miotto, F. Wang, S. Wang, X. Jiang, and J. T. Dudley, “Deep Learning for healthcare: review, Opportunities and Challenges,” Briefings in Bioinformatics, vol. 19, no. 6, pp. 1236–1246, 2018, doi: https://doi.org/10.1093/bib/bbx044

M. Ghassemi, T. Naumann, P. Schulam, A. L. Beam, and R. Ranganath. Opportunities in machine learning for healthcare. arXiv preprint arXiv:1806.00388. 2018 Mar 30. Available: https://healthyml.org/publication/ghassemi-2018-opportunities/

W. Liu, C. Stansbury, K. Singh, A. M. Ryan, D. Sukul, E. Mahmoudi, A. Waljee, J. Zhu, and B. K. Nallamothu. Predicting 30-day hospital readmissions using artificial neural networks with medical code embedding. Plos One. vol. 15, no. 4, p. e0221606, Apr. 2020. doi: https://doi.org/10.1371/journal.pone.0221606

S. Bhattacharyya, V. Snasel, A. E. Hassanien, S. Saha, and B. K. Tripathy, editors. Deep Learning: Research and Applications. Walter de Gruyter GmbH & Co KG; 2020 Jun 22.

N. V. Chawla, K. W. Bowyer, L. O. Hall, and W. P. Kegelmeyer. SMOTE: Synthetic minority over-sampling technique. Journal of Artificial Intelligence Research. vol. 16: 321-57 Jun 2002. doi: https://doi.org/10.1613/jair.953

H. He, Y. Bai, E. A. Garcia, and S. Li, “ADASYN: Adaptive synthetic sampling approach for imbalanced learning,” in Proc. IEEE Int. Joint Conf. Neural Netw. (IEEE World Congr. Comput. Intell.), Hong Kong, China, Jun. 2008, pp. 1322–1328. doi: https://doi.org/10.1109/IJCNN.2008.4633969

J. M. Johnson and T. M. Khoshgoftaar. Survey on deep learning with class imbalance. Journal of Big Data. vol. 6, no. 1, pp. 1-54. Dec. 2019. doi: https://doi.org/10.1186/s40537-019-0192-5

F. Ma, J. Gao, Q. Suo, Q. You, J. Zhou, and A. Zhang, “Risk Prediction on Electronic Health Records with Prior Medical Knowledge,” Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Jul. 2018, doi: https://doi.org/10.1145/3219819.3220020.

S. M. Lundberg and S. I. Lee. A unified approach to interpreting model predictions. Advances in Neural Information Processing Systems. vol. 30, 2017. Available: https://dl.acm.org/doi/abs/10.5555/3295222.3295230

S. Barbieri, J. Kemp, O. Perez-Concha, S. Kotwal, M. Gallagher, A. Ritchie, and L. Jorm. Benchmarking deep learning architectures for predicting readmission to the ICU and describing patients-at-risk. Scientific Reports. vol. 10, no. 1, p. 1111, Jan. 2020. doi: https://doi.org/10.1038/s41598-020-58053-z

M. Buda, A. Maki, and M. A. Mazurowski. A systematic study of the class imbalance problem in convolutional neural networks. Neural Networks. vol. 106, pp. 249–259, oct. 2018. doi: https://doi.org/10.1016/j.neunet.2018.07.011

N. Japkowicz and S. Stephen, “The class imbalance problem: A systematic study,” Intelligent Data Analysis, vol. 6, no. 5, pp. 429–449, Nov. 2002, doi: https://doi.org/10.3233/ida-2002-6504.

C. K. Reddy and C. C. Aggarwal. Healthcare data analytics. CRC Press; 2015 Jun 23.

A. Rajkomar et al., “Scalable and accurate deep learning with electronic health records,” Npj Digital Medicine, vol. 1, no. 1, May 2018, doi: https://doi.org/10.1038/s41746-018-0029-1.

H. He and E. A. Garcia, “Learning from Imbalanced Data,” IEEE Transactions on Knowledge and Data Engineering, vol. 21, no. 9, pp. 1263–1284, Sep. 2009, doi: https://doi.org/10.1109/tkde.2008.239.

A. E. W. Johnson et al., “MIMIC-IV, a freely accessible electronic health record dataset,” Scientific Data, vol. 10, no. 1, Jan. 2023, doi: https://doi.org/10.1038/s41597-022-01899-x.

Published

2025-09-09

Issue

Section

Articles