Vincent B. Liu , Laura Y. Sue , Oscar Madrid Padilla , Yingnian Wu
{"title":"Optimizing blood glucose predictions in type 1 diabetes patients using a stacking ensemble approach","authors":"Vincent B. Liu , Laura Y. Sue , Oscar Madrid Padilla , Yingnian Wu","doi":"10.1016/j.endmts.2025.100253","DOIUrl":null,"url":null,"abstract":"<div><h3>Introduction</h3><div>The diabetes pandemic, including 828 million adults worldwide in 2022, would benefit from continued development of novel, effective and accurate blood glucose prediction systems. Using the DiaTrend dataset, this study used stacking machine learning optimized by Grey Wolf Optimizer to construct and assess prediction models for blood glucose levels in type 1 diabetes patients.</div></div><div><h3>Methods</h3><div>The DiaTrend dataset includes 27,561 days of continuous glucose monitoring and 8220 days of insulin pump data for 54 patients with type 1 diabetes. Grey Wolf optimization was used to tune and evaluate three machine learning algorithms – Random Forest, LSTM, GRU – for blood glucose predictions, whose predictions were then combined into an XGBoost stacking ensemble meta-learner.</div></div><div><h3>Results</h3><div>This study looked at three baseline algorithms for predicting blood glucose levels. Machine learning models Random Forest, LSTM, and GRU served as baselines, with MAE, RMSE, and MARD values. GRU had the best predictive accuracy of the initial models. Grey Wolf optimization contributed to achieving optimal baseline model results. Stacking ensemble learning via XGBoost meta-learner (MAE = 10.65, RMSE = 14.59, MARD = 6.98) achieved higher performance than the baseline models.</div></div><div><h3>Conclusion</h3><div>The GRU method with Grey Wolf optimization outperformed the other models with the lowest MAE, RMSE, and MARD, but the Stacked XGBoost model fared best. These findings emphasize the need to improve parameter selection with approaches such as Grey Wolf or stacking ensemble methods to achieve accurate blood glucose predictions. These prediction models can aid in the continued development of monitoring devices, and algorithms for these devices, which contain alert systems for impending abnormal blood glucose levels, allowing for timely diabetes self-management.</div></div>","PeriodicalId":34427,"journal":{"name":"Endocrine and Metabolic Science","volume":"18 ","pages":"Article 100253"},"PeriodicalIF":0.0000,"publicationDate":"2025-05-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Endocrine and Metabolic Science","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2666396125000391","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"Medicine","Score":null,"Total":0}
引用次数: 0
Abstract
Introduction
The diabetes pandemic, including 828 million adults worldwide in 2022, would benefit from continued development of novel, effective and accurate blood glucose prediction systems. Using the DiaTrend dataset, this study used stacking machine learning optimized by Grey Wolf Optimizer to construct and assess prediction models for blood glucose levels in type 1 diabetes patients.
Methods
The DiaTrend dataset includes 27,561 days of continuous glucose monitoring and 8220 days of insulin pump data for 54 patients with type 1 diabetes. Grey Wolf optimization was used to tune and evaluate three machine learning algorithms – Random Forest, LSTM, GRU – for blood glucose predictions, whose predictions were then combined into an XGBoost stacking ensemble meta-learner.
Results
This study looked at three baseline algorithms for predicting blood glucose levels. Machine learning models Random Forest, LSTM, and GRU served as baselines, with MAE, RMSE, and MARD values. GRU had the best predictive accuracy of the initial models. Grey Wolf optimization contributed to achieving optimal baseline model results. Stacking ensemble learning via XGBoost meta-learner (MAE = 10.65, RMSE = 14.59, MARD = 6.98) achieved higher performance than the baseline models.
Conclusion
The GRU method with Grey Wolf optimization outperformed the other models with the lowest MAE, RMSE, and MARD, but the Stacked XGBoost model fared best. These findings emphasize the need to improve parameter selection with approaches such as Grey Wolf or stacking ensemble methods to achieve accurate blood glucose predictions. These prediction models can aid in the continued development of monitoring devices, and algorithms for these devices, which contain alert systems for impending abnormal blood glucose levels, allowing for timely diabetes self-management.