Skip to main content

ORIGINAL RESEARCH article

Front. Pediatr., 31 January 2024
Sec. Children and Health
Volume 12 - 2024 | https://doi.org/10.3389/fped.2024.1330420

Predicting preterm birth using auto-ML frameworks: a large observational study using electronic inpatient discharge data

Deming Kong1,† Ye Tao1,† Haiyan Xiao1,† Huini Xiong1,† Weizhong Wei1* Miao Cai2*
  • 1Wuhan Children’s Hospital (Wuhan Maternal and Child Healthcare Hospital), Tongji Medical College, Huazhong University of Science and Technology, Wuhan, Hubei, China
  • 2Department of Epidemiology, School of Public Health, Sun Yat-sen University, Guangzhou, Guangdong, China

Background: To develop and compare different AutoML frameworks and machine learning models to predict premature birth.

Methods: The study used a large electronic medical record database to include 715,962 participants who had the principal diagnosis code of childbirth. Three Automatic Machine Learning (AutoML) were used to construct machine learning models including tree-based models, ensembled models, and deep neural networks on the training sample (N = 536,971). The area under the curve (AUC) and training times were used to assess the performance of the prediction models, and feature importance was computed via permutation-shuffling.

Results: The H2O AutoML framework had the highest median AUC of 0.846, followed by AutoGluon (median AUC: 0.840) and Auto-sklearn (median AUC: 0.820), and the median training time was the lowest for H2O AutoML (0.14 min), followed by AutoGluon (0.16 min) and Auto-sklearn (4.33 min). Among different types of machine learning models, the Gradient Boosting Machines (GBM) or Extreme Gradient Boosting (XGBoost), stacked ensemble, and random forrest models had better predictive performance, with median AUC scores being 0.846, 0.846, and 0.842, respectively. Important features related to preterm birth included premature rupture of membrane (PROM), incompetent cervix, occupation, and preeclampsia.

Conclusions: Our study highlights the potential of machine learning models in predicting the risk of preterm birth using readily available electronic medical record data, which have significant implications for improving prenatal care and outcomes.

Introduction

Preterm birth, defined as delivery before 37 weeks of gestation, is a major public health challenge that affects around 15 million babies worldwide each year (1). It is a leading cause of neonatal mortality and morbidity, as well as long-term health problems such as neurodevelopmental disabilities and chronic diseases (2). The causes of preterm birth are multifactorial and complex, and include maternal factors such as infections, stress, and chronic diseases, as well as fetal and environmental factors (3). Despite efforts to reduce the incidence of preterm birth, the rate has not significantly decreased in recent years. This highlights the need for a better understanding of the underlying mechanisms and risk factors, as well as improved prevention and management strategies.

Health service and outcome research utilizing administrative medical databases is becoming increasingly popular and gaining more attention (49). Administrative data are routinely gathered from various healthcare institutions such as hospitals, clinics, and pharmacies. These data are extensive and provide comprehensive service utilization information, which has led to a surge in researchers using them for cost-effectiveness analysis, risk adjustment, and mortality and health outcome prediction. Prior research primarily uses traditional statistical models such as generalized linear models to construction prediction models, which fails to capture the nonlinear and complex relationships between potential risk factors and the outcome, and this bottleneck in model predictive performance may overshadow potentially important risk factors or bias the importance of each factor.

With the recent advancements in artificial intelligence and the associated applications in the medical field, machine learning models are closer to generate innovative solutions for healthcare than ever. For example, a study utilizing electronic medical records from Vanderbilt Hospital in the United States, highlighted the substantial advantages of machine learning in improving healthcare throughout the prenatal period via its superior performance in accuracy, classification, and portability (10). Additionally, a recently published review comprehensively examined the role of artificial intelligent as a promising tool for clinicians facing daily obstetric challenges (11). The review highlighted the potential of advanced machine learning algorithms to analyze vast amounts of medical data, aiding in the identification of new risk factors associated with premature birth. However, in real world practice, there is limited research on constructing powerful machine learning models for administrative medical data to predict premature birth.

Several challenges, including feature engineering, hyperparameter tuning, model specification, and model evaluation, hinder the hands-on applications of accurate machine learning models in the medical research. Automatic machine learning (AutoML) aims to address these challenges and generate machine learning products by automating algorithm selection, hyperparameter tuning, model evaluation, and others (12). In this study, we utilized three open-source AutoML frameworks (AutoGluon, Auto-sklearn, and H2O) to predict preterm birth based on data from 715,962 women who were hospitalized for childbirth in China.

Related work

The existing research primarily focuses on using machine learning models for preterm birth prediction has significantly advanced our understanding in this field. Machine learning is a subdivision of artificial intelligence, possesses the capability to forecast patient clinical outcomes by extracting valuable features or predictor variables from data, and has been increasingly used in the prediction of preterm birth (13). For several decades, many researchers have used popular machine learning algorithms, including Support Vector Machine, K-Nearest Neighbors, and Convolutional Neural Networks. For instance, a study employed ordinary logistic regression, random forest, and KNN to identify risk factors for preterm birth in the USA. The study highlighted that a history of prior stillbirth, hypertension, and diabetes mellitus were significant risk factors for preterm birth (14).

A major limitation of these machine learning applications in preterm birth prediction is that they only smaller datasets (15) and this is partially from a lack of machine learning applications on electronic health records, which encompass millions of medical records, diagnoses, prescriptions, along with high-dimensional medical images (16). This study analyzes a large-scale electronic medical record database with 715,962 participants, which maximizes the statistical power to identify the risk factors associated with preterm birth in more depth, providing a scientific basis for evidence-based decision-making and guidelines.

Another notable limitation in earlier studies is the lack of applications on the vast repository of high-dimensional medical data available within electronic medical records. The current research acknowledges the untapped potential of these datasets and actively harnesses their breadth and depth to enhance the predictive modeling of preterm birth, enabling machine learning practitioners to derive meaningful insights and develop precise guidelines.

A common trend in previous studies involves heavy reliance on expert input for feature engineering, hyperparameter tuning, and model specification. This reliance can introduce subjectivity and make it challenging to reproduce superior performance. In contrast, the present study adopts a more automated approach. By leveraging three open-source AutoML frameworks—AutoGluon, Auto-sklearn, and H2O AutoML—the research automates critical steps such as algorithm selection, hyperparameter tuning, and model evaluation (17). The automation implemented in our study improves the reproducibility and scalability of the predictive models by boosting efficiency and reducing the reliance on extensive expert-defined phenotyping and ad-hoc feature engineering.

Methods

Data sources

This study used an administrative inpatient discharge dataset collected by the Health Commission of Shanxi Province, China from January 1, 2014 to December 31, 2017 (49). This database routinely collects data on individual's demographic information (age, sex, and ethnicity), socioeconomic status (marital status and occupation), disease severity, one main diagnosis code and up to 10 secondary diagnosis codes based on International Classification of Diseases, Tenth Revision (ICD-10). All female individuals with the main ICD-10 diagnosis code for pregnancy and childbirth (starting with “O”) were included in the study (18), yielding a final analytic sample of 715,962 participants. All individual and hospital identifiers, such as individual name, ID card number, and insurance card number, were excluded before the study team had access to the data. Observations with any missing data on the predictors were excluded from the study.

Outcomes

Preterm birth was defined as a binary variable, with the value of one assigned to preterm birth that occurred before 37 weeks of gestation (identified by ICD-10 code O60.1) was coded as one, while the value of zero assigned to all other births (19).

Predicting variables

We selected a few predicting variables (features) based on preexisting studies and data availability (2023). Age was computed as the difference between the date of hospitalization and the date of birth. We coded ethnicity as Han Chinese and non-Han Chinese (minorities), and biological sex as male or female. Socioeconomic status was assessed through two multi-category variables: marital status (married, unmarried, widowed, divorced, and other) and occupation (public sector, private sector, agriculture, unemployed, and other) (24). Admission status was classified as normal, emergent, or dangerous, while admission source was categorized as inpatient admission, emergency department transfer, outpatient department transfer, or transfer from other medical facilities. Payment method included the New Rural Cooperative Medical Scheme (NRCMS), the Urban Employee Basic Medical Insurance (UEBMI), the Urban Resident Basic Medical Insurance (URBMI), self-payment, or others. We also identified a set of clinical risk factors for preterm birth, including gestational diabetes, premature rupture of membranes, preeclampsia, incompetent cervix, and nuchal cord. The ICD-10 clinical diagnosis codes for these risk factors are listed in Supplementary Table S1.

Automl frameworks

AutoML can automate the process of algorithm optimization, hyperparameter tuning, model iteration, and model evaluation, and reduce the difficulty and time-consuming bottlenecks of machine learning models. We considered several popular AutoML frameworks for supervised machine learning. Auto-WEKA ceased development in March 2022. Auto-kera and Auto-PyTorch primarily focus on image pattern recognition and lack sufficient support for tabular classification data and its model evaluation metrics. Therefore, we did not use these three tools and this study chose three popular AutoML frameworks (AutoGluon, Auto-sklearn, and H2O AutoML).

Autogluon

A robust and accurate AutoML framework for structural data sets (25). We used AutoGluon-Tabular for binary classification for structural tabular dataset. It supports Graphics Processing Unit (GPU) training for most of its models including lightGBM, CatBoost, XGBoost, MXNet and FastAI Neural Networks (2628). AutoGluon framework for tabular data, also known as AutoGluon-Tabular, is designed to train highly accurate machine learning models via ensembling multiple models and stacking these models in multiple layers (25). Therefore, the models implemented in AutoGluon are not generic models such as generalized linear models. The philosophy of AutoGluon is to train a curated list of most powerful models within a reasonable amount of time, so only around 10 complex ensembling models are implemented. Less performant generic models such as generalized linear models are not fitted in AutoGluon and can be implemented in other autoML frameworks. Instead of focusing on combined algorithm selection and hyperparameter optimization (CASH) problem, AutoGluon replies more on advanced data processing, deep learning, and multi-layer stack ensembling to maximize prediction performance (25).

Auto-sklearn

Auto-sklearn is an AutoML tool using meta-learning (29). The aim of meta-learning is to search for a better start for hyperparameter tuning using efficient and parameterized Bayesian optimization method, so that the initial start-off values are better than random. It is based on an existing Python machine learning package scikit-learn and does not support GPU acceleration in the current version. This framework so far supports 15 classifiers, 14 methods for feature preprocessing, and 4 data preprocessing methods, resulting in a total of over 100 hyperparameter combinations.

H2O AutoML

H2O AutoML is an open-source, highly scalable, fully automated, and distributed supervised learning algorithm set based on Java (30). The H2O AutoML tools includes hundreds of efficient supervised machine learning algorithms such as generalized linear models (GLMs), XGBoost, GBM (Gradient Boosting Machine), random forest, GBMs, and deep neural nets. It does not support GPU acceleration except for XGBoost models. The H2O AutoML also introduces a convenient tool of leaderboard to allow users to compare the predictive performance, including k-fold cross-validation, across models fitted by H2O AutoML framework.

Table 1 presents the number of machine learning models by model type and automatic machine learning frameworks. In this study, we only select the best 10 models (based on model evaluation metrics) for each model type to reduce the imbalance number of models per framework. Auto-sklearn fits the largest number of models (N = 60) among the three frameworks, followed by H2O AutoML (N = 33) and AutoGluon (N = 12).

Table 1
www.frontiersin.org

Table 1. The number of machine learning models by model type and automatic machine learning frameworks.

Model evaluation metrics

To compare the model performance across different AutoML tools using a unified metric, we chose area under curve (AUC) for receiver operating characteristic curves as the model evaluation metric. The ROC curve depicts the relationship between the true positive and false positive rates while selecting cut-off values for predicting binary outcomes such as preterm birth (31). AUC, also known as the concordance statistic or c-statistics, is a measure of goodness of fit for binary classification models. It ranges between 0.5 and 1, where larger values indicate better prediction performance.

Model interpretability was evaluated as feature importance computed via permutation-shuffling (32). The feature importance scores measure the decrease in the predictive performance of a trained model when the values of a feature are randomly shuffled across rows. The features with high importance scores have the most significant contribution to prediction accuracy. These feature importance scores aid in interpreting the model's overall prediction performance by identifying the prioritized features it relies on for predictions.

Computing environment and model setup

The original data (N = 715,962) were split into a training set (N = 536,971) and a test set (N = 178,991) based on the 75/25 criteria. We did not set up a validation data set as the hyperparameters were tuned using 10-fold cross validation on the training set. All the training, testing, and model evaluation were performed in a high-performance computing environment (96 threads and 512 GB RAM), with Python 3.9 installed on Linux CentOS 7.8 and one NVIDIA Tesla T4 enabled for GPU computing (CUDA version 11.7). The training was performed with a time limit of 24 h. The versions of the AutoML tools were AutoGluon (v0.7.0), Auto-sklearn (v0.15.0), and H2O AutoML (v3.40.0.2). The data and associated Python code to replicate the results can be found in Supplementary Materials.

Results

Descriptive statistics

Table 2 presents the descriptive features of patients, including their demographics, covariates, and health characteristics. In total, 715,962 participants were included in the study, with an average [standard deviation, (SD)] age of 28.4 (5.0) years. The overwhelming majority (99.9%) of the participants were of Han Chinese ethnicity. Of the total participants, 45,763 (6.39%) had a preterm birth, and they tended to be younger and have a higher prevalence of gestational diabetes, premature rupture of membranes, preeclampsia, and incompetent cervix.

Table 2
www.frontiersin.org

Table 2. Overall participant characteristics and by preterm birth.

Predictive performance of machine learning models

To assess the predictive performance of machine learning models used in this study, we employed AUC as the evaluation metric. Results are presented in Figure 1 and Table 3, showing the AUC statistics for different models and AutoML frameworks. Our analyses revealed that GBM/XGBoost, Stacked Ensemble and random forrest/trees models performed better, with median AUC scores of 0.846, 0.846 and 0.842, respectively. Conversely, KNN, Naive Bayes, and LDA models achieved lower AUC scores with medians of 0.725, 0.791 and 0.818 respectively, which are even lower than a traditional logistic regression (AUC = 0.823). H2O AutoML had the best median predictive ability among the three frameworks, achieving a median AUC of 0.846, followed by AutoGluon (AUC = 0.840) and H2O AutoML (AUC = 0.820). It is notable that the maximum AUCs of the three H2O modeling frameworks were identical (AUC = 0.847).

Figure 1
www.frontiersin.org

Figure 1. Area under the curve (AUC) for different AutoML frameworks and machine learning models. (A) Raincloud plot of the area under the curve (AUC) for three AutoML frameworks (Auto-sklearn, AutoGluon, and H2O AutoML). Each raincloud plot panel consists of three components: a jittered dot plot on the left side, a boxplot in the middle, and a cloud plot of the distribution of AUCs on the right side. (B) Boxplots of AUCs by machine learning models. GBM: Gradient Boosting Machines; GLM: Generalized Linear Models; KNN: K-Nearest Neighbors; LDA: Linear Discriminant Analysis.

Table 3
www.frontiersin.org

Table 3. Summary statistics for area under the curve (AUC) for different autoML frameworks and machine learning models.

Training time

Figure 2 and Table 4 shows that H2O AutoML had the lowest median training time (0.14 min) among the evaluated AutoML frameworks, followed by AutoGluon AutoML (median training time: 0.16 min) and Auto-sklearn (median training time: 4.33 min). Regarding machine learning models, GLM had the fastest training times, taking less time than Stacked Ensemble, Deep Learning, Naive Bayes, LDA, GBM/XGBoost, Random Forest/Trees and KNN.

Figure 2
www.frontiersin.org

Figure 2. Training time in minutes for different AutoML frameworks and machine learning models. (A) Raincloud plot of training time in minutes (training set sample size N = 536,971) for three AutoML frameworks (Auto-sklearn, AutoGluon, and H2O AutoML). Each raincloud plot panel consists of three components: a jittered dot plot on the left side, a boxplot in the middle, and a cloud plot of the distribution of AUCs on the right side. (B) Boxplots of training time in minutes by machine learning models. GBM: Gradient Boosting Machines; GLM: Generalized Linear Models; KNN: K-Nearest Neighbors; LDA: Linear Discriminant Analysis.

Table 4
www.frontiersin.org

Table 4. Summary statistics for training time in minutes for different autoML frameworks and machine learning models (training set sample size N = 536,971).

Overall, when it comes to selecting an AutoML framework or machine learning model, the training time is an important factor to consider. The choice of framework or model should be based on the specific requirements of the task at hand, including the available computational resources and the desired level of accuracy.

Variable importance

Figure 3 illustrates the important features for predicting preterm birth identified by AutoGluon frameworks. We utilized permutation feature importance to prioritize the variables by their importance in the FastAl Neural Net model. The top five variables were payment methods, premature rupture of membrane (PROM), incompetent cervix, occupation, and preeclampsia. These variables had a more substantial influence on the model's accuracy compared to the other variables in the dataset.

Figure 3
www.frontiersin.org

Figure 3. Overall feature importance (95% confidence intervals) plots for predicting preterm birth via permutation-shuffling in AutoGluon. PROM: premature rupture of membranes.

Discussion

Preterm birth is the primary cause of neonatal death and disability, and early prediction of preterm birth has great potential to improve the survival rate of preterm infants (33). In this study, electronic medical record data from the hospital discharge database were selected to construct machine learning model to predict preterm birth. We found that payment methods, PROM, incompetent cervix, occupation, and preeclampsia were strongly associated with preterm birth patients.

AutoML has increasingly gain popularity since it substantially reduces the difficulty of building a machine learning pipeline via complex, iterative, and time-consuming technical details including hyperparameter tuning, algorithm selection, or model evaluation. All these processes are automated within a few lines of code, so it saves the effort and time of an expensive machine learning engineer. Our findings revealed that H2O AutoML, surpassing AutoGluon and Auto-sklearn, achieved reasonably well predictive performance within a short amount of time (median train time per model was 0.1 min for over half a million rows and 12 features).

This study confirmed that PROM and incompetent cervix are significant risk factors for preterm birth, consistent with previous research (22, 34). Several mechanisms may explain these associations. Intrauterine infection or inflammation is believed to be a major contributing factor to preterm birth PROM, subsequent preterm birth (35). The cervix is normally closed during pregnancy, but in cases of incompetent cervix, the cervix may dilate too early, triggering contractions and labor, which may lead to preterm birth (36).

The findings in our study show that payment method is associated with increased risk of preterm birth. The association between payment method and preterm birth is a complex issue. Research has revealed that socioeconomic status can be a contributing factor to preterm birth, with women who lack insurance or have public insurance being more susceptible to preterm birth compared to those with private insurance (37). One possible explanation for this correlation is that women with lower incomes may encounter more obstacles in accessing healthcare services, including prenatal care, which can elevate the risk of preterm birth. Additionally, women with public insurance may have limited choices in healthcare providers and receive less comprehensive care. However, it is important to note that the relationship between payment method and preterm birth is not entirely straightforward, and other factors, such as age, race/ethnicity, and marital status, may also play a role. More research is needed to comprehensively understand the complex factors.

We also observed the association between occupation and the risk of preterm birth. Specifically, public sectors are more susceptible to preterm birth than other occupations in China. One possible reason for this is that the job of a public sector is demanding and requires them to be constantly engaged in decision-making, multitasking, and dealing with high levels of stress. This pressure can take a toll on their mental health, leading to anxiety, depression, and other mental health issues, which can lead to preterm birth.

Strengths and limitations

The main strength of this study is the large sample size of over 700,000 women in Shanxi province of China, which enhances the reliability and statistical power of the models conducted in this study. Additionally, we evaluate three metrics including the predictive performance, training times, and feature importance, which could give medical practitioners and machine learning engineers a more practical guidelines on choosing the most appropriate tool to work with. Third, we compared the results of different AutoML frameworks and machine learning models in this study, allowing us to precisely identify risk factors for preterm birth.

Several limitations should be noted. First, our study is an observational study that depends on secondary data, it is subject to potential information bias and residual confounding caused by inaccurate coding, hospital characteristics or unobserved patient. Second, a predominate proportion of the study participants were Han Chinese, thus the results cannot be generalized to populations with different ancestries. Third, since the three autoML frameworks were independently developed by different teams with different optimizing philosophies, the type and number of models implemented in each framework are not directly comparable.

Conclusion

We employed several popular AutoML frameworks and machine learning to analyze a large Chinese electronic medical record database, assessing the risk factors associated with the risk of preterm birth. Our findings have the potential to screen high-risk populations for preterm birth in China, which can help doctors tailor treatments for pregnant women with different risks of preterm birth.

Data availability statement

The code and data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Author contributions

DK: Conceptualization, Data curation, Formal Analysis, Investigation, Validation, Writing – review & editing. YT: Data curation, Formal Analysis, Methodology, Project administration, Validation, Writing – review & editing. HaX: Data curation, Methodology, Project administration, Software, Supervision, Validation, Writing – review & editing. HuX: Data curation, Methodology, Project administration, Resources, Software, Supervision, Validation, Writing – review & editing. WW: Conceptualization, Project administration, Resources, Supervision, Writing – review & editing. MC: Conceptualization, Data curation, Formal Analysis, Funding acquisition, Visualization, Writing – original draft.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article.

This work was supported by the Guangdong Basic and Applied Basic Research Foundation (grant number: 2022A1515110995).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fped.2024.1330420/full#supplementary-material

References

1. Cao G, Liu J, Liu M. Global, regional, and national incidence and mortality of neonatal preterm birth, 1990–2019. JAMA Pediatr. (2022) 176(8):787–96. doi: 10.1001/jamapediatrics.2022.1622

PubMed Abstract | Crossref Full Text | Google Scholar

2. Walani SR. Global burden of preterm birth. Int J Gynaecol Obstet. (2020) 150(1):31–3. doi: 10.1002/ijgo.13195

PubMed Abstract | Crossref Full Text | Google Scholar

3. Vogel JP, Chawanpaiboon S, Moller AB, Watananirun K, Bonet M, Lumbiganon P. The global epidemiology of preterm birth. Best Pract Res Clin Obstet Gynaecol. (2018) 52:3–12. doi: 10.1016/j.bpobgyn.2018.04.003

PubMed Abstract | Crossref Full Text | Google Scholar

4. Cai M, Lin X, Wang X, Zhang S, Wang C, Zhang Z, et al. Long-term exposure to ambient fine particulate matter chemical composition and in-hospital case fatality among patients with stroke in China. The Lancet Reg Health West Pac. (2023) 32:1–13. doi: 10.1016/j.lanwpc.2022.100679

Crossref Full Text | Google Scholar

5. Cai M, Zhang S, Lin X, Qian Z, McMillin SE, Yang Y, et al. Association of ambient particulate matter pollution of different sizes with in-hospital case fatality among stroke patients in China. Neurology. (2022) 98:e2474–86. doi: 10.1212/WNL.0000000000200546

Crossref Full Text | Google Scholar

6. Cai M, Liu E, Tao H, Qian Z, Fu QJ, Lin X, et al. Does a medical consortium influence health outcomes of hospitalized cancer patients? An integrated care model in shanxi. China. Int J Integr Care. (2018) 18(2):7. doi: 10.5334/ijic.3588

Crossref Full Text | Google Scholar

7. Cai M, Liu E, Tao H, Qian Z, Lin X, Cheng Z. Does level of hospital matter? A study of mortality of acute myocardial infarction patients in Shanxi, China. Am J Med Qual. (2018) 33(2):185–92. doi: 10.1177/1062860617708608

PubMed Abstract | Crossref Full Text | Google Scholar

8. Lin X, Cai M, Tao H, Liu E, Cheng Z, Xu C, et al. Insurance status, inhospital mortality and length of stay in hospitalised patients in Shanxi, China: a cross-sectional study. BMJ Open. (2017) 7(7):e015884. doi: 10.1136/bmjopen-2017-015884

PubMed Abstract | Crossref Full Text | Google Scholar

9. Cai M, Liu E, Bai P, Zhang N, Wang S, Li W, et al. The chasm in percutaneous coronary intervention and in-hospital mortality rates among acute myocardial infarction patients in rural and urban hospitals in China: a mediation analysis. Int J Public Health. (2022) 67:1–10. doi: 10.3389/ijph.2022.1604846

Crossref Full Text | Google Scholar

10. Abraham A, Le B, Kosti I, Straub P, Velez-Edwards DR, Davis LK, et al. Dense phenotyping from electronic health records enables machine learning-based prediction of preterm birth. BMC Med. (2022) 20(1):333. doi: 10.1186/s12916-022-02522-x

PubMed Abstract | Crossref Full Text | Google Scholar

11. Sarno L, Neola D, Carbone L, Saccone G, Carlea A, Miceli M, et al. Use of artificial intelligence in obstetrics: not quite ready for prime time. Am J Obstet Gynecol MFM. (2023) 5(2):100792. doi: 10.1016/j.ajogmf.2022.100792

PubMed Abstract | Crossref Full Text | Google Scholar

12. Ferreira L, Pilastri A, Martins CM, Pires PM, Cortez P. A comparison of AutoML tools for machine learning, deep learning and XGBoost. IEEE. (2021):1–8. doi: 10.1109/IJCNN52387.2021.9534091

Crossref Full Text | Google Scholar

13. Zhang Y, Du S, Hu T, Xu S, Lu H, Xu C, et al. Establishment of a model for predicting preterm birth based on the machine learning algorithm. BMC Pregnancy Childbirth. (2023) 23(1):779. doi: 10.1186/s12884-023-06058-7

PubMed Abstract | Crossref Full Text | Google Scholar

14. Weber A, Darmstadt GL, Gruber S, Foeller ME, Carmichael SL, Stevenson DK, et al. Application of machine-learning to predict early spontaneous preterm birth among nulliparous non-hispanic black and white women. Ann Epidemiol. (2018) 28(11):783–789 e1. doi: 10.1016/j.annepidem.2018.08.008

PubMed Abstract | Crossref Full Text | Google Scholar

15. Akazawa M, Hashimoto K. Prediction of preterm birth using artificial intelligence: a systematic review. J Obstet Gynaecol. (2022) 42(6):1662–8. doi: 10.1080/01443615.2022.2056828

PubMed Abstract | Crossref Full Text | Google Scholar

16. Xiao C, Choi E, Sun J. Opportunities and challenges in developing deep learning models using electronic health records data: a systematic review. J Am Med Inform Assoc. (2018) 25(10):1419–28. doi: 10.1093/jamia/ocy068

PubMed Abstract | Crossref Full Text | Google Scholar

17. Cai M, Yazdi MAA, Mehdizadeh A, Hu Q, Vinel A, Davis K, et al. The association between crashes and safety-critical events: synthesized evidence from crash reports and naturalistic driving data among commercial truck drivers. Transport Res C-Emer. (2021) 126:1–19. doi: 10.1016/j.trc.2021.103016

Crossref Full Text | Google Scholar

18. Xu C, Zhong W, Fu Q, Yi L, Deng Y, Cheng Z, et al. Differential effects of different delivery methods on progression to severe postpartum hemorrhage between Chinese nulliparous and multiparous women: a retrospective cohort study. BMC Pregnancy Childbirth. (2020) 20(1):660. doi: 10.1186/s12884-020-03351-7

PubMed Abstract | Crossref Full Text | Google Scholar

19. Sarayani A, Wang X, Thai TN, Albogami Y, Jeon N, Winterstein AG. Impact of the transition from ICD-9-CM to ICD-10-CM on the identification of pregnancy episodes in US health insurance claims data. Clin Epidemiol. (2020) 12:1129–38. doi: 10.2147/CLEP.S269400

PubMed Abstract | Crossref Full Text | Google Scholar

20. Howson CP, Kinney MV, McDougall L, Lawn JE. Born too soon preterm birth action G. Born too soon: preterm birth matters. Reprod Health. (2013) 10 Suppl 1(Suppl 1):S1. doi: 10.1186/1742-4755-10-S1-S1

PubMed Abstract | Crossref Full Text | Google Scholar

21. Dean SV, Mason E, Howson CP, Lassi ZS, Imam AM, Bhutta ZA. Born too soon: care before and between pregnancy to prevent preterm births: from evidence to action. Reprod Health. (2013) 10 Suppl 1(Suppl 1):S3. doi: 10.1186/1742-4755-10-S1-S3

PubMed Abstract | Crossref Full Text | Google Scholar

22. Goldenberg RL, Culhane JF, Iams JD, Romero R. Epidemiology and causes of preterm birth. Lancet. (2008) 371(9606):75–84. doi: 10.1016/S0140-6736(08)60074-4

PubMed Abstract | Crossref Full Text | Google Scholar

23. Muglia LJ, Katz M. The enigma of spontaneous preterm birth. N Engl J Med. (2010) 362(6):529–35. doi: 10.1056/NEJMra0904308

PubMed Abstract | Crossref Full Text | Google Scholar

24. Cai M, Lin X, Wang X, Zhang S, Qian ZM, McMillin SE, et al. Ambient particulate matter pollution of different sizes associated with recurrent stroke hospitalization in China: a cohort study of 1.07 million stroke patients. Sci Total Environ. (2023) 856(Pt 2):159104. doi: 10.1016/j.scitotenv.2022.159104

PubMed Abstract | Crossref Full Text | Google Scholar

25. Shirkov A, Zhang H, Larroy P, Li M, Smola A. AutoGluon-Tabular: Robust and Accurate AutoML for Structured Data. 2020).

26. Ke G, Meng Q, Finley T, Wang T, Chen W, Ma W, et al. Lightgbm: a highly efficient gradient boosting decision tree. Adv Neural Inf Process Syst. (2017):30.

Google Scholar

27. Prokhorenkova L, Gusev G, Vorobev A, Dorogush AV, Gulin A. Catboost: unbiased boosting with categorical features. Adv Neural Inf Process Syst. (2018):31.

Google Scholar

28. Chen T, Guestrin C. Xgboost: a scalable tree boosting system. Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. New York: Association for Computing Machinery (2016). p. 785–94. doi: 10.1145/2939672.2939785

Crossref Full Text | Google Scholar

29. Feurer M, Eggensperger K, Falkner S, Lindauer M, Hutter F. Auto-sklearn 2.0: hands-free automl via meta-learning. J Machine Learn Res. (2020) 23(261):1–61. doi: 10.48550/ARXIV.2007.04074

Crossref Full Text | Google Scholar

30. LeDell E, Poirier S. H2O AutoML: Scalable Automatic Machine Learning. 2020).

31. Cai M, Liu E, Zhang R, Lin X, Rigdon SE, Qian Z, et al. Comparing the performance of charlson and elixhauser comorbidity indices to predict in-hospital mortality among a Chinese population. Clin Epidemiol. (2020) 12:307–16. doi: 10.2147/CLEP.S241610

PubMed Abstract | Crossref Full Text | Google Scholar

32. Altmann A, Tolosi L, Sander O, Lengauer T. Permutation importance: a corrected feature importance measure. Bioinformatics. (2010) 26(10):1340–7. doi: 10.1093/bioinformatics/btq134

PubMed Abstract | Crossref Full Text | Google Scholar

33. Care A, Nevitt SJ, Medley N, Donegan S, Good L, Hampson L, et al. Interventions to prevent spontaneous preterm birth in women with singleton pregnancy who are at high risk: systematic review and network meta-analysis. Br Med J. (2022) 376:e064547. doi: 10.1136/bmj-2021-064547

Crossref Full Text | Google Scholar

34. Menon R, Richardson LS. Preterm prelabor rupture of the membranes: a disease of the fetal membranes. Semin Perinatol. (2017) 41(7):409–19. doi: 10.1053/j.semperi.2017.07.012

PubMed Abstract | Crossref Full Text | Google Scholar

35. Romero R, Espinoza J, Goncalves LF, Kusanovic JP, Friel L, Hassan S. The role of inflammation and infection in preterm birth. Semin Reprod Med. (2007) 25(1):21–39. doi: 10.1055/s-2006-956773

PubMed Abstract | Crossref Full Text | Google Scholar

36. Koullali B, Westervelt AR, Myers KM, House MD. Prevention of preterm birth: novel interventions for the cervix. Semin Perinatol. (2017) 41(8):505–10. doi: 10.1053/j.semperi.2017.08.009

PubMed Abstract | Crossref Full Text | Google Scholar

37. Montemor MS, Demarque GF, Rodrigues AS, Francisco RPV, de Carvalho MHB. Association between preterm births and socioeconomic development: analysis of national data. BMC Public Health. (2022) 22(1):2014. doi: 10.1186/s12889-022-14376-2

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: preterm birth, machine learning, administrative data, China, autoML

Citation: Kong D, Tao Y, Xiao H, Xiong H, Wei W and Cai M (2024) Predicting preterm birth using auto-ML frameworks: a large observational study using electronic inpatient discharge data. Front. Pediatr. 12:1330420. doi: 10.3389/fped.2024.1330420

Received: 30 October 2023; Accepted: 16 January 2024;
Published: 31 January 2024.

Edited by:

Tetyana Chumachenko, Kharkiv National Medical University, Ukraine

Reviewed by:

Ricardo Valentim, Federal University of Rio Grande do Norte, Brazil
Qutaiba Saleh, Rochester Institute of Technology (RIT), United States

© 2024 Kong, Tao, Xiao, Xiong, Wei and Cai. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Weizhong Wei weiweizhong@zgwhfe.com Miao Cai miao.cai@outlook.com

These authors have contributed equally to this work and share first authorship

Abbreviations AUC, area under the curve; AutoML, automatic machine learning; GBM, gradient boosting machines; GPU, graphics processing unit; ICD-10, international classification of diseases, tenth revision; LDA, linear discriminant analysis; NRCMS, new rural cooperative medical scheme; PROM, premature rupture of membrane; SD, standard deviation; UEBMI, the urban employee basic medical insurance; URBMI, the urban resident basic medical insurance; XGBoost, extreme gradient boosting.

Download