An explainable machine learning based prediction model for Alzheimer's disease in China longitudinal aging study

Alzheimer's disease (AD) is the most common cause of dementia. Accurate prediction and diagnosis of AD and its prodromal stage, i.e., mild cognitive impairment (MCI), is essential for the possible delay and early treatment for the disease. In this paper, we adopt the data from the China Longitudinal Aging Study (CLAS), which was launched in 2011, and includes a joint effort of 15 institutions all over the country. Four thousand four hundred and eleven people who are at least 60 years old participated in the project, where 3,514 people completed the baseline survey. The survey collected data including demographic information, daily lifestyle, medical history, and routine physical examination. In particular, we employ ensemble learning and feature selection methods to develop an explainable prediction model for AD and MCI. Five feature selection methods and nine machine learning classifiers are applied for comparison to find the most dominant features on AD/MCI prediction. The resulting model achieves accuracy of 89.2%, sensitivity of 87.7%, and specificity of 90.7% for MCI prediction, and accuracy of 99.2%, sensitivity of 99.7%, and specificity of 98.7% for AD prediction. We further utilize the SHapley Additive exPlanations (SHAP) algorithm to visualize the specific contribution of each feature to AD/MCI prediction at both global and individual levels. Consequently, our model not only provides the prediction outcome, but also helps to understand the relationship between lifestyle/physical disease history and cognitive function, and enables clinicians to make appropriate recommendations for the elderly. Therefore, our approach provides a new perspective for the design of a computer-aided diagnosis system for AD and MCI, and has potential high clinical application value.


. Introduction
Alzheimer's disease (AD) is the most common dementia in the elderly, which is a slow and lengthy progressive neurodegenerative disorder and accounts for 60-80% of dementia cases.The population of AD is projected to reach 106.8 million by 2050 (Brookmeyer et al., 2007).Although numerous therapies have been investigated, there has been no successful trial that can modify the course of the disease.On the other hand, according to the data from epidemiologic studies and clinical trials, it has indicated that early intervention may delay the AD progression (Brookmeyer et al., 1998;Norton et al., 2014;Ngandu et al., 2015).The prodromal stage of AD, termed as mild cognitive impairment (MCI), can lead to cognitive decline and has a high risk to develop AD.Thus, accurate prediction and diagnosis of AD and MCI are very critical for the prevention and therapy of the disease.
Previous studies reported that some clinical and demographic features are considered to have strong predictive abilities (Livingston et al., 2017).However, none of them is strong enough to differentiate AD/MCI among the community elderly independently.It is more likely that clinical/demographic features may have complex relationships and, as a whole, jointly predict AD progression.Hence, the artificial intelligence (AI) approach may be a suitable way to combine these data to solve the problem.
Recently, many researchers have applied AI techniques for AD prediction.Zhang et al. (2019) propose a deep learning approach based on two convolutional neural networks (CNN) and multimodal medical images.Then correlation analysis is applied to judge the consistency of the output of the two CNN.Salvatore et al. (2015) extract features from MRI data using principal component analysis, and apply a machine learning algorithm to predict whether MCI patients will convert to AD. Loddo et al. (2022) presents a deep learning approach for Alzheimer's disease diagnosis using brain images.It compares different deep learning models and proposes a fully automated deep-ensemble approach for dementia-level classification.Discusses the challenges in detecting Alzheimer's disease (AD) in its early stages and reviews the current research on machine learning techniques for its detection and classification, with a focus on neuroimaging.The review suggests that deep learning techniques hold promise for AD diagnosis, and new algorithms have yet to be explored for AD diagnosis.These studies above apply various machine learning and deep learning methods to predict AD based on data from different modalities.But they only focus on the models' performance while neglecting the interpretation of the output of these models.
The following studies not only design a new model, but also analyze the output of the model.El-Sappagh et al. (2021) develop a two-layer model with random forest (RF), and use the SHapley Additive exPlanations (SHAP) framework to make overall and individual explanations for the result of each layer.Additionally, 22 explainers are developed based on decision tree and fuzzy rulebased systems to provide supplementary justifications for every RF decision in each layer.Danso et al. (2021) develop a framework that integrates transfer learning and ensemble learning algorithms to develop explainable personalized risk prediction models for dementia, SHAP is used to visualize the risk factors responsible for the prediction.
In this paper, we adopt the data from China Longitudinal Aging Study (CLAS;Haibo et al., 2013;Xiao, 2013;Xiao et al., 2016), which is a community-based cohort study launched in 2011.The project was conducted jointly by 15 institutions located in eastern, middle, and western parts of China.A total of 4,411 people at least 60 years old participated in the project, where 3,514 people completed the baseline survey.The survey collected data including demographic information, daily lifestyle, medical history, and routine physical examination.In addition, a variety of psychological and psychosocial measures were assessed by psychologists.A normal diagnostic method was adopted to classify the cognitive condition of all subjects, i.e., normal control (NC), MCI, or AD.
Based on this data, we aim to propose a joint detection of interpretable machine learning models and predictive indicators for predicting AD and MCI as follows: 1) we have processed the missing and default values in the data through a unified arrangement and data cleaning steps as part of data preprocessing.At the same time, we compared five feature selection methods to reduce the dimensionality of the data and the complexity of the model calculation.Finally, we used nine classifiers of general interpretable machine learning for classification comparison.2) based on previous research, our dataset includes more comprehensive information, including lifestyle, physical diseases, and medical check-up results.To our knowledge, this is the first work that aims to predict cognitive status using large-scale and multi-faceted information, especially detailed lifestyle and clinical information. .

Materials and methods
This section describes the details of our proposed system.As shown in Figure 1, it has four stages, which are data preprocessing, feature engineering, classification, and explanation, respectively.The framework also displays the methods adopted in these stages.The detailed introduction of every stage can be found in the following subsections.

. . Study participants and data collection
The population of our study is a community-based cohort study, named CLAS.The Chinese Longitudinal Aging Study (CLAS) was designed to provide information about the cognitive, mental, and psychosocial health of older people in China (Xiao, 2013).This survey was a joint effort of 15 institutions located in the eastern, middle, and western parts of China.The sample was randomly selected from all permanent residents aged over 60 in the 2010 national census (Xiao, 2013).
As reported in the above protocol (Haibo et al., 2013;Xiao, 2013;Xiao et al., 2016).These clinical diagnoses were made according to accepted criteria and with consideration of comorbid conditions.MCI was classified using the Petersen criteria (Petersen et al., 2001) and AD dementia were diagnosed according to the DSM-IV criteria (American Psychiatric Association, 2000), both of which were clinically diagnosis.
Out of the 3,514 participants who completed the survey, a total of 2,658 people had cognitive condition results, which includes 98 individuals (3.69%) with AD, 556 individuals (20.92%) with MCI, and 2,004 individuals with NC.
The dataset has 53 features, including demographic information, daily lifestyle, medical history, and routine physical examination.Tables 1, 2 shows the summary results of the standard deviation, mean, and interquartile range (IQR) of every feature for three classes.

. Missing value
In the dataset, most features have missing values, yet the missing rate is low (<7%).For the features with missing values, we first treat a feature with missing value as a new tag, and the remaining features and original tags form new input values.Then we apply a random forest algorithm to predict the missing values in the new tag (Liaw and Wiener, 2002).All features are filled up in turn following the steps above.

. . . Data augmentation
In our dataset, the number of AD (98 samples) and MCI (556 samples) are far less than that of NC (2,004 samples).The data imbalance may seriously degrade the performance of the machine learning algorithm.For example, overfitting may occur due to the imbalanced training data.We use the adaptive synthetic sampling approach (ADASYN) to handle the issue (He et al., 2008).ADASYN can adaptively generate samples for the minority class based on its distribution.

. . . Data normalization
In the dataset, every feature has a different value range.This may lead to unreasonable results, since the feature with larger values will have higher weights on the learned model.Thus, it is necessary to use data normalization to mitigate this effect.Max-min normalization is applied to each feature, which can be expressed by where X is the standard deviation of X, given by . .Feature selection There are a total of 53 features in the dataset, including demographics, daily lifestyle, medical history, and routine physical examination as shown in Tables 1, 2. Dimensionality reduction is a fundamental requirement for achieving simplicity and assessing the complexity of the model.The curse of dimensionality can adversely impact the model in terms of runtime and exhaustion of storage resources, particularly for non-scalable classifiers.For these reasons, we need to use the feature selection methods.
Feature selection is the preprocessing step before applying the classifier, which aims to eliminate unrelated and redundant features while preserving the key information of the original dataset by selecting the representative features.

. . Machine learning algorithm
Several machine learning models are compared to select the best classifier for AD/MCI prediction, which include three basic classifiers and six ensemble classifiers.
Ensemble learning is an integrated approach that can combine multiple base learners to achieve better performance, where many machine learning algorithms can be applied as the base learners, such as DT, neural network, etc.The base learners can be generated by two styles, i.e., the parallel style and the sequential style.Then all base learners will be combined to form a better learner, where the most common combination schemes are the majority for voting and weighted averaging for regression.To find out the best classifier, six ensemble learning methods are applied for comparison, which are Adaptive Boosting (AdaBoost; Freund and Schapire, 1996), eXtreme Gradient Boosting (XGBoost; Chen and Guestrin, 2016), Light Gradient Boosting Machine (LightGBM; Ke et al., 2017), Bootstrap Aggregation (Bagging;Breiman, 1996), Random Forest (RF;Breiman, 1996), and Extra Tree (ET; Geurts et al., 2006), respectively.

. . Model explainer
SHapley Additive exPlanation (SHAP) is a game-theoretic approach to explain the output of any machine learning model, which is proposed by Lundberg and Lee (2017).The goal of SHAP is to explain the prediction of a sample x i by computing the influence score of each feature to the prediction.The prediction y i can be expressed as follows: where y base is the average of predictions of all samples, and f (x ij ) is the SHAP value of x ij which is the contribution of j-th feature to the prediction of x i .When f (x ij ) > 0, the j-th feature can boost the prediction, otherwise, it has a negative effect.Compared with traditional measurement of feature importance, the strength of SHAP is that it can reflect the specific contribution of each feature to the model's output.

. . Performance metrics
To evaluate the model's performance, we use five performance criteria : Accuracy, Sensitivity, Specificity, G-mean, and Area Under Curve (AUC).Accuracy is the ratio between the correctly classified samples and all samples, which is defined as where TP represents the true positive, TN represents the true negative, FP represents the false positive, and FN represents the false negative.Since the data imbalance exists among different classes, accuracy is not enough for evaluating performance.It may cause misleading if the model only predicts the majority class correctly while neglecting the minority class.It is necessary to use more metrics to evaluate the performance of each class.Sensitivity is a measure of how well a model can predict for positive samples, and specificity is a metric of how well a model can predict for negative samples.The definition of the above metrics are shown as follows: G-mean is a reliable metric in the situation of overfitting the negative class and underfitting the positive class.As shown in Equation ( 7), it combines the sensitivity and specificity into a single score to balance both concerns.A model has a high G-mean, meaning that a classifier is not biased toward any of the classes (Kotsiantis et al., 2006).
AUC is another helpful metric to evaluate how effective the classifier is in separating different classes.The receiver operating characteristic curve (ROC) plots the Sensitivity against the 1 − Specificity at various threshold settings, where the area of 1 indicates the model is excellent, and the area of 0.5 denotes it is a worthless model.

. Experiments and results
In this section, we commence by revisiting hyperparameter optimization techniques and other pertinent works, followed by a detailed exposition of our specific experimental setup.

. . Hyperparameter optimization
In facing a plethora of algorithms mentioned within, each bearing distinct types of hyperparameters, the impact on model performance may manifest in varying manners.Take the Random Forest algorithm as an instance, wherein the number of estimators and the depth of trees serve as hyperparameters, casting a profound influence upon the model's performance.Currently, hyperparameter tuning can predominantly be categorized into the following methods: Grid Search: A classic technique diligently employed by examining all plausible parameter combinations.Grid Search contemplates the entire parameter space, partitioning it into a grid-like structure, where each point within the grid is evaluated as a hyperparameter (Bergstra and Bengio, 2012;Shekhar et al., 2021).This near-exhaustive optimization approach is apt for low-dimensional hyperparameter spaces, albeit our classifier algorithms necessitate multi-dimensional space optimization.Random Search: Randomized selection of hyperparameters marks the hallmark of this method, offering simplicity in implementation, yet challenging the adjustment of model hyperparameters based on commonality (Bergstra and Bengio, 2012).Bayesian Hyperparameter Optimization: Adopting a Bayesian rules, this method fine-tunes the evaluation function through posterior distribution, markedly reducing the search process within the parameter space (Dewancker et al., 2016).In our experimental setup, we employed the HyperOpt library (Bergstra et al., 2013) for hyperparameter optimization.HyperOpt operates on the paradigm of Sequential Model-Based Optimization (SMBO; Hutter et al., 2011), with the Tree of Parzen Estimators (TPE) orchestrating the search within the current space.
The parameter search space across different classifiers is delineated in Table 3, showcasing the breadth and scope entailed in tuning the hyperparameters to adeptly tailor the model to our dataset.

. . System setup and implementation
We developed our framework using the Python 3.6 environment.Essential libraries employed in our study included scipy, matplotlib, pandas, sklearn, Hyperopt, and numpy.The computational experiments were conducted on a laptop equipped with an Intel Core i5-10310U CPU and 16 GB of RAM.The simulations demanded ∼10 h to produce the outcomes.
The experiments are carried out based on the K-fold crossvalidation technique with K=10.The dataset is divided into K subsets, where each subset is treated as a testing set in turn, while the rest of the data is used to train the model.Then the final result is the average of these K results.This method guarantees the training and testing processes are both applied to the whole dataset.During the generation of each fold, stratified sampling is also applied to ensure that the proportion of samples of each class in the training and testing sets is the same as that in the original dataset, which is important to have more representative samples and reduce sampling errors.

. . Performance analysis of all classifiers with oversampling
In order to select the optimal classifier for the classification task, we apply nine classifiers for comparison.As the disparity of the sample number between different classes is too large, we conduct the experiments repeatedly using different oversampling ratios.Every classifier's final results are obtained with the optimal oversampling ratio.
Furthermore, we also compare the results of the classifiers on the original data and the oversampling data.In the experiment, since the number of MCI and NC are 556 and 2,004, respectively, we use ADASYN to oversample for MCI at different ratios, starting from 100 to 300%, and the experiment is repeated three times for each classifier.Table 4 shows the results of all classifiers on original data and oversampling data with 10-fold cross validation.Compared with the results obtained from the original data, the specificity of all classifiers from the oversampling data decrease, which means the prediction ability of all classifiers for NC declines.On the contrary, the sensitivity of all models increases significantly, which indicates that the prediction ability of MCI has been greatly improved by using the oversampling method.Moreover, according to the increment in accuracy, G-mean, and AUC among most of the classifiers, more samples generated by oversampling technique make the whole performance improve.Although SVM achieves the best with respect to the accuracy and the G-mean, its overall performance is not as good as AdaBoost after applying the feature selection methods, which will be discussed in the next subsection.
Table 5 shows the classification results of AD and NC with 10-fold cross validation.Since the number of AD is only 98, the oversampling ratio is from 100 to 2000%.In the experiment, except for the specificity, the other four metrics of all classifiers have been greatly improved by applying oversampling, especially the sensitivity.The AdaBoost achieves accuracy of 0.996, sensitivity of 0.999, specificity of 0.993, G-mean of 0.996, and AUC of 1.Although the sensitivity of AdaBoost is slightly worse than that  .

. Performance analysis of feature selection methods under AdaBoost and oversampling
The aim of this experiment is to use the feature selection method to decrease the dimension of the dataset and computational complexity.Five feature selection methods are applied in the experiment, which are Gini index, IG, reliefF, LASSO, and SGL.They are used to reduce the dimension from 53 to 15 for the oversampling dataset, which is selected from the prior experiments with an optimal oversampling ratio when AdaBoost achieves the best performance.Then we train the AdaBoost model on these datasets.Tables 6, 7 show the results for different classification tasks with 10-fold cross validation under different feature selection methods.In the classification task of MCI/NC, reliefF achieves the best performance with respect to accuracy, sensitivity, G-mean, and AUC, which are 0.892, 0.877, 0.892, and 0.957, respectively.It also achieves the optimal values for the four metrics in the AD/NC classification task, where accuracy is 0.992, sensitivity is 0.997, Gmean is 0.992 and AUC is 1.Therefore, the reliefF is selected as the final feature selection method.
We also compare Adaboost with SVM to classify MCI/NC on the dataset processed by reliefF.The results indicate that Adaboost indeed outperforms SVM, as shown in Table 8.
The runtime of the experiment with feature selection is 111.6 seconds in this experiment, which is much smaller than that without feature selection, 393.3 seconds.Although the performance of classifiers slightly reduces after feature selection, it decreases the computational complexity of the model and screens out the most important features, which lays the foundation for further analysis of these features.

. . Model explainability
The learned model is further analyzed by using the SHAP.The 15 features selected by reliefF in the MCI/NC classification task are Education, Sleeping time (elder), Heart rate, Height, Memory decline, Age, Eat fish, Diastolic blood pressure, Sport, Systolic blood pressure, Tea, Hypertension, Frequency of nap (elder), Smoke and Family history of physical disease.Regarding the classification of     We notice that Education is the most important feature for the MCI class, where high value of Education has a negative impact on predicting MCI class, meaning that Education is a factor that decreases MCI risk.Some features [e.g., Sleeping time (elder), Heart rate, Height, etc.] are globally less critical than Education, but they have more impacts in some cases.For instance, the largest SHAP value for Sleeping time (elder) is 0.0497, which is greater than the maximum SHAP value of Education, 0.0275.Similarly, for AD, the top feature is the Daily life function decline.The feature Age is less critical, however, when its value is very small, it has more negative impact than the Daily life function decline on the model for predicting AD.In addition, there are nine features that are identical for MCI and AD, but the importance of these features in MCI class is not as high as theirs in AD class.
We also analyze the impact of a single feature on the prediction.The Figure 3 shows the SHAP dependency plots for MCI class, where the x-axis represents the value distribution of each feature in all samples, and the y-axis represents the SHAP value.As seen in Figures 3A, C, D, E, G, with the increasing values of these features, the overall trend of the SHAP values is downward, indicating these features have negative effects on predicting MCI class.On the contrary, as the value of Age and Systolic blood pressure increase, their SHAP values also increase as shown in Figures 3B, F.
In the initial assessment targeting the prediction of Mild Cognitive Impairment (MCI) and Alzheimer's Disease (AD), as portrayed in Figures 4A, B respectively, the features manifesting the most substantial influence were "Memory Decline" and "Daily Life Function Decline."This observation harmonizes with the discernments encountered in clinical diagnostic realms.Aiming for a more meticulous evaluation of our model's stability, we pivoted our attention toward the interplay of daily life features with MCI and AD prognostics.Specifically, we excised crucial features conventionally harnessed for clinical recognition: "Memory Decline, " "Daily Life Function Decline, " and "Disability in Work and Study."The subsequent SHAP plots, delineated in Figure 5, vindicate that notwithstanding the exclusion, daily life features retain a pivotal role in rendering credible predictive values in our model's framework.This meticulous endeavor not only  underscores the robustness of our model but also illumines the nuanced daily life attributes that contribute significantly to the predictive landscape of MCI and AD.
Figure 6 displays the SHAP dependency plots for AD class.We notice that the SHAP values of the Daily life function decline and Age increase as their feature values increase.In addition, the downward trend is observed in Figures 6C-E, which is the same as their trend for MCI.We also discover that the Eating habits (1 = Vegetarian-based diet, 2 = Meat-based diet, 3 = Meat, and vegetables) have a negative impact on classifier when its value is 1 or 3, which indicates that eating vegetables is helpful to prevent AD.Meanwhile, as shown in Figure 6G, the SHAP value is below 0 if the value of the Frequency of nap (elder) is 4 (0 = None, 1 = Sometimes, 2 = 1-3 days a week, 3 = 4-6 days a week, 4 = Every day), which means regular naps may help to reduce the risk of AD.
SHAP can also conduct an analysis of a single sample.Figure 4 shows the contribution of each feature value to the classifier's judgement of MCI and AD instances.Each feature value is a force that either increases or decreases the prediction.As shown in Figure 4A, the sample is classified as MCI with a probability of 62%.The top four features are Memory decline, Diastolic blood pressure, Eat fish, and Education.And these values of features increase the probability that the classifier will judge the sample as MCI. Figure 4B shows the same thing for AD class.The model is 81% confident that the sample is AD.The Daily life function decline, Diastolic blood pressure, Education, and Sleeping time (elder) play important roles to push the prediction decision toward AD class.

. Discussion
This study develops an explainable machine-learning framework to predict AD/MCI based on clinical data obtained from CLAS.The performance of the framework has been improved by oversampling.We also apply multiple classification methods and feature selection methods for comparison, so that the best methods for prediction are selected.The resulting model achieves accuracy of 89.2%, sensitivity of 87.7%, specificity of 90.7%, G-mean of 89.2%, and AUC of 0.957 for MCI/NC prediction, while it achieves accuracy of 99.2%, sensitivity of 99.7%, specificity of 98.7%, G-mean of 99.2%, and AUC of 1 for AD/NC prediction.Then we make a detailed analysis by visualizing the specific contributions of the features to the classifier's output.To the best of our knowledge, this is the first attempt to employ ensemble learning with feature selection to develop models for AD prediction based on the large-size lifestyle and medical information.The strengths of this study include an unprecedentedly large-size dataset, an advanced machine learning-based algorithm that jointly considers the associations among the clinical and lifestyle features toward an effective feature set, as well as an explainable prediction model.
Our results are compatible with previous intuitions and scientific knowledge.Tables 9, 10 summarize the existing studies about the relationship between some features and AD/MCI.In the two tables, 13 features and 10 features are shown to be associated with MCI and AD, respectively, which overlaps with our selected features.This validates that our feature selection method is fairly reasonable.Hebert et al. (1995) and Petersen et al. (2018) discover that the risk of MCI and AD increases with the increasing age.High education may reduce the risk of MCI and AD claimed by Sattler et al. (2012).Marshall et al. (2012) believe that daily life function decline aggravates the severity of dementia.(2020).
Regarding the lifestyle, eating fish is beneficial for preventing MCI and AD, discussed by Barberger-Gateau et al. (2007) and Sinn et al. (2012).Barberger-Gateau et al. (2007) also find that frequent consumption of vegetables may decrease the risk of AD.Tea intake may reduce the risk for dementia discussed by Kakutani et al. (2019).Shi et al. (2018) andBrachem et al. (2020) find that the poor sleep quality can enhance the risk of MCI and AD.Cross et al. (2015) discover that the relationship between nap and the risk of dementia exists.The conclusions of these papers above are reflected accordingly in Figures 3, 6, which indicates that our model is fairly reasonable.In addition to these supportive research, our results further demonstrate that AD is a complicated disease that is affected by multiple factors, including daily lifestyle and physical disease.With an advanced feature selection and a unified framework of machine learning, we are able to detect the combination of such contributive features.

. Conclusion
We develop an explainable machine-learning based model with oversampling and feature selection methods.The oversampling method is used to generate new samples for the minority class to solve the data imbalance issue.The feature selection method is applied to reduce the data dimension, so as to lower the computational complexity of the model, and to find out the most important features.We adopt the ensemble learning method to implement the prediction.Our model not only realizes the prediction, but also provides the specific contribution of each feature to the prediction classifiers by building an explainer.Experimental results demonstrate that the model achieves excellent performance, which coincides with other prior research.In sum, our model not only provides the prediction outcome, but also helps to understand the relationship between lifestyle/physical disease and cognitive function, and enables clinicians to make appropriate recommendations for the elderly.Therefore, our approach provides a new perspective for the design of a computeraided diagnosis system for AD, and has potential high clinical application value.

. Future work
The study has several limitations.Firstly, the cross-sectional study is not able to examine causal relationships between life style and individual cognitive decline.Follow-ups are needed to make the final outcome for these population.Furthermore, we did not include FDG-PET, Aβmarkers, and APOE genotype in this work, so the true extent of AD pathology remains unknown.Additionally, the exploration of multimodal data encompassing neuropsychological tests, and functional neuroimaging data, genetic information, and other relevant biological indicators will be undertaken in research to provide a multifaceted understanding of the pathophysiology of MCI and AD.The analysis of multimodal data through advanced machine learning and artificial intelligence techniques will be employed to unveil hidden patterns and relationships, aiding in the better understanding of cognitive decline risk factors and pathophysiological mechanisms.Moreover, these technologies will be harnessed to develop predictive models for the early identification of MCI and AD risks.

FIGURE
FIGUREOverview of the structure of the proposed system.

FIGURE
FIGURE SHAP summary plots for MCI and AD prediction.(A) MCI.(B) AD.

FIGURE
FIGURESHAP force plots for (A) MCI and (B) AD instances.

FIGURE
FIGURESHAP force plots for (A) MCI and (B) AD instances focusing on lifestyle features.
TABLE Statistics summary of the full data set for , patients (Part ).
TABLE Hyperparameter space explored for each model.
TABLE Performance under di erent classifiers for MCI/NC prediction (OR, Original data; OS, Oversampling data).
TABLE Performance under di erent classifiers for AD/NC prediction (OR, Original data; OS, Oversampling data).
of SVM, its other metrics are the best among all classifiers.In summary, we select AdaBoost as the classifier.
TABLE Performance for di erent feature selection methods for MCI/NC classification using ADASYN and AdaBoost.
TABLE Performance for di erent feature selection methods for AD/NC classification using ADASYN and AdaBoost.
TABLE The performance comparison between AdaBoost and SVM in MCI/NC classification task using ADASYN and ReliefF.
TABLE Current studies on the relationship between selected features and MCI.
TABLECurrent studies on the relationship between selected features and AD.These results elucidate that sleep disturbance can enhance the risk of developing dementia.Insomnia may increase the risk of incident AD, and sleep disordered breathing (SDB) is a risk factor for all-cause dementia, AD, and vascular dementia.Cross et al. (2015)Frequency of nap (elder) This study highlights that nap is associated with underlying neurobiological changes such as depression and cognition.Thus it is necessary for older individuals to monitor the nap routinely to elucidate their relationship with psychological and cognitive outcomes.The trends of the Diastolic blood pressure and Systolic blood pressure in Figures3F, G, 6E also verify the conclusion in Ou et al.