ORIGINAL RESEARCH article

Front. Psychol., 14 April 2022

Sec. Quantitative Psychology and Measurement

Volume 13 - 2022 | https://doi.org/10.3389/fpsyg.2022.855379

Predicting Verbal Learning and Memory Assessments of Older Adults Using Bayesian Hierarchical Models

  • 1. Department of Statistics, Faculty of Science and Literature, Institute of Graduate Studies, Ondokuz Mayis University, Samsun, Turkey

  • 2. Department of Statistics, College of Natural and Computational Sciences, Debre Tabor University, Gondar, Ethiopia

Article metrics

View details

3

Citations

2,9k

Views

951

Downloads

Abstract

Verbal learning and memory summaries of older adults have usually been used to describe neuropsychiatric complaints. Bayesian hierarchical models are modern and appropriate approaches for predicting repeated measures data where information exchangeability is considered and a violation of the independence assumption in classical statistics. Such models are complex models for clustered data that account for distributions of hyper-parameters for fixed-term parameters in Bayesian computations. Repeated measures are inherently clustered and typically occur in clinical trials, education, cognitive psychology, and treatment follow-up. The Hopkins Verbal Learning Test (HVLT) is a general verbal knowledge and memory assessment administered repeatedly as part of a neurophysiological experiment to examine an individual’s performance outcomes at different time points. Multiple trial-based scores of verbal learning and memory tests were considered as an outcome measurement. In this article, we attempted to evaluate the predicting effect of individual characteristics in considering within and between-group variations by fitting various Bayesian hierarchical models via the hybrid Hamiltonian Monte Carlo (HMC) under the Bayesian Regression Models using ‘Stan’ (BRMS) package of R. Comparisons of the fitted models were done using leave-one-out information criteria (LOO-CV), Widely applicable information criterion (WAIC), and K-fold cross-validation methods. The full hierarchical model with varying intercepts and slopes had the best predictive performance for verbal learning tests [from the Advanced Cognitive Training for Independent and Vital Elderly (ACTIVE) study dataset] using the hybrid Hamiltonian-Markov Chain Monte Carlo approach.

1. Introduction

Verbal learning and memory tests are highly varied among older-aged adults due to various influences. Early cognitive intervention in older adults is a critical program to reduce the future risk of dementia (Thomas et al., 2019). The efficacy of the Chinese form Hopkins Verbal Learning Test (HVLT) for screening dementia and mild cognitive impairment in a Chinese population showed that HVLT scores were affected by age, education, and sex (Shi et al., 2012). The dataset of Advanced Cognitive Training for Independent and Vital Elderly (ACTIVE) study consists of two hierarchies in which four different repeated measures are nested within each participant (Luo and Wang, 2014). The outcome measures of the cognitive training interventions were the total HVLT from three learning trials and the baseline measure (Gross, 2011).

Bayesian logistic and hierarchical probit models of accuracy data that allow two levels of mixed-effects in repeated-measures designs have been implemented. The Bayes factor through the Bayesian information criterion estimate and the Widely applicable information criterion (WAIC) model selection techniques were used (Song et al., 2017). Duff (2016) used stepwise regression model to scrutinize the effect of age, education, and gender on HVLT scores in 290 cognitively intact older adults. The study revealed that age was negatively correlated with the HVLT score, while education status was positively correlated. Moreover, there were fewer gender differences among four repeatedly measured verbal learning tests (Lekeu et al., 2009).

Another study showed that besides capabilities through training, personal characteristics like age, unmarried status, and lower occupational cognitive requirements increased the likelihood of cognitive risk (Silva et al., 2012). Higher educational levels and active engagement in exercise may contribute to cognitive reserve and have a protective effect on cognitive decline in late life (Shen et al., 2021).

Gender effects on neuropsychological performance were negligible when the age and educational status of elderly people were controlled (Welsh-Bohmer et al., 2009). Recently, the Markov chain Monte Carlo (MCMC) methods have been widely used to generate samples from complicated and high-dimensional distributions (Hadfield, 2017). Among all Bayesian computational methods, the Hamiltonian Monte Carlo (HMC) (Almond, 2014) approach is the most efficient for approximating complex data structure models and converges faster than the traditional Metropolis-Hastings and Gibbs methods (Kruschke and Vanpaemel, 2015). The common MCMC approaches show poor performance and tremendously slow convergence in complex parameter structures (Yao and Stephan, 2021).

The HVLT is the ultimate in situations calling for multiple neuropsychological assessments (Benedict et al., 1998). Classical statistical inferences and single-level models have limitations for predicting naturally nest data. Bayesian hierarchical models (Congdon, 2020) were able to predict verbal learning test and memory scores from baseline personal characteristics, such as age, gender, cognitive status [mini-mental state exam (MMSE) score], years of education, and participants’ booster training and reasoning ability measured by training progress (Kuslansky et al., 2004).

In Bayesian inference, the WAIC, the leave one out information criterion (LOO-IC), and K-fold cross-validation (K-fold-CV) are recently developed measures of complexity penalized fitting models (Almond, 2014; Sivula et al., 2020). In this article, model comparisons and model selections were performed using these three methods under the Bayesian Regression Models using ‘Stan’ (BRMS) package of R (Bürkner, 2018). In most cases, WAIC and LOO-IC showed a slight preference for the random slope model over other models (Bürkner, 2018). However, the general model selection principle shows to choose the null model when diffuse priors are used in the parameters to be included or rejected by the algorithms (Liu, 2000). Therefore, in this article, we used the HMC approach to fit the three different Bayesian hierarchical models and select the best predictive model.

2. Materials and Methods

2.1 Data and Variables

The ACTIVE study was a randomized controlled trial conducted in 1999–2001 at six diverse research centers in the United States and organized by the New England Research Institutes (NERI). A total of 1,575 purposively selected older adults were included in this study (Willis et al., 2015), in which 26% of the participants were African American. The ACTIVE dataset accessed from the study of Willis et al. (2015) has 13 variables. However, this modeling paper used six explanatory variables, and the dependent variable HVLT is used as repeated measures of learning tests and memory ability. In this dataset, HVLT has four different repeated measurement scores doi: 10.3886/ICPSR04248.v3.

2.2 Bayesian Hierarchical Model for Repeated Measures Data

Suppose X is the matrix of explanatory variables, and Y is the outcome variable that is the Total Hopkins Verbal Learning Test Score (THVLTS). Besides the classical statistics, a more flexible Bayesian model is required that can accommodate the varying correlation between covariates and independent variables that occur in repeated measures-type longitudinal data. The general form of the Bayesian hierarchical model for repeated measures data can be expressed as:

Where Y denotes the vector of outcome variable; β denotes a vector of fixed effects parameters; U denotes a vector of associated random effects (specifictoeachsubject); X is a matrix of covariates (explanatory variables); Z denotes a block diagonal matrix of covariates for the random effects as a complement of X embraced of m blocks that each block has ni × q dimension matrix and ε denote a column vector of residuals. We assumed that the random effects U∼N(0d, Ω) and the residuals . Where U and ε are independently distributed. Based on the unknown vector of φΩ and φR, the unknown random effects in Ω and R can be written as Σ = (φΩ, φR) (Laird and Ware, 1982).

Where X is divided into two columns corresponding to fixed effects and a corresponding random effects design matrix denoted as and , respectively. And the parameters are divided into fixed effects β(F) and random effects β(R) = U. Cov(ui,ui) = Var(ui) = Ω and

It can be assumed that the hyperparameters of both the intercept and the coefficient/slope model have uniform hyper-prior distributions with appropriate assumptions for the parameters μu, μβ, σu, σβve ρ. Then, the mathematical form of the three possible Bayesian hierarchical models (Nalborczyk and Vilain, 2019) for predicting the verbal learning and memory test with two (group/subject and time) random effects (Hilbe, 2009) can be written as follows:

Model 1: Null Model

Here, the model is fitted by varying the intercept without including any predictor variable. Thus, this model shows the overall within and between-subject variations of the outcome variable (Goldstein et al., 2009).

Model 2: Varying Intercept Model

Here, the BRMS command is fitted in R with varying intercepts for both clusters (i.e., participating subjects) and repeated measures (i.e., measurement time point) by including all predictor variables in the model. Thus, this model can be called a random intercept and fixed slope model (McGlothlin and Viele, 2018).

Model 3: Varying Slopes

Here, we can focus on examining the dependence between the random intercepts and the random coefficients (Bafumi and Gelman, 2011). In this case, we are interested in whether the effects of age and reasoning skill have correlations with variations in verbal and memory test skills measured by trail scores.

Where S is the covariance matrix, is the corresponding correlation matrix, and ρ is the association between intercepts and coefficients used in the calculation of S. The prior matrix R is the LKJ-correlation (Lewandowski et al., 2009) with a parameter ζ(zeta) which regulates the strength of the association.

As shown in Figure 1 above, each component of the mixed effect model appears in the graph as a node. The dotted arrows represent deterministic (fixed) dependencies between the parameters (e.g., from β to μij), whereas the solid arrows represent probabilistic (random) dependencies (e.g., from to Yij) (Bürkner, 2018). The hyper-parameters of the varying both intercept and slope model (μα, μβ, σα, σβ, and ρ) can be assumed to have hyper-prior distributions with appropriate assumptions for the parameters (Liu, 2016; Congdon, 2020).

FIGURE 1

FIGURE 1

A varying intercept and slope model (Bayesian Framework).

2.3 Bayesian Information Criterion for Model Comparison and Selection

Watanabe’s Widely Applicable (WAIC)

WAIC (Watanabe, 2010) could be achieved as an improvement over the divergence-based information criterion (DIC) for Bayesian models. The deviation term used in the calculation of the WAIC is Log-Point Based -Requires Predictive-Density (LNTTY). LNTTY is calculated as:

The whole ppost(θ) is the posterior distribution used in the calculation of LNTTY. Similar to LNTTY, WAIC’s penalty term is purely Bayesian and is computed as:

Where pWAIC is the penalty term which is the variance of the log-predictive-density terms aggregated over N data points. Thus, the WAIC can be calculated as:

Leave-One-Out Information Criteria (LOO-CV)

Bayesian leave-one-out cross-validation (LOO-CV) is different from the WAIC. Because there is no penalty term in its calculation. LOO-CV can be computed as:

Where ppost(−i)(θ) is the posterior distribution based on a sub-set of the data at point i from the dataset. LNTTY used ith data points to calculate both the posterior distribution and the parameter estimation. Here, in contrast, the log-pointwise predictive density (LPPDloo) is used the same for prediction only. Therefore, there is no need for a penalty term to correct potential bias by using the data twice (Vehtari et al., 2017).

K-Fold Cross-Validation

Sometimes, multiple Pareto Corrected Significance Sampling (PSIS-LOO) fails, and it takes too long to remodel in the iteration. Therefore, we can estimate LOO-CV using K-fold-CV by separating the data into completely random multiples, which leads to looking at each cross-validation estimate distinctly (Vehtari et al., 2018).

The Bayesian K-fold-CV partitions the dataset into k subsets yk(k = 1, 2, …, K). The Bayesian hierarchical model (BHM) generates each training dataset yke separately, which returns a ppost(e)(θ) = p(θ|y(ke)) posterior distribution (Vehtari and Gelman, 2014). To preserve reliability with WAIC and LOO-IC, defining the predictive accuracy of every point in the dataset is essential. Therefore, the log-predictive distribution function is

Using “S” simulations corresponding to a subset of k (usually K = 10) containing the ith data point and the posterior distribution P(θ|y(ke)). The overall estimate of the expected log point predictive density for a new dataset is determined as follows:

Therefore, a point estimate of the k-fold value is the sum of the iterative folds from the data points.

2.4 The Hamiltonian Monte Carlo Algorithm in Bayesian Regression Models Using ‘Stan’ Package of R

Similar to Gibbs sampling, HMC practices a proposal distribution that changes subject to the recent location in the parameter space (Liu, 2000). However, unlike the Gibbs algorithm, HMC does not rely on computing the conditional posterior distribution of parameters and sampling from it (Mai and Zhang, 2018). HMC has two advantages over other MCMC methods: little or no autocorrelation of the samples and fast mixing, i.e., the chain converges to the distribution immediately (Nalborczyk and Vilain, 2019). Therefore, it is the best approach for continuous distributions with low (auto) correlation and low rejection of samples.

When the model parameters are continuous rather than discrete, HMC, also known as Hybrid Monte Carlo, can overpower such random walk behavior using a clever scheme of supplementary variables that converts the tricky of sampling from the targeted function into the simulating Hamiltonian dynamics (Britten et al., 2021). HMC is an MCMC algorithm that avoids the random walk behavior and sensitivity to correlated parameters that outbreak other MCMC approaches by performing a series of steps informed by first-order gradient information (Hilbe, 2009).

The HMC algorithm is based on the Hamiltonian (total energy) calculating the trajectory for a time t = 0, …, T and then taking the final position X(T) = Xn+1.

The steps of the algorithm are as follows:

HMC algorithm

  • 1.

    Choose a starting point and a velocity distribution θ0 = X0q(v)

  • 2.

    for n = 0, …

  • 3.

    Set the initial position as X(t = 0) = Xn

  • 4.

    Draw a random initial velocity, v(t = 0)∼q(v);

  • 5.

    Integrate the orbit numerically with the total energy for some time (use the Leapfrog method):

  • 6.

    Calculate the probability of acceptance:

  • 7.

    Set Xn+1 = X(t = T)

  • 8.

    Increment

3. Results

In practice, the three basic Bayesian hierarchical models have been fitted in BRMS default settings, and population-level (fixed) effects and subject-level (random) effects were obtained (Luo et al., 2021). All three models (Models 1, 2, and 3) had both fixed and random (mixed) parts but with different estimated parameter types. In the result, the estimate shows the posterior mean and Est. Error is the SD for each parameter. Model convergence was achieved well enough both the bulk effective sample size (Bulk_ESS) and the tail effective sample size (Tail_ESS) for the 95% CIs were adequate (Vehtari et al., 2017). In general, every parameter is summarized using the posterior distribution’s mean (“Estimate”) and SD (“Est. Error”), as well as two-sided 95% credible intervals as lower and upper bounds based on quintiles.

Table 1 of the fixed effects shows that the posterior mean verbal testing score was estimated to be 26.33 with an SD of 0.73. The 95% credible interval shows that the posterior distribution mean (intercept) was significant. On the other hand, the random effect showed significant verbal score test variation between groups (participant subjects) and within-subjects (between different measurements of different time points). Thus, according to the null model, the HVLT score showed more between-group/subject variation than within-group (between repeated measurements) variation.

TABLE 1

Outcome variableCovariates
EstimateEst. ErrorBulk_ESSTail_ESS95% CI
Fixed effectsLowerUpper
Total hopkins verbal learning test score (THVLTS)Intercept26.33120.7331137118751.0124.850127.7214
Random EffectsLowerUpper
σintercept,subject4.31050.085281014501.004.15244.4751
σintercept,time1.30350.6456204724291.000.57543.0562
σe(sigma)3.11340.0256331532961.013.04623.1662

Results from the fitted null model: Model 1.

Table 2 showed that the coefficient of booster training was positive with a zero overlapping 95% CI. This indicates that, on average, there is little evidence that taking booster training increases elderly adults’ verbal learning and memory test scores by 0.1865, but the evidence-based on the data and random intercept model. On the other hand, adults’ years of education (edu) estimate was negative with a zero overlapping 95% CI. This negative estimate indicates that, on average, in the random intercept model, there is little evidence that increasing the years of education decreases elderly adults’ verbal learning and memory test scores by 0.0034 units.

TABLE 2

Outcome variableCovariates
EstimateEst. ErrorBulk_ESSTail_ESS95% CI
Fixed effectsLowerUpper
Total hopkins verbal learning test score (THVLTS)Intercept9.23141.941112601891.005.471212.9510
Age−0.12110.021292617021.01−0.1611−0.0854
Edu−0.00340.0011413928381.01−0.01010.0042
Booster0.18650.175464518381.00−0.15110.5432
Gender2.65640.201591016071.002.27523.0654
Reason0.14640.011298016731.000.13100.4232
MMSE0.60120.0462103221281.000.50130.7012
Random effectsLowerUpper
σintercept,subject3.03120.0654114622711.002.88453.1645
σintercept,time1.26540.6572185221211.000.58323.0812
σe(sigma)3.11020.0312426430291.003.04623.1761

Results from the fitted varying intercept model: Model 2.

According to the predictive effects of each explanatory variable shown in Figure 2 and Table 3, taking booster training, age, and gender were the most influential factors affecting participants’ cognitive verbal test and memory ability. Table 3 reveals that there is also an adverse association between the intercepts and coefficients for reasoning ability, which implies reasoning ability has a large average score value showing additional variability by poor reasoning ability than by good reasoning ability. Nevertheless, it can be seen that the slope estimate of such a model is even further unreliable than that of the preceding models, as it can be clearly understood from the associated standard error and the size of the 95% CIs. Table 3 also showed that booster training had a significant positive predictive effect on elderly adults’ verbal learning and memory test scores. In contracts, adults’ years of education had a significant negative impact on elderly adults’ verbal learning and memory test scores.

FIGURE 2

FIGURE 2

Bayesian hierarchical varying slope convergence diagnosis.

TABLE 3

Outcome variableCovariates
EstimateEst. ErrorBulk_ESSTail_ESS95% CI
Fixed effectsLowerUpper
Total hopkins verbal learning test score (THVLTS)Intercept9.84122.0602315729181.005.852313.9344
Age−0.12110.0213284627201.01−0.1513−0.0823
Edu (education)−0.00330.0012552327701.00−0.01220.0012
Booster0.14120.1703336228761.00−0.21450.4831
Gender2.55050.2004323628661.012.17122.9331
Reason0.14440.0113308728671.000.13130.4402
MMSE0.58030.0512325630421.000.48220.6840
Random effectsLowerUpper
σintercept,subject1.92221.28331114881.000.07244.3111
σintercept,time1.30220.8004202722701.000.57023.1343
σage0.04240.01331008331.000.01230.0625
σreason0.04050.01321383911.000.02120.0732
corIntercept,age0.10330.43331112551.00−0.70420.8303
corIntercept,reason−0.39020.42041003881.00−0.90110.6212
corage,reason−0.59220.264551910531.00−0.90420.1407
σe(sigma)3.11020.0333376727481.003.05333.1710

Results from the fitted varying slope mode: Model 3.

We also noticed in Figure 2 and Figure 3 below that adding any term to the early model showed predictive performance improvements on the fitted models are ordered from Models 1 to 3 (full model). However, such a result may not be interpreted as a universal rule, subsequent adding extra terms to a unique model may also result in overfitting, which corresponds to a condition in which the fitted model is over-specified about the data, making the model good at clarifying the sample dataset but poor at predicting no observed data. The model convergence diagnosis plots are hairy caterpillars which showed the model converged. On the other hand, the models have well converged based on the estimated statistical values. This means that the R-hat statistics were close to 1 and the (bulk and tail) ESSs values were sufficiently high when ESS > 100 was chosen as the cutoff (Vehtari et al., 2021). The majority of parameters still showed sufficiently high ESS values when more conservative cutoffs were chosen (i.e., ESS > 400 or even 1,000, see Zitzmann and Hecht, 2019).

FIGURE 3

FIGURE 3

Bayesian hierarchical varying slope convergence diagnosis (Continuous).

Based on the fitted varying slope model, which accounted for six predictors from the data, fixed effects showed that age, gender, reasoning ability, and booster training were significant predictors of verbal learning and memory test scores, whereas random-effect showed that much of the variation in test scores occurred within-subjects (between measurement time points) than between subjects.

After we have built the three different models, it is necessary to identify relatively the best model that can be used to predict the outcome variable and make inferences. However, choosing the model that has the best predictive and a better fit on the actual data is complicated with diverse information criteria since all selected models on the actual data might not essentially achieve as fit on a different dataset. In its place, it is necessary to decide on a model that fits best in terms of predicting new data which had not been practiced.

In case of the non-existence of extra information, cross-validation methods such as WAIC and LOO-CV can be used. According to Table 4, the varying slope model has the lowest WAIC, LOO-IC, and 10-fold estimates. However, the difference is relatively small when we compare the difference in estimates of criteria for each model and the corresponding standard errors (in the column SE).

TABLE 4

Model typeModel selection criteria from BRMS package
WAIC
LOO-IC
10-fold
EstimateSEEstimateSEEstimateSE
Null model (Model 1)33638.0134.633744.1136.233923.8136.4
Varying Intercept model (Model 2)33494.5139.533574.9140.633717.0141.4
Varying slopes model (Model 3)33488.4141.833567.5143.033685.2140.8

Model comparisons based on predictive performance.

Among the fitted models above, it looks like the final model (Model 3) in the HMC algorithm is the best model. Therefore, as a function of the six explanatory variables and the random coefficient for age and reasoning ability, Model 3 has the best predictive performance for the cognitive HVLT.

According to Figure 4, the varying slope and intercept model fit well and produced nearly identical posterior observed density and posterior predictive distribution plots of the outcome variable of THVLTS from the ACTIVE study.

FIGURE 4

FIGURE 4

Bayesian hierarchical varying slope fitted model on the observed and predicted outcomes.

Furthermore, the marginal effect of each predictor variable revealed (Figure 5) that age and reasoning skills are the most significant explanatory variables that predict the THVLTS of the ACTIVE study.

FIGURE 5

FIGURE 5

Bayesian hierarchical varying slope model marginal prediction effects.

4. Discussion

Based on the selected sample participants in the ACTIVE study dataset (Willis et al., 2015), the Bayesian hierarchical linear models of three types were fitted by considering only six explanatory variables as predictors of the cognitive verbal learning test. The null model without any predictor effect but with only the intercept term was fitted, and it shows a mass of cognitive verbal learning ability variability across subjects. The varying intercept model with the addition of all predictor variables was fitted; and getting booster training, age, and reasoning ability were significant predictor of verbal test scores (Duff, 2016). The varying coefficient/slope model (i.e., Model 3) is the best-fitted model than the other fitted models since it had the lowest WAIC, LOO-IC, and 10-fold estimates (Bafumi and Gelman, 2011). A bulk of participants’ cognitive verbal test scores variations were observed between subjects (Ryoo, 2011). The full hierarchical model with varying intercepts and slopes has the best performance for predicting verbal learning tests (from ACTIVE study dataset) using the hybrid Hamiltonian Markov Chain Monte Carlo approach.

Socio-demographic and training-related characteristics influence elderly verbal learning tests that can be measured in multiple occupations (Welsh-Bohmer et al., 2009).

5. Conclusion

Total Hopkins Verbal Learning Test Score from the ACTIVE study can be used as a measure of elderly adults’ cognitive verbal learning ability. Four demographic characteristics of adults, such as age, gender, educational status, and cognitive status (MMSE score), were measured at the baseline, and characteristics measured after cognitive training such as reasoning ability and booster training were considered. THVLTS from the ACTIVE study can be used as a measure of elderly adults’ cognitive verbal learning ability. According to the findings, the varying intercept and slope model fit best, and age, gender, booster, and reasoning ability are the main significant predictors for THVLTS, which measures cognitive verbal learning. Taking booster training had a positive significant predictive effect, while years of education (edu) had a negative significant predictive effect on THVLTS.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Statements

Data availability statement

The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found in the article/supplementary material.

Ethics statement

Ethical review and approval was not required for the study on human participants in accordance with the local legislation and institutional requirements. Written informed consent for participation was not required for this study in accordance with the national legislation and the institutional requirements. Written informed consent was not obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article. This is because this quantitative analysis and modeling paper used open-access secondary data on repeated measurements.

Author contributions

EE participated in all aspects of the study: designing the study, performing data management, conducting the data analysis, writing the first draft of the manuscript, and discussing with MC to improve the manuscript, as it is a part of the first author’s Ph.D. dissertation. MC participated in revising the manuscript, commenting, and proofreading. Both authors listed have made a substantial, direct, and intellectual contribution to the manuscript and approved it for publication.

Acknowledgments

Our great appreciation and thanks are forwarded to NACDA members for the helpful repeated measured data collected and accessibility.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  • 1

    AlmondR. G. (2014). A comparison of two MCMC algorithms for hierarchical mixture models.CEUR Workshop Proc.1218119. 10.1016/j.neuroimage.2008.02.017

  • 2

    BafumiJ.GelmanA. (2011). Fitting multilevel models when predictors and group effects correlate.SSRN Electron. J.2, 814. 10.2139/ssrn.1010095

  • 3

    BenedictR. H. B.SchretlenD.GroningerL.BrandtJ. (1998). Hopkins verbal learning test – revised: normative data and analysis of inter-form and test-retest reliability.Clin. Neuropsychol.124355. 10.1076/clin.12.1.43.1726

  • 4

    BrittenG. L.MohajeraniY.PrimeauL.AydinM.GarciaC.WangW.-L.et al (2021). Evaluating the benefits of Bayesian hierarchical methods for analyzing heterogeneous environmental datasets: a case study of marine organic carbon fluxes.Front. Environ. Sci.9:491636. 10.3389/FENVS.2021.491636

  • 5

    BürknerP. C. (2018). Advanced Bayesian multilevel modeling with the R package brms.R J.10395411. 10.32614/rj-2018-017

  • 6

    BürknerP. C. (2019). thurstonianIRT: Thurstonian IRT models in R. J. Open Sour. Softw.4:1662. 10.21105/joss.01662

  • 7

    CongdonP. D. (2020). Book Bayesian Hierarchical Models: With Applications Using R Second Edition.London: Chapman & Hall.

  • 8

    DepaoliS.van de SchootR. (2017). Improving transparency and replication in Bayesian statistics: the WAMBS-checklist.Psychol. Methods22240261. 10.1037/met0000065

  • 9

    DepaoliS.WinterS. D.VisserM. (2020). The importance of prior sensitivity analysis in bayesian statistics: demonstrations using an interactive shiny app.Front. Psychol.11:608045. 10.3389/FPSYG.2020.608045/FULL

  • 10

    DominiqueA. (2015). “Doing Bayesian data analysis,” in Doing Bayesian Data Analysis, ed.KruschkeJ. K. (Bloomington, IN: Dept. of Psychological and Brain Sciences, Indiana University). 10.1016/c2012-0-00477-2

  • 11

    DuffK. (2016). Demographically corrected normative data for the hopkins verbal learning test-revised and brief visuospatial memory test-revised in an elderly sample.Appl. Neuropsychol Adult.23179185. 10.1080/23279095.2015.1030019

  • 12

    GelmanA. (2006). Prior distributions for variance parameters in hierarchical models (Comment on Article by Browne and Draper). Bayesian Anal.1, 515534. 10.1214/06-BA117A

  • 13

    GoldsteinH.CarpenterJ.KenwardM. G.LevinK. A. (2009). Multilevel models with multivariate mixed response types.Stat. Model.9173197. 10.1177/1471082X0800900301

  • 14

    GrossA. L. (2011). Memory Strategy use Among Older Adults: Predictors and Health Correlates in the Advanced Cognitive Training for Independent and Vital Elderly (Active) Study. Ph.D. Dissertation. Available online at: https://www.zhangqiaokeyan.com/academic-degree-foreign_mphd_thesis/02061282945.html(accessed August 19, 2021).

  • 15

    HadfieldJ. (2017). MCMCglmm Course Notes. Available online at: http://cran.nexr.com/web/packages/MCMCglmm/vignettes/CourseNotes.pdf(accessed September 9, 2021).

  • 16

    HilbeJ. M. (2009). Data analysis using regression and multilevel/hierarchical models.J. Stat. Softw.30530548. 10.18637/jss.v030.b03

  • 17

    KruschkeJ. K.VanpaemelW. (2015). “Bayesian estimation in hierarchical models,” in The Oxford Handbook of Computational and Mathematical Psychology, edsBusemeyerJ. R.WangZ.TownsendJ. T.EidelsA. (Oxford: Oxford University Press), 279299.

  • 18

    KuslanskyG.KatzM.VergheseJ.HallC. B.LapuertaP.LaRuffaG.et al (2004). Detecting dementia with the Hopkins verbal learning test and the mini-mental state examination.Arch. Clin. Neuropsychol.1989104. 10.1016/S0887-6177(02)00217-2

  • 19

    LairdN. M.WareJ. H. (1982). Random-effects models for longitudinal data.Biometrics38:963. 10.2307/2529876

  • 20

    LekeuF.MagisD.MariqueP.DelbeuckX.BechetS.GuillaumeB.et al (2009). Journal of clinical and experimental neuropsychology the California verbal learning test and other standard clinical neuropsychological tests to predict conversion from mild memory impairment to dementia.Exp. Neuropsychol.32164173. 10.1080/13803390902889606

  • 21

    LewandowskiD.KurowickaD.JoeH. (2009). Generating random correlation matrices based on vines and extended onion method.J. Multivar. Anal.10019892001. 10.1016/j.jmva.2009.04.008

  • 22

    LiuH. (2000). Efficiency of Markov Chain Monte Carlo Algorithms for Bayesian Inference in Random Regression Models.Ph.D. Dissertation. 12347. Available online at: https://lib.dr.iastate.edu/rtd/12347(accessed October 20, 2021).

  • 23

    LiuX. (2016). Methods and Applications of Longitudinal Data Analysis.Bethesda, MD: Uniformed Services University of the Health Sciences.

  • 24

    LuoS.WangJ. (2014). Bayesian hierarchical model for multiple repeated measures and survival data: an application to Parkinson’s disease.Stat. Med.3342794291. 10.1002/sim.6228

  • 25

    LuoW.LiH.BaekE.ChenS.LamK. H.SemmaB. (2021). Reporting practice in multilevel modeling: a revisit after 10 years.Rev. Educ. Res.91311355. 10.3102/0034654321991229

  • 26

    MaiY.ZhangZ. (2018). Software packages for Bayesian multilevel modeling.Struct. Equ. Model.25650658. 10.1080/10705511.2018.1431545

  • 27

    McElreathR. (2020). Statistical Rethinking: A Bayesian Course with Examples in R and STAN, 2nd Edn.London: Chapman and Hall/CRC.

  • 28

    McGlothlinA. E.VieleK. (2018). Bayesian hierarchical models.JAMA32023652366. 10.1001/jama.2018.17977

  • 29

    NalborczykL.VilainA. (2019). An introduction to Bayesian multilevel models using brms: a case study of gender effects on vowel variability in standard Indonesian.J. Speech Lang. Hear. Res.6212251242. 10.1044/2018_JSLHR-S-18-0006

  • 30

    PiironenJ.VehtariA. (2016). “On the hyperprior choice for the global shrinkage parameter in the horseshoe prior,” in Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, AISTATS 2017, Lauderdale, FL. 10.48550/arxiv.1610.05559

  • 31

    PolsonN. G.ScottJ. G. (2012). On the half-Cauchy prior for a global scale parameter. Bayesian Anal.7, 887902. 10.1214/12-BA730

  • 32

    RyooJ. H. (2011). Model selection with the linear mixed model for longitudinal data.Multivar. Behav. Res.46598624. 10.1080/00273171.2011.589264

  • 33

    ShenL.TangX.LiC.QianZ.WangJ.LiuW. (2021). Status and factors of cognitive function among older adults in urban China.Front. Psychol.12:728165. 10.3389/FPSYG.2021.728165

  • 34

    ShiJ.TianJ.WeiM.MiaoY.WangY. (2012). The utility of the hopkins verbal learning test (Chinese version) for screening dementia and mild cognitive impairment in a Chinese population.BMC Neurol.12:136. 10.1186/1471-2377-12-136

  • 35

    SilvaD.GuerreiroM.MarocoJ.SantanaI.RodriguesA.MarquesJ. B.et al (2012). Comparison of four verbal memory tests for the diagnosis and predictive value of mild cognitive impairment.Dement. Geriatr. Cogn. Disord. Extra2:120. 10.1159/000336224

  • 36

    SivulaT.MagnussonM.VehtariA. (2020). Uncertainty in Bayesian Leave-One-Out Cross-Validation Based Model Comparison. Available online at: https://arxiv.org/abs/2008.10296v2(accessed October 11, 2021).

  • 37

    SongY.NathooF. S.MassonM. E. J. (2017). A Bayesian approach to the mixed-effects analysis of accuracy data in repeated-measures designs.J. Mem. Lang.967892. 10.1016/j.jml.2017.05.002

  • 38

    ThomasK. R.RebokG. W.WillisS. L. (2019). Advanced cognitive training for independent and vital elderly (ACTIVE).Encycloped. Gerontol. Popul. Aging17. 10.1007/978-3-319-69892-2_1075-1

  • 39

    VehtariA.GelmanA.SimpsonD.CarpenterB.BurknerP. C. (2021). Rank-normalization, folding, and localization: an improved (formula presented) for assessing convergence of MCMC (with discussion).Bayesian Anal.16667718. 10.1214/20-BA1221

  • 40

    VehtariA.GelmanA. (2014). WAIC and cross-validation in Stan.ArXiv [Preprint]. Available online at: https://statmodeling.stat.columbia.edu/2014/05/26/waic-cross-validation-stan/(accessed December 5, 2021).

  • 41

    VehtariA.GelmanA.GabryJ. (2017). Practical Bayesian model evaluation using leave-one-out cross-validation and WAIC.Stat. Comput.2714131432. 10.1007/s11222-016-9696-4

  • 42

    VehtariA.SimpsonD. P.YaoY.GelmanA. (2018). Limitations of “limitations of Bayesian leave-one-out cross-validation for model selection”.Comput. Brain Behav.22227. 10.1007/S42113-018-0020-6

  • 43

    WatanabeS. (2010). Asymptotic equivalence of Bayes cross validation and widely applicable information criterion in singular learning theory.J. Mach. Learn. Res.1135713594.

  • 44

    Welsh-BohmerK. A.ØstbyeT.SandersL.PieperC. F.HaydenK. M.TschanzJ. A. T.et al (2009). Neuropsychological performance in advanced age: influences of demographic factors and apolipoprotein E: findings from the Cache County memory study.Clin. Neuropsychol.237799. 10.1080/13854040801894730

  • 45

    WillisS. L.JonesR.BallK.MorrisJ.MarsiskeM.TennstedtS.et al (2015). Advanced Cognitive Training for Independent and Vital Elderly (ACTIVE), United States, 1999–2008.Ann Arbor, MICH: Inter-University Consortium for Political and Social Research [Distributor].

  • 46

    YaoY.StephanK. E. (2021). Markov chain monte carlo methods for hierarchical clustering of dynamic causal models.Hum. Brain Mapp.42:2973. 10.1002/HBM.25431

  • 47

    ZitzmannS.HechtM. (2019). Going beyond convergence in Bayesian estimation: why precision matters too and how to assess it.Struct. Equ. Model.26646661. 10.1080/10705511.2018.1545232

Appendix

The Priors in Bayesian Hierarchical Models and Sensitivity Analysis

Different scholars suggested various priors for a component in a hierarchical model of variability parameters depending on the fitted Bayesian model structure and MCMC method. Some researchers proposed non-informative prior distributions, including uniform and inverse-gamma families in Gibbs sampling. Other researchers suggested a half-t family for the hierarchical model and demonstrate relatively weakly informative prior distribution. Half-student-t prior, a default prior in BRMS for SD parameters leads to better convergence. Still, the local shrinkage parameters lead to an increased number of divergent transitions in the BRMS of Stan (Piironen and Vehtari, 2016). The robust options for group-level standard deviations in Bayesian hierarchical models are half-normal (0,100), half-Student t with 3 degrees of freedom, or half-Cauchy prior distributions (Congdon, 2020).

Bürkner (2019) and McElreath (2020) proposed a half-normal distribution for SD priors in BRMS. Choosing a truncated normal distribution considers a good idea that the standard deviation cannot be less than 0. However, a prior on the random effect parameter with a long right tail has been revealed as “conservative” because it allows for bulky value estimates of the SD parameters.

Gelman (2006) suggested half-Cauchy prior with a mode at 0 and scale set to a considerable value by reasonably explaining the restrictive nature of half-Cauchy prior in providing enough information for the small numbers of groupings in the hierarchical nature of the data. Bayesian hierarchical model, to reduce the occurrence of unrealistically large SD estimates, BRMS-Stan documentation suggested half-Cauchy is the prior that automatically bound at 0. R-Stan renormalizes the distribution used so that the sum of the area between the bounds is 1.

The half-Cauchy (0, 1) prior is a case of the half-Student t-distribution with v1 degrees of freedom parameterized in the SD metric. It occupies a reasonable middle ground of different prior classes that performs well near the origin. It does not lead to drastic compromises estimates of the population-level (location) and group-level effect of the parameter space (Polson and Scott, 2012).

Sensitivity analysis of priors Appendix Table 1 shows the robustness of a Bayesian analysis when choosing different prior distributions in the fitted models. According to Bürkner (2019), for each SD component of the random parameters in hierarchical models, any prior distribution is practically well-defined on the non-negative real numbers only. In this study, we used the default in BRMS, the truncated Student’s t distribution with 3 degrees of freedom considered a reference prior. Then, because negative values are incredible for a standard deviation, we practiced a very strong informative truncated normal prior with a mean of 5 and a standard deviation of 0.01, and a half Cauchy prior for the sensitivity analysis of the impact of the prior on the Bayesian hierarchical models for the applied dataset.

Models with an effective sample size greater than 1000 and R-hat closest to 1.00 but not greater than 1.10 showed the consistency of an ensemble of Markov chains (Dominique, 2015). Moreover, both Bulk-ESS and Tail-ESS should be at least 100 (approximately) per Markov chain to be reliable and indicate that estimates of the respective posterior quantiles are reliable (Vehtari et al., 2021). In this paper, the R-hat, Bulk-ESS, and Tail-ESS results of the null, varying intercept, and varying coefficient models fulfilled these convergence diagnosis metrics. Therefore, the effective sample sizes (ESS) and potential scale reduction (R-hat) convergence diagnostic metrics are sufficient for stable estimates in each fitted model.

We used the sensitivity analysis for priors to scrutinize the final fully hierarchical specified model (Model 3) results, based on the default (or reference) prior, with the results obtained using different prior distributions. The posterior distributions and 95% posterior density (HPD) intervals by the median for the fixed effects and random effects, including SD of the verbal learning test score, did not change much depending on the priors specified, which indicates a practically identical interpretation of the estimates depending on the priors. Thus, because of no significant percentage deviation among models depending on the alternative prior specification, we reported the model results with the half-Cauchy prior that yielded good model convergence and sufficient ESS values (i.e., greater than or equal to 100). Besides, a good-looking posterior density plot of the predicted versus the observed data in Figure 4 was considered.

As it can be explained by Depaoli and van de Schoot (2017), and Depaoli et al. (2020), sensitivity analysis results could be provided through visuals, akin to the Shiny app plots or it may be in a table format indicating the degree of discrepancy in estimates or HPD intervals across parameters as we presented in the Appendix table below.

APPENDIX TABLE 1

Alternative priorsParameter/CovariatesEstimate (SD)Median (50%)5–95% HDPDefault estimate (SD)Percentage deviation
Alternative prior I: Half- Cauchy (0,1)Intercept9.8412 (1.521)9.83315.8523, 13.93449.8321 (1.932)0.0926

Age−0.1211 (0.021)−0.1201−0.1513, −0.0823−0.1223 (0.423)−0.9812
Edu (education)−0.0033 (0.001)−0.0033−0.0122, 0.0012−0.0034 (0.005)−2.9412
Booster0.1412 (0.102)0.1413−0.2145, 0.48310.1411 (0.623)0.0709
Gender2.5505 (0.112)2.55042.1712, 2.93312.5487 (0.222)0.0706
Reason0.1444 (1.902)0.14430.1313, 0.44020.1443 (2.081)0.0693
MMSE0.5803 (0.028)0.59210.4822, 0.68400.5801 (0.082)0.0345
σintercept,subject1.92221.92210.0724, 4.31111.92120.0521
σintercept,time1.30221.31020.5702, 3.13431.3032−0.0767
σage0.04240.04030.0123, 0.06250.04250.0126
σreason0.04050.04020.0212, 0.07320.0401−0.9975
corIntercept,age0.10330.1032−0.7042, 0.83030.10310.1040
corIntercept,reason−0.3902−0.8902−0.9011, 0.6212−0.3904−0.0512
corage,reason−0.5922−0.5887−0.9042, 0.1407−0.5923−0.0169
σe(sigma)3.11023.20413.0533, 3.17103.1112−0.0321
ParameterEstimate (SD)Median (50%)5–95% HDPDefault estimate (SD)Percentage deviation

Alternative prior II: Normal (5, 0.01)Intercept9.8423 (1.543)9.82316.1415, 13.65529.8321 (1.932)−0.0112

Age−0.1212 (0.034)−0.1212−0.1514, −0.0855−0.1223 (0.423)−0.0825
Edu (education)−0.0034 (0.011)−0.0034−0.0124, 0.0015−0.0034 (0.005)−2.9412
Booster0.1413 (0.124)0.1412−0.2165, 0.48710.1411 (0.623)−0.0708
Gender2.5514 (0.142)2.55052.1722, 2.93712.5487 (0.222)−0.0353
Reason0.1445 (2.013)0.14440.1453, 0.45620.1443 (2.081)−0.0692
MMSE0.5802 (0.035)0.58720.4852, 0.69510.5801 (0.082)0.0172
σintercept,subject1.92131.92210.0724, 4.34131.92120.0468
σintercept,time1.30331.31020.5622, 3.13441.3032−0.0844
σage0.04250.04030.0123, 0.06270.0425−0.2353
σreason0.04060.04020.0212, 0.07330.0401−0.2463
corIntercept,age0.10340.1032−0.7044, 0.83040.10310.0969
corIntercept,reason−0.3903−0.8902−0.9021, 0.6217−0.3904−0.0256
corage,reason−0.5923−0.5987−0.8045, 0.1404−0.5923−0.0169
σe(sigma)3.21153.21413.0533, 3.57103.1112−3.1543

Posterior estimates with the verity of priors: Sensitivity analysis results.

The relative percentage deviation can be computed as: {[(estimate using new alternative prior)–(estimate using default/reference prior)]/ (estimate using default/reference prior)}*100. Interpreting percentage deviation results is largely subjective and dependent on the metric of the parameters. However, percentage deviation under 10% would likely be considered negligible (Depaoli and van de Schoot, 2017). Default estimate = posterior estimate (mean) of analysis with the BRMS default/reference prior [Student’s t (3); 5–95% is the highest posterior density (HPD) interval].

Summary

Keywords

predicting, Hamiltonian Monte Carlo, Verbal Learning Test, hierarchical, model

Citation

Ebrahim EA and Cengiz MA (2022) Predicting Verbal Learning and Memory Assessments of Older Adults Using Bayesian Hierarchical Models. Front. Psychol. 13:855379. doi: 10.3389/fpsyg.2022.855379

Received

15 January 2022

Accepted

14 March 2022

Published

14 April 2022

Volume

13 - 2022

Edited by

Holmes Finch, Ball State University, United States

Reviewed by

Steffen Zitzmann, University of Tübingen, Germany; Luis Benites, Pontifical Catholic University of Peru, Peru

Updates

Copyright

*Correspondence: Endris Assen Ebrahim, , orcid.org/0000-0002-8959-6052

This article was submitted to Quantitative Psychology and Measurement, a section of the journal Frontiers in Psychology

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics