Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Psychol., 29 January 2026

Sec. Sport Psychology

Volume 17 - 2026 | https://doi.org/10.3389/fpsyg.2026.1743725

This article is part of the Research TopicPsychological Factors in Physical Education and Sport - Volume VIIView all 17 articles

Predicting college students’ exercise dependence: a machine learning approach

  • 1Department of Physical Education, Neijiang Normal University, Neijiang, China
  • 2Embodied Media Laboratory, Graduate School of Media Design, Keio University, Hiyoshi Campus, Yokohama, Japan
  • 3National Institute of Education, Nanyang Technological University, Singapore, Singapore
  • 4Zhuhai Research Center for Women and Children’s Sports Culture, College of Sports, Jinan University Zhuhai Campus, Zhuhai, Guangdong, China

Exercise dependence behavior among college students is a critical issue in sports psychology that deserve closer examination, and artificial intelligence offer a useful ways to explore its mechanisms and predicting associated risks. In this study, data were collected from 2,745 college students using three standardized questionnaires, covering (i) exercise dependence behavior, (ii) psychological characteristics (e.g., exercise identity, weight biases), and (iii) basic demographic information. We used four widely used machine learning algorithms: logistic regression, random forest, extreme gradient boosting (XGBoost), and multilayer perceptron, and their outputs were further integrated through an ensemble learning techniques to further enhance the robustness and predictive power of the models. The stacking ensemble model achieved a mean AUC of 0.96 in identifying exercise dependence risk among college students, demonstrating that integrating multiple machine learning approaches can yield robust and highly accurate risk prediction in this setting. Among the variables, the most influential predictors of exercise dependence behavior included prolonging exercise to obtain the desired effect, allocating most leisure time on exercise, experiencing difficulty in reducing exercise frequency, and actual exercise time longer than originally planned. These findings uncovers the key psychological and behavioral mechanisms underlying in exercise dependence among college students and show that artificial intelligence methods can be effectively applied to support risk monitoring in sport and psychological health contexts.

Highlights

• This study applies advanced machine learning algorithms to predict exercise dependence among Chinese college students.

• Deviation regulation theory and exercise psychology provide the theoretical foundation for identifying key psychological predictors.

• Exercise identity and weight bias are found to be significant psychological variables influencing exercise dependence.

• The use of data-driven methods improves prediction accuracy compared to traditional statistical approaches.

• Findings offer new perspectives for early identification and intervention strategies targeting exercise dependence in university populations.

Introduction

Regular physical activity is universally recognized as a building block of both physical and mental health, which has consistently been associated with positive health, better performance at school and work place (Calestine et al., 2017; Mahindru et al., 2023; Ruegsegger and Booth, 2018). Yet, as exercise load increases, the beneficial effects often decrease and may even be converted to maladaptive or addictive behavior. A highly relevant behavioral condition in this context is exercise dependence (ED). Exercise dependence (ED) is defined as a dysfunctional behavior pattern that is accompanied by withdrawal-like effects, such as anxiety and depression if the individual has to refrain from exercising, and an uncontrollable desire to increase workout intensity or frequency (de Coverley Veale, 1987). Systematic reviews estimate that approximately 3–7% of college students and regular exercisers are at risk for ED, with prevalence rates notably higher in athletic populations (Lichtenstein et al., 2018).

Importantly, this is a phenomenon that deserves special consideration in the Chinese higher education system given rapid sociocultural changes happening to student lives. Inspired by the nationwide “Healthy China 2030” program, fitness activities participation has increased significantly among Chinese youth (Chen et al., 2019). However, this wave of fitness meets some special stressors: college students in China face fierce academic pressure (commonly known as “involution”) and severe body image anxiety brought by the increasingly popular trend of social media (Wang et al., 2023). There is empirical evidence that these pressures may increase the risk for maladaptive exercise. Data from recent epidemiological surveys have shown that the proportion of Chinese college students with the potential risk for exercise dependence is 6.6 to 10.5% in active subpopulations, which is similar or even higher than the prevalence rates described in some Western data sources (Gong et al., 2023; Zhang, 2025). Despite these alarming trends, exercise dependence is rarely studied in the context of Chinese universities.

Despite several decades of research on exercise dependence (de Coverley Veale, 1987), most empirical studies have focused on competitive athletes and fitness enthusiasts (Szabo and Demetrovics, 2022), and fewer studies have examined college students in a systematic way (Terry et al., 2004). Moreover, much of the literature has been framed through a pathological lens, underscoring symptoms and diagnostic features (Berczik et al., 2012; Hausenblas and Symons Downs, 2002) rather than applying psychological theories to elucidate underlying mechanisms. For instance, prior work has predominantly relied on diagnostic criteria, such as tolerance, withdrawal, and loss of control, to classify high-risk individuals, but has paid comparatively less attention to the motivational and social-cognitive processes that may lead students to develop maladaptive exercise patterns (Griffiths et al., 2005; Szabo et al., 2013). By contrast, comparatively fewer studies have applied psychological theories to explain maladaptive exercise behaviors. One promising framework is deviance regulation theory (DRT), which argues that people regulate their behavior in order to preserve a valued social identity in relation to salient group norms.

According to DRT, people’s behavior is responsive to their estimates of the group norm with which they compare themselves: individuals are motivated to self-enhance when that category good is rare (e.g., exercising when most peers do not exercise) or to avoid attention when it is prevalent but potentially harmful (Blanton et al., 2001; Blanton and Christie, 2003). Within the exercise domain, message framing consistent with DRT has been found to affect intentions and subsequent behavior in reliance on perceived group norms and identity-relevant depictions (Gallagher and Updegraff, 2012). In addition, DRT has been used to explore exercise identity alongside social factors like weight bias in predicting excessive training or muscle building exercise (Palermo et al., 2021).

Drawing from this framework, the present study views each predictor as a theoretically justified proxy for either (a) perceived deviation from salient exercise norms or (b) conflict between one’s exercise identity and broader social expectations. So, for example, it has been shown that stronger exercise identity is associated with greater risk of exercise dependence when athletes perceive their own training volume or intensity exceeds what the general group norm would predict because deviation from the group norm may be a metaphor of why was believed to be a “discipline” self (Hausenblas and Symons Downs, 2002; Miller and Mesagno, 2014). On the other hand, pupils with a less strong exercise identity may not be as proned to risk compulsive behaviors because they are not highly motivated and do not contribute to develop different compulsive patterns in order to maintain an exclusive in-group status via high training commitment.

However, a major drawback of the current exercise dependence field is that much of the research has been conducted using conventional linear statistical methods [i.e., multiple regression, ANOVA, and structural equation modeling (SEM)] (Chang et al., 2019; Lichtenstein et al., 2014a). Although these methods have their strengths, they base on the assumption of linearity and additivity, i.e., that the association between psychological correlates (e.g., identity, anxiety) and exercise dependence is also linear across all participants (Yarkoni and Westfall, 2017). As a consequence, these models can suffer from modeling only part of the non-linear interactions that are known to describe behavioral psychopathology, thus leading to rather modest explained variance and predictive accuracy (Bzdok et al., 2018). For example, the relationship between exercise identity and dependence may not be constant, but rather depend on different levels of social pressure or body dissatisfaction, an exchange that is difficult to account for using traditional linear interaction terms in high-dimensional datasets (Strobl et al., 2009). In the specific exercise dependence literature, prior work has typically modeled a restricted set of predictors using such linear models, yielding only modest explained variance and offering limited insight into how identity-related and contextual factors jointly shape maladaptive exercise patterns.

To address these methodological shortcomings, machine learning (ML) techniques offer a robust alternative. Unlike traditional inference-based statistics, ML prioritizes out-of-sample prediction and can flexibly model high-dimensional, non-linear relationships without strict parametric assumptions (Breiman, 2001; Yarkoni and Westfall, 2017). In psychology research, these methods have already been used in areas such as emotion recognition (Khare et al., 2024), psychopathology (Bartlett et al., 2014; Livieris et al., 2018), and health behaviors (Abdul Rahman et al., 2023; Goh et al., 2022). Specifically, ML offers distinct advantages for this study: (1) the use of cross-validation (splitting data into training, validation, and test sets) minimizes overfitting and provides more generalizable performance estimates (Lu, 2010); (2) ensemble algorithms (e.g., random forest, XGBoost) can aggregate weak learners to significantly enhance classification accuracy over single-model approaches (Chen and Guestrin, 2016); (3) integration with explainable AI tools, such as SHAP (SHapley Additive exPlanations), allows for the transparent identification of variable contributions, overcoming the “black box” criticism of complex models (Lundberg and Lee, 2017); and (4) ML-based models can generate individual risk estimates, facilitating early detection and targeted intervention. Despite these clear advantages, to our knowledge, no study has yet combined DRT with ML techniques to disentangle the complex psychological architecture of exercise dependence among college students.

Grounding in deviance regulation theory (DRT) and by means of machine learning techniques, the present study examines exercise dependence among Chinese college students. Importantly, by positing exercise dependence as a maladaptive mode of social identity regulation, DRT directs us in selecting our features. We chose exercise identity as a core predictor, a dispositional motive to differentiate oneself positively from others (Strachan et al., 2011), and weight bias as an important contextual stressor—the concern about standing apart from group body norms (Puhl and Suh, 2015). And demographic variables (gender, grade, BMI) were not just included only as covariates but correlated with social comparison groups for which individuals’ behaviors are calibrated against (Blanton et al., 2001). By including these variables in machine-learning models, this study seeks to (1) identify essential factors that are primarily related with exercise dependence, (2) examine their relative contributions in predicting exercise dependence, and (3) offer scientific insights for the design of specific prevention and intervention plans.

Building directly on DRT, we formulated the following theory-driven hypotheses: (H1) Machine learning models will show improved predictive performance (higher accuracy and AUC) as compared to traditional logistic regression baselines because they account for non-linear dependencies; and (H2) Consistent with DRT’s focus on identity preservation, exercise identity and weight bias will rise to be the top-ranking predictors that interact such that students with high exercise identity who also experience high weight bias are predicted to have the highest probability of exercise dependence.

Methods

Participants

This study was approved by the Academic Ethics Committees of the University (Ethics No. YZUHL2020102), and informed consent was obtained from the students’ respective colleges and administrative units. Undergraduate students were recruited from several colleges, and classes were randomly selected from each of the four-year levels (freshman to senior). Data were collected online using Wenjuanxing (Questionnaire Star), which enabled students to complete the survey on their smartphones. A total of 2,920 questionnaires received, 106 questionnaires with less than 70% completion and 69 questionnaires missing values on key outcome variables were excluded. The final sample consisted of 2,745 students, giving a valid response rate of 94.01%. The average age was 20.37 ± 1.53 years, with males comprising 43.4% of participants. Table 1 summarizes the demographic characteristics (grade, gender, BMI).

Table 1
www.frontiersin.org

Table 1. Basic characteristics of the study population.

Measures

Outcome variable measurement: Exercise Dependence Scale-Revised

Exercise dependence behavior was assessed using the Chinese version of the Exercise Dependence Scale-Revised (EDS-R) adapted by Yang et al. (2021). The EDS-R includes 21 items across seven dimensions, scored on a 6-point scale (1 = never, 6 = always). Higher scores means more dependence-like exercise behavior. The scale demonstrated excellent internal consistency (Cronbach’s alpha = 0.977). Participants scoring 15 or higher on at least three dimensions were classified as at risk for exercise dependence (ARED). Those not classified as ARED but scoring 7 or higher on three dimensions were categorized as non-dependent symptomatic (NDS). Participants not meeting the criteria for either group were classified as non-dependent.

Predictor variable measurement: Exercise Identity Scale

Exercise identity was accessed with the Chinese version of the Exercise Identity Scale (EIS), originally developed by Anderson and Cychosz (1994) and revised by Li et al. (2019). This unidimensional scale consists of 9 items rated on a 7-point scale (1 = strongly disagree, 7 = strongly agree), with higher scores indicating a stronger exercise identity. The Cronbach’s alpha for the EIS in this study was 0.937. From a DRT perspective, exercise identity functions as an internalized standard for “being a disciplined exerciser,” making discrepancies from perceived peer norms particularly salient.

Predictor variable measurement: Eating Pathology Symptoms Inventory

Weight bias was measured with a 5-item subscale from the Eating Pathology Symptoms Inventory (ESPI) that measures negative attitudes towards obesity (α = 0.89). The ESPI is a 45-item self-report instrument evaluating eating disorder symptoms, with items rated on a 5-point scale (1 = never, 5 = very often), reflecting symptom frequency over the past 4 weeks (Forbush et al., 2013). Because no Chinese version of this subscale was available, we translated and evaluated its psychometric properties. Detailed confirmatory factor analysis (CFA) results supporting its reliability and validity in this sample are reported in the Results section. Within DRT, internalized weight bias is interpreted as a contextual pressure to avoid the stigmatized identity of being “unfit,” potentially motivating excessive exercise among students striving to conform to perceived body norms.

Demographic information questionnaire

The self-developed demographic questionnaire included three sections: gender, year of study (freshman to senior), height, and weight, BMI was calculated from height and weight.

Theoretical mapping of predictors to DRT

In line with DRT, all predictors were coded as indicators of either (a) perceived deviations from salient exercise norms (for example, training volume and dependence symptoms) or (b) discrepancies between the exercise self-concept and social norms regarding what it means to fit in with a given social group (internalized weight bias, demographic reference groups). Greater exercise identity and stronger weight bias would be expected to predict dependence risk, especially with high levels of behavioral engagement (e.g., long duration, frequent engagement).

Data analysis

Machine learning model construction

Descriptive statistics were conducted using SPSS 27.0 to summarize sample characteristics. Subsequent machine learning modeling was implemented in Python (version 3.9, using Scikit-learn and XGBoost libraries). Following the best-practice guidelines for prediction modeling as outlined in the TRIPOD statement (Collins et al., 2015), we adopted a systematic workflow to ensure model transparency and reproducibility. Given that the relations between the predictors and exercise dependence were likely to be non-linear, we trained four classifiers with four different modeling assumptions: multinomial logistic regression (MLR), random forest (RF), eXtreme gradient boosting (XGBoost), and multilayer perceptron (MLP). These algorithms were selected to capture both simpler and more complex patterns in the data, so that key factors related to exercise dependence in college students could be captured more fully. A stacking ensemble was finally used to combine the four base models in order to obtain a more stable and accurate classifier.

Four machine learning algorithms and procedures

The modeling steps comprised: (1) Data preprocessing: Continuous variables were Z-score standardized and categorical variables were dummy coded. (2) Data splitting: The data were split randomly into a training set (64%), a test set (16%), and a hold-out set (20%). This allocation effectively retains 80% of the data for model development (encompassing training and internal tuning) while reserving a strictly independent 20% hold-out sample to provide an unbiased estimate of generalization performance (Varma and Simon, 2006; Yarkoni and Westfall, 2017). (3) Model design and training: Five machine-learning algorithms are adopted, including logistic regression (LR), random forests (RF), XGBoosting (XGB), a fully-connected neural network model (MLP) and one stacking ensemble. Hyperparameters settings were chosen in a way to guarantee reproducibility and convergence:

• LR: With multinomial strategy (multi_class = “multinomial”) and maximum number of iterations 1,000 (max_iter = 1,000) to ensure convergence of the solver.

• RF: Used the default parameters and set the random seeds to get deterministic results.

• XGB: Set up for multi-class log loss optimization (eval_metric = “mlogloss”).

• MLP: A two hidden layers model with 64 and 32 neurons (hidden_layer_sizes = (64, 32)), trained at most for 1,000 iterations.

• Stacking classifier: Combine the above four models (LR, RF, XGB and MLP) as base models and a logistic regression model (same hyperparameters as a single model of LR) is used as meta-model.

(4) Model assessment: Model performance was measured on test set using accuracy, precision, recall and F1-score. (5) Robustness validation: For alleviation of sampling bias, the training-test process was repeated 100 times with random splits. The last model of all models was validated in the holdout set to evaluate generalization.

Evaluation metrics

To strictly and comprehensively evaluate the performance of the machine learning models, we employed several key metrics: area under the curve (AUC), accuracy, precision, recall, and average precision (AP). Additionally, model stability was assessed using the standard deviation (SD) of accuracy scores across cross-validation folds.

ROC curve and AUC: The area under the curve (AUC) is calculated as the integral of the receiver operating characteristic (ROC) curve, which plots the true positive rate (TPR) against the false positive rate (FPR). The AUC is defined as:

AUC = 0 1 TPR ( FPR ) d ( FPR )

where TPR = TP TP + FN ( sensitivity ) , FPR = FP FP + TN (1-sensitivity)

Confusion matrix-based metrics: The calculations for precision, recall, and accuracy are derived from the confusion matrix, which consists of true positives (TR), false positives (FP), true negatives (TN), and false negatives (FN).

Accuracy: Accuracy represents the overall correctness of the model and is calculated as the ratio of correctly predicted observations to the total observations:

Accuracy = TP + TN + FP + FN TP + TN

Precision: Precision measures the accuracy of positive predictions. It is defined as the ratio of correctly predicted positive observations to the total predicted positive observations:

Precision = TP + FP TP

Recall: Recall (also known as sensitivity) measures the ability of the model to identify all relevant cases (positive samples). It is the ratio of correctly predicted positive observations to all observations in the actual class:

Recall = TP + FN TP

F1-score: F1-score is employed as a comprehensive metric to evaluate the overall performance of the model. Defined as the harmonic mean of precision and recall, the F1-score provides a balanced assessment by taking both false positives and false negatives into account. It is particularly useful for comparing classifiers, as it penalizes extreme values in either precision or recall. The formula is calculated as follows:

F 1 score = 2 × Precision + Recall Precision × Recall

Average precision: Average precision (AP) summarizes the precision-recall curve into a single value representing the average of all precision values calculated at each threshold level. It is calculated as the weighted mean of precisions achieved at each threshold, with the increase in recall from the previous threshold used as the weight:

AP = 0 ( R n R n 1 ) P n

where R n and R n 1 are the precision and recall at the n-th threshold, respectively.

Model stability: To evaluate model stability and robustness across the cross-validation folds, we utilized the standard deviation (SD) of the accuracy scores. A lower SD indicates higher stability: SD = 1 k 1 i = 1 k ( ACC i ACC ¯ ) 2

where k is the number of folds (or repetitions), ACC i is the accuracy of the i-th fold, and ACC ¯ is the mean accuracy.

Integration and evaluation

In the end, predictions of the four base models were stacked in a stacking ensemble where logistic regression acted as meta-learner, performance was checked using five-fold cross-validation. Model performance was described as AUC and SHAP. The combination of these set-up allowed identifying both model accuracy and the relative significance of sensitive parameters.

Results

Confirmatory factor analysis

Confirmatory factor analysis (CFA) was conducted via the AMOS 26.0 program to evaluate the structural validity of both the Weight Bias Subscale and Exercise Dependence Scale. The goodness of fit was checked against well-accepted indices and criteria (Hair et al., 2009; Hu and Bentler, 1999). The model demonstrated an excellent fit to the data, with reported indices as follows: χ2/df = 1.033 (suggesting a good fit as it is <3.0; McIver and Carmines, 1981), RMSEA = 0.023 (<0.06; Hu and Bentler, 1999), CFI = 0.983 (>0.90; Bentler, 1990), IFI = 0.996, TLI = 0.978 (>0.90; Tucker and Lewis, 1973), and PNFI = 0.654 (>0.50; Mulaik et al., 1989). This suggests that the measurement model was statistically sound and suitable for further machine learning analyses (see Figure 1 and Table 2).

Figure 1
Diagram illustrating weight bias with values linked to five nodes: w1 to w5. Arrows from a central

Figure 1. Confirmatory factor analysis of the weight bias scale.

Table 2
www.frontiersin.org

Table 2. Model fit indices for the weight bias scale.

Descriptive statistics

Participant grouping by exercise dependence

For machine learning classification purposes, participants (N = 2,745) were categorized into three levels based on their Exercise Dependence Scale scores: no dependence symptoms (No-DS; n = 952, 34.68%), non-dependence symptoms (NDS; n = 874, 31.84%), and at risk for exercise dependence (ARED; n = 919, 33.48%). However, a chi-square test showed that the proportions of participants across these categories did not differ significantly (χ2 = 3.36, p > 0.05). Therefore, the three-level structure was retained for the machine learning analyses, whereas descriptive and inferential statistics were reported for the total sample rather than by group (see Figure 2).

Figure 2
Bar chart showing sample distribution by exercise dependence group: 0 (low), 1 (middle), and 2 (high). Each group has similar sample sizes around 850, with group 1 slightly lower.

Figure 2. Distribution of participants by exercise dependence total score groups.

Differences in exercise dependence behavior across demographic factors

To explore whether exercise dependence varied by gender, year in school and BMI, we ran independent samples t-tests and one-way ANOVAs. As shown in Table 3, male students scored higher than females on exercise identity, weight bias, and exercise dependence (all p-values <0.001). For Grade difference, the first year and the second years’ students reported higher scored than students in other grades. For BMI, students in the normal-weight range scored shows higher exercise identity, higher weight bias, and higher exercise dependence than both underweight and overweight groups. In addition, overweight/obese students reported higher exercise dependence than underweight students. Overall, these findings indicate clear demographic differences in factors linked to exercise dependence.

Table 3
www.frontiersin.org

Table 3. Difference analysis of college students’ exercise identity, weight bias and exercise dependence behavior.

Exercise dependence behavior and its associated factors in college students

To establish the major features of exercise dependence behavior and its associated variables, we calculated correlation coefficients between the demographic factors, physical measures, psychological predictors and EDS-R scores (see Figure 3). The analysis revealed that gender exhibited a significant negative correlation with height (r = −0.74), weight (r = −0.54), and BMI (r = −0.51), reflecting significant sex differences in body shape and weight distribution. Moderate positive correlations were observed among exercise intensity, duration, and frequency (r = 0.20–0.29), suggesting a consistent activity pattern in the multidimensional characteristics of college students’ exercise behaviors. Most EDS-R items correlated strongly with one another, with several coefficients above 0.60; in particular, A15–A17 (r = 0.80), A17–A19 (r = 0.80), and A15–A19 (r = 0.75) formed a tightly linked cluster, suggesting a well-coupled dependence structure. Exercise identity and weight bias were moderately to strongly positively associated with the core exercise dependence items (e.g., A17, A13, A11, A21; r-values ≈ 0.30–0.50). Overall, correlations were generally medium to high within the dependence scale (r = 0.60–0.70), while strong correlations (r > 0.75) were concentrated in a few potential core items. These findings provide evidence of the internal consistency of the ECDS and emphasize the structure beneath exercise dependence behaviors, as well as weak relationships with physiological measures. This pattern supports the approach to use machine learning to determine a relevant predictor.

Figure 3
Correlation heatmap depicting relationships among variables such as grade, gender, height, weight, BMI, exercise identity, and exercise-related behaviors. Positive correlations are shown in red, negative in blue. The color intensity represents the correlation strength. Values and significance levels are marked in each cell.

Figure 3. Correlation heatmap of demographic, predictor variables related to exercise dependence.

Machine learning analysis

Overview of model evaluation metrics

In summary, considering that we wanted an exhaustive evaluation of the predictive models, several performance metrics were used: area under the receiver operating characteristic curve (AUC), precision, recall and F1-score. AUC represents the overall ability of the model to differentiate between classes (0.5 means by chance, and 1.0 denotes perfect discrimination). Precision measures the fraction of true high-risk students among the predicted ones, while recall (sensitivity) measures how many real positives are captured from all actual positive instances. The F1-score takes the harmonic mean of precision and recall, which allows a balance in measuring model’s performance. Also, we used both ROC and precision-recall (PR) curves. ROC curves demonstrate the trade-off between sensitivity and specificity over decision thresholds, but may give an overly optimistic view in unbalanced datasets. PR curves are a useful supplement to ROC analysis, as they concentrate on the performance of the positive class, and offer a more stringent evaluation of model applicability by avoiding false short cuts through genuine negatives.

Analysis of AUC values: discrimination performance of algorithms

As illustrated in Figure 4 and Table 4, all five predictive models demonstrated excellent discrimination. To rigorously evaluate the generalization ability of the proposed models, we report the performance metrics exclusively based on the independent test set using repeated (n = 100) hold-out validation.

Figure 4
Dot plot showing AUC scores for five classifiers: Logistic Regression, Random Forest, XGBoost, Multi-Layer Perceptron, and Stacking across 100 repeats. Colored dots represent scores for each method, with Stacking Mean indicated by a purple dashed line.

Figure 4. AUC scores of five machine learning models across 100 repeats.

Table 4
www.frontiersin.org

Table 4. Performance metrics of machine learning algorithms on the independent test set.

The stacking learning framework demonstrated superior performance, achieving a maximum mean AUC of 0.96 and an accuracy of 0.91, outperforming individual base classifiers (LM, RF, and XGBoost). While the multilayer perceptron (MLP) showed a slightly lower accuracy of 0.87 compared to the ensemble method, it still maintained a robust predictive capability (AUC = 0.95).

All models were significantly superior to chance level (50%, p < 0.001). The confidence intervals across Table 4 confirm that these models are not overfitting or simply memorizing the training data, but rather learning stable patterns that generalize well to novel instances. The AUC distribution further indicates that, although single models performed well, stacking learning provided the best clear separation for identifying different risk levels of exercise dependence.

Precision-recall performance: reliability across risk levels

Figure 5 shows the precision-recall (PR) curves for the four base models and the stacking model for each of the three exercise dependence levels. Average precision was high for every model and for each risk group, with the PR curves were all pushed towards the upper-right corner. All models performed well in three risk levels, with the PR curves approaching the ideal point (1, 1). The Stacking ensemble model did better than any single model, with AP values of 0.9922, 0.9632, and 0.9928 for the low, medium, and high-risk groups, respectively. These high AP scores suggest that the models maintain high precision even at high recall levels, effectively minimizing false positives and confirming that the high accuracy is not an artifact of overfitting to the majority class.

Figure 5
Precision-Recall curves for different models: a) Logistic Regression, b) Random Forest, c) XGBoost, d) Multi-Layer Perceptron, and e) Stacking. Each graph displays three curves for low, middle, and high groups, illustrating precision versus recall performance.

Figure 5. PR curves of machine learning models for exercise dependence risk. (a) Precision-recall curve of the logistic regression model. (b) Precision-recall curve of the random forest model. (c) Precision-recall curve of the XGBoost model. (d) Precision-recall curve of the multilayer perceptron model. (e) Precision-recall curve of the stacking model.

Classification accuracy and confusion matrix analysis results

Figure 6 shows the confusion matrices for the four models. Cells on the main diagonal represent correct classification; off-diagonal cells indicate errors. All four models distinguished the three risk groups effectively. Logistic regression and the multilayer perceptron (MLP) reached the highest accuracy (accuracy = 0.98), random forest reached 0.97, and XGBoost reached 0.96. Across the three risk groups, all models classified most students correctly. MLP and logistic regression had the highest correct counts, random forest was close, and XGBoost was slightly behind.

Figure 6
Four confusion matrices of different models. a) Logistic Regression shows high accuracy of 0.93 with most true positives in the diagonal. b) Random Forest has an accuracy of 0.92, similar diagonal pattern. c) XGBoost also displays an accuracy of 0.92, with minor misclassifications. d) Multi-Layer Perceptron, also at 0.92 accuracy, follows the same pattern, highlighting effective classification in all models.

Figure 6. Confusion matrix diagrams for four machine learning models. (a) Logistic regression model. (b) Random forest model. (c) XGBoost model. (d) Multilayer perceptron model.

Model stability and robustness verification

To alleviate the overfitting and validate model calibration, we further carried out a rigorous robustness test through 100 repetitions of 5-fold cross-validation. The accuracy distribution of four single models and stacked model is shown in Figure 7. The results indicate excellent stability. Performance of the stack is the most accurate and stable, having the highest median accuracy (≈0.94) and smaller IQR (0.93–0.95). Logistic regression had relatively low variances as well. The small deviation of accuracy over the 100 runs indicates a well-calibrated model which is robust against random splits in the training data.

Figure 7
Box plot comparing the accuracy of five machine learning models: Logistic Regression, Random Forest, XGBoost, Multi-Layer Perceptron, and Stacking. Logistic Regression and Stacking exhibit higher median accuracy around 0.94, while Random Forest shows the lowest at approximately 0.92. XGBoost and Multi-Layer Perceptron are in between, with the presence of outliers noted in Random Forest and XGBoost.

Figure 7. Accuracy distribution of five machine learning models over 100 repeated 5-fold cross-validations.

In contrast, the median accuracy for random forest, XGBoost and MLP models was slightly lower (around 0.92–0.93) and somewhat more variable—in some cases down to around 0.90 but also retained high baseline performance. The stronger stability provided by the stacking model and logistic regression is consistent with their ranking in interpreting the conclusion, that high level prediction power results from systematic rather than overfitting pattern of findings.

Feature importance analysis in the random forest model

To understand how the model has “decided” to predict the outcomes, the team performed an analysis of the feature importance of the predictors. For RF models, feature importance is commonly measured by the mean decrease in impurity approach, in which the impurity is determined by splitting a specific tree on a certain variable, and then averaged across all trees. A higher score reflects a more important role of the variable in classification according to the factor. Figure 8 presents the most important predictive variables estimated by the RF model for exercise dependence risk, encompassing the top 20 by contribution. The data in the chart reflect feature importance scores on the x axis and predictors under consideration on the y axis. The most important feature included specific indicators of exercise dependence severity. More specifically, the three most important variables were A17 (“Increase duration,” with the score of A17 = 0.1179), A13 (“Most leisure time,” with the score of A13 = 0.1049), and A21 (“Exceeds planned time,” with the score of A21 = 0.0957). These variables with the 3 top importance scores for the model emphasized more exercise due to tolerance and the consumption of more time to be critical for the identification of exercise dependence development. Additionally, the most important variables were A11 (“Not reduce frequency”); A15 (“Avoid tension”); A19 (“Avoid socializing”); A5 (“Prefer over social”); A9 (“With injury”); A7 (“Exceeds intended time”); A3 (“Increase intensity”) and A1 (“Avoid irritability”).

Figure 8
Bar chart showing feature importance scores. The highest scores include A17 (increase duration) at 0.1179, A13 (most leisure time) at 0.1049, and A21 (exceeds planned time) at 0.0957. Scores decrease with attributes like gender and academic year having the lowest scores, around 0.0041 to 0.0045.

Figure 8. Random forest feature importance ranking (top 20).

Overall, withdrawn symptoms and preferred activities over socialization show major importance for dependent behaviors. In contrast, psychological and physiological variables, such as weight bias internalization, exercise identity, and BMI, show low importance scores. However, they are crucial for maximizing the overall prediction accuracy of the model. Lastly, demographic factors such as gender and year of study show the lowest importance scores. This means that demographic components had minimal prediction power for exercise dependence in this specific group.

Shapley value analysis of predictors in the final model from the stacking learning method

To further examine how each predictors contributed to exercise dependence, we computed for all variables in the stacked model using the holdout set and get the SHAP values. All predictors showed SHAP values significantly above zero (p-values <0.05), indicating that each variable added unique information to the model’s predictions. According to the ranking of Shapley values, Table 5 and Figure 9 summarizes the SHapley Additive exPlanations (SHAP) values for key predictors whose marginal contributions to the prediction of exercise dependence risk levels were greater than or equal to 1%. The clear differences between variables indicate that some predictors played a much larger role than others in estimating exercise dependence risk in college students. The predictors could be grouped into four categories based on their importance, from highest to lowest.

Table 5
www.frontiersin.org

Table 5. Mean SHAP values of predictors across exercise dependence risk groups in the stacking ensemble model.

Figure 9
Bar chart showing various factors related to exercise habits and characteristics across three groups: Low (blue), Middle (orange), and High (yellow). Factors include increased duration, leisure time, social preferences, exercise identity, and demographics. High group consistently shows larger values across most factors.

Figure 9. Summary of SHAP values in the random forest model.

The first group is the core contributing variables (Shapley values ≥5%), including A17, A13, A11, and A21, with Shapley values significantly higher than other variables, reaching approximately 17.5, 16, 15, and 14%, respectively. In the low-risk group, items A17 (0.072), A11 (0.068), A21 (0.053), and A13 (0.051) showed noticeably higher SHAP values than the remaining predictors, indicating that they played the main role in the model for this group. The same items stayed at the top in the medium- and high-risk groups (mean SHAP ≈ 0.041–0.066), which supports their status as core indicators. In the high-risk group, for example, A13 and A21 reached 0.066 and 0.057, respectively, well above most other variables. Paired-sample t tests did not detect significant differences in the average SHAP values of A17 and A21 across groups (t = 1.82, df = 2, p > 0.05), but both items remained consistently higher than the rest, confirming their central contribution to distinguishing exercise-dependence levels in this sample.

The second group of variables had Shapley values between 2 and 4% and included A9, A19, A15, A5, A7, A3 and A1. These made a moderate contribution, with A19 and A9 standing out in the high-risk group (0.047 and 0.046), close to or slightly above 4%. Their values fluctuated less than those of the core items and form a clear step below the first tier. This middle level showed fairly even predictive strength, which made the model’s importance structure appear graded.

The third group of variables had Shapley values between about 0.01 and 0.02, consisted mainly of the total scores for exercise identity and weight bias. In the high-risk group, their SHAP values were 0.018 and 0.009, and although slightly lower in the other groups, they remained above 0.01. Thus, while weaker than the first two tiers, these variables still added explanatory value to the model. These results indicate that although exercise identity and weight bias totals were weaker than the first two tiers, they still added useful explanatory power to the model.

The fourth group consisted of variables with low marginal contributions (SHAP <1%), primarily demographic and basic physical measurements such as height, weight, BMI, gender, and academic year. For example, height and weight had SHAP values below 0.01 in every group, and BMI, gender, and year of study were even lower. Although these variables reached statistical significance, their marginal contribution to the model was small. This suggests that college students’ primary predictors of exercise dependence behavior lie in psychological and behavioral factors rather than in basic demographic or physical measurements.

The Shapley value analysis results in this study demonstrate that the key predictors of exercise dependence behavior among college students are mainly concentrated in the core variables A17 (Increase duration), A13 (Most leisure time), A11 (Not reduce frequency), and A21 (Exceeds planned time). These variables not only align closely with the core pathological features of exercise dependence, but also exhibit dominant contributions in the Stacking ensemble model. Additionally, scale-based variables such as the total scores for exercise identity and weight bias, along with certain behavioral indicators, provided auxiliary support to the model’s predictive power. In contrast, demographic and physical measurement indicators (e.g., height, weight and BMI) played a relatively limited role.

Discussion

Prevalence, group differences, and ML-based identification of key predictors among college students

The present study was motivated by two gaps in the literature: (a) exercise dependence has rarely been examined in Chinese university contexts from a theory-driven perspective, and (b) applications of deviance regulation theory (DRT) have seldom been combined with machine-learning models capable of capturing non-linear identity processes. Our findings address both gaps by linking SHAP-derived importance patterns to core DRT constructs of identity regulation and norm-contingent motivation. This study seeks to provide a characterization of how prevalent exercise dependence is in a large Chinese college student sample, across basic demographic categories, and which psychological/behavioral identifiers are most useful for demarcating type patterns. Drawing on high volume survey data and machine learning techniques (Pedregosa et al., 2011), we provided important insights into the complex relationship patterns between psychological and demographic factors from DRT. The main findings are presented in the following below.

First, the present study revealed that some 33% of university students are potentialy exposed to exercise dependence which approximate similar results between domestic and foreign literature (Allegre et al., 2006; Garman et al., 2004; Živkov et al., 2022). The data imply that when the frequencies and intensities of exercising become elevated, a sub-group of students cross from healthy involvement to unhealthy dependence, which is marked by compulsive control and inflexible routines. In the absence of effective self-regulation, or environmental cues to reduce these behaviors, they become established, and early identification is essential. This result reflects the widespread but unreported phenomenon of exercise dependence in universities and confirms earlier findings about increasing prevalence of exercise addiction (Berczik et al., 2014; Symons Downs et al., 2019). Moreover, we obtained better classification performance as compared to traditional psychometric cut-offs by using an ensemble of machine learning models (logistic, random forest, tree-based XGBoost and MLP). The ensemble learning model was the best predictor (mean AUC = 99.16%). The model had excellent predictive discrimination when the AUC was converted to Cohen’s d effect sizes (>2.0). Unlike the previous studies that have used only linear statistical techniques, our method also demonstrates that data-driven models can successfully disentangle non-linear aspects of exercise behaviors and can be used as a powerful instrument to distinguish high risk individuals from general population (Goh et al., 2022; Gulhane and Sajana, 2021).

Second, demographic profiling revealed that male students, freshmen and sophomores, and those with normal BMI exhibited significantly higher exercise dependence scores compared to their counterparts. These results corroborate Garman’s et al. (2004) observation that the early university years represent a critical window of vulnerability for developing exercise dependence. According to correlation analyses, dependence was moderately related to psychological variables—partially with exercise identity and weight cognition bias—but not with physiological measurements (e.g., BMI). This is consistent with the interpretation that EDS reflects mainly psychological reinforcement and habit formation (Hausenblas and Symons Downs, 2002) At least from DRT perspective, however, these demographic and psychological trends imply those students could be at risk for perceiving their exercise behavior as a necessary tool to control perceived self-discrepancies.

Third, Shapley value analysis disentangled the relative importance of specific predictors, revealing that tolerance and time-allocation behaviors—specifically items A17 (Increasing duration), A13 (Spending leisure time on exercise), A11 (Inability to reduce frequency), and A21 (Exercising longer than planned)—were the paramount predictors, with contributions ranging from 14 to 17.5%. These findings strongly implicate loss of control and excessive time investment as the core pathology of the disorder, consistent with previous literature (Allegre et al., 2006; Symons Downs et al., 2019). Critically, the prominence of A17 accords with predictions based on deviation regulation theory. When people notice that they do not meet an attractive standard (e.g., physical fitness, body image availability), in the DRT model, they enhance whatever behavior differs distinctly most from what is believed to be necessary so that they will fit this standard or return to it (Blanton and Christie, 2003). In the context of exercise dependence, this manifests as a maladaptive feedback loop: to eliminate cognitive dissonance regarding body image or performance, individuals escalate their investment, leading to the observed “tolerance” effect. In DRT terms, these patterns imply that students escalate exercise precisely because they perceive themselves as deviating from an internalized ideal, thereby creating a norm-contingent feedback loop in which identity threats are managed through increasingly rigid exercise routines. The high Shapley values for A13, A11, and A21 further illustrate the collapse of self-regulatory systems, where the motivation to exercise overrides other life priorities (Köpetz et al., 2013). Secondary predictors (e.g., A9, A19; contribution 2–4%) and tertiary psychological factors (e.g., exercise identity; contribution 1–2%) provided supplementary predictive power. One potential hierarchical model of this nature is such that overarching psychological forces lay the groundwork, but it is students’ observable behaviors regarding a “loss of control” that are the only culturally sanctioned markers of at-risk status.

On the other hand, in contrast to the biases of psychological factor preferences discussed above, basic demographic and physiological measurement-related factors held consistent below 1% Shapley values. This discovery contributes a particularly important theoretical insight in that it implies the constructs of exercise dependence are not intrinsic to certain demographic profiles, but rather reflective of cognitive self-regulatory processes. More specifically, these findings expand upon DRT as a theoretical framework and specify the process whereby “identity regulation” and “stigma avoidance” work. DRT postulates that people conform on norms to develop a positive self concept, or avoid the “stigma of nonconformity. That psychological, but not demographic, predictors had relatively high feature importance suggests exercise dependence represents a compensatory phenomenon. Excessive exercise among students probably has less to do with biological factors (e.g., the BMI) than maintaining identity in the face of perceived social norms, namely warding off stigma internalized from being seen as “lazy” or “unfit.” This reading goes beyond descriptive validation, and suggests a behavioral dependence that is, at its core, a paradoxical maladaptive byproduct of successful identity management.

This trend is also consistent with recent machine learning work on behavioral addictions such as Hong et al. (2024) about smartphone addiction in which self-regulation and motivational control play a key role inherent to more static traits. But it contrasts with the results of Ferrer et al. (2011) and Dvorak et al. (2016, 2018) related to substance use disorder, while the focus in these studies was mainly on socio-demographic risk factors. Our finding that the distinctive psychological dominancy (exercise dependence) it represents contributes to a literature review in sports psychology, adding information rather than demographic vulnerability, what is specifically for exercise dependence in particular—norm-contingent motivation.

While the direct effects of demographic factors are unimpressive; yet, their subtle possible roles should be taken into account. For example, gender or grade may not be an overarching determinant of dependence, but rather act as a moderating factor on the strength with which psychological pressures affect individuals (e.g., weight stigma effects may differ by gender). Tentative forays demonstrated that the role of core predictors such as A17 was slightly more pronounced in males and first-year students, indicating possible moderator effects. Future studies may use multi-level linear models to better disentangle these associations.

Innovative use of machine-learning modeling and interpretation

Here, we combined ensemble machine-learning models to break down exercise dependence. As noted by Bartlett et al. (2014), these models suggest enhanced exclusive tracking of behavioral data via non-linear content that traditional linear counterparts might miss: importantly and uniquely, our SHAP translation of “black-box” algorithms allowed us to create theoretically interpretable outputs and trace a prediction back to distinct “norm-deviation” behaviors and identity indicators. This interpretability serves as the bridge between advanced data science and actionable intervention planning. This insight allows for the operationalization of findings into two distinct practical avenues. First, in university counseling settings, the high-ranking item-level predictors can be utilized to develop rapid screening protocols. Instead of administering full-length diagnostic batteries, counselors can prioritize these “red flag” questions to efficiently triage students at risk (Lichtenstein et al., 2014b). Second, in athletic settings, these insights enable coaches to distinguish between high-performance training and pathological dependence. By monitoring specifically identified behaviors—such as the inability to reduce exercise intensity despite injury—staff can implement targeted, symptom-specific interventions rather than generic exercise reduction strategies, thereby preventing the escalation of dependence (Nogueira et al., 2018; Szabo et al., 2015).

Limitations and future directions

This study has some limitations, although it makes a significant contribution. First of all, the model did not involve variables such as social support or peer norms so some variance in DRT’s “social context” component was unaccounted for. Second, self-report questionnaires may be subject to reporting biases. Third, the sample is not regionally diverse. Studies combining objective physiological assessments, inclusion of multi center populations and incorporating data from wearable technology to capture identity regulation through exercise in vivo are also warranted.

Conclusion

To the best of our knowledge, for the first time various artificial intelligence models are used to investigate exercise dependence in Chinese college students. Using data on 2,745 students, we examined prevalence trends and demographic correlates as well as the main predictors. 33% of the students exhibited risk signals, whereas males and lower class levels and normal BMI students had a higher proportion of risk. The AUC of the 4-classifier-based stacked model was very high (99.16%), allowing accurate risk stratification. The model-interpretation results demonstrated that for prediction, behavioral/psychological items were more informative than demographics. There are 4 of them: extending exercise time to achieve the desired effect (A17), spending maximum time with leisure activity-exercising way (A13), not able to reduce frequency (A11) and exercising longer than plan script A21. These measures are directly related to excessive control and excess investment. They are also congruent with the compensatory regulation theory of exercise dependence.

The study, in general, highlights the promise of artificial intelligence towards capacity-building within sport and exercise psychology by providing new methods for detecting risk and developing evidence-informed intervention. It would be a valuable task for the future to refine current models, validate them in other university settings, and integrate them with more realistic intervention scenarios aimed at enhancing healthy exercise behaviors in college students.

Data availability statement

The original contributions presented in the study are included in the article/Supplementary material, further inquiries can be directed to the corresponding author.

Ethics statement

The studies involving humans were approved by the Academic Ethics Committees of Yangzhou University (Ethics No. YZUHL2020102), and informed consent was obtained from the students’ respective colleges and administrative units. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

YD: Formal analysis, Project administration, Validation, Writing – original draft. WL: Conceptualization, Methodology, Supervision, Writing – review & editing. MS: Conceptualization, Methodology, Validation, Writing – review & editing. YR: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Supervision, Validation, Writing – original draft, Writing – review & editing.

Funding

The author(s) declared that financial support was not received for this work and/or its publication.

Acknowledgments

The author(s) would like to acknowledge that this research was independently completed without external funding or assistance.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that Generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyg.2026.1743725/full#supplementary-material

References

Abdul Rahman, H., Kwicklis, M., Ottom, M., Amornsriwatanakul, A., Abdul-Mumin, K. H., Rosenberg, M., et al. (2023). Machine learning-based prediction of mental well-being using health behavior data from university students. Bioengineering 10:575. doi: 10.3390/bioengineering10050575,

PubMed Abstract | Crossref Full Text | Google Scholar

Allegre, B., Souville, M., Therme, P., and Griffiths, M. (2006). Definitions and measures of exercise dependence. Addict. Res. Theory 14, 631–646. doi: 10.1080/16066350600903302

Crossref Full Text | Google Scholar

Anderson, D. F., and Cychosz, C. M. (1994). Development of an exercise identity scale. Percept. Mot. Skills 78, 747–751. doi: 10.2466/pms.1994.78.3.747

Crossref Full Text | Google Scholar

Bartlett, M. S., Littlewort, G. C., Frank, M. G., and Lee, K. (2014). Automatic decoding of facial movements reveals deceptive pain expressions. Curr. Biol. 24, 738–743. doi: 10.1016/j.cub.2014.02.009,

PubMed Abstract | Crossref Full Text | Google Scholar

Bentler, P. M. (1990). Comparative fit indexes in structural models. Psychol. Bull. 107:238. doi: 10.1037/0033-2909.107.2.238

Crossref Full Text | Google Scholar

Berczik, K., Griffiths, M. D., Szabó, A., Kurimay, T., Urban, R., and Demetrovics, Z. (2014). “Exercise addiction” in Behavioral addictions (London: Elsevier), 317–342.

Google Scholar

Berczik, K., Szabó, A., Griffiths, M. D., Kurimay, T., Kun, B., Urbán, R., et al. (2012). Exercise addiction: symptoms, diagnosis, epidemiology, and etiology. Subst. Use Misuse 47, 403–417. doi: 10.3109/10826084.2011.639120,

PubMed Abstract | Crossref Full Text | Google Scholar

Blanton, H., and Christie, C. (2003). Deviance regulation: a theory of action and identity. Rev. Gen. Psychol. 7, 115–149. doi: 10.1037/1089-2680.7.2.115

Crossref Full Text | Google Scholar

Blanton, H., Stuart, A. E., and Van den Eijnden, R. J. J. M. (2001). An introduction to deviance-regulation theory: the effect of behavioral norms on message framing. Personal. Soc. Psychol. Bull. 27, 848–858. doi: 10.1177/0146167201277007

Crossref Full Text | Google Scholar

Breiman, L. (2001). Random forests. Mach. Learn. 45, 5–32. doi: 10.1023/A:1010933404324

Crossref Full Text | Google Scholar

Bzdok, D., Altman, N., and Krzywinski, M. (2018). Statistics versus machine learning. Nat. Methods 15, 233–234. doi: 10.1038/nmeth.4642,

PubMed Abstract | Crossref Full Text | Google Scholar

Calestine, J., Bopp, M., Bopp, C. M., and Papalia, Z. (2017). College student work habits are related to physical activity and fitness. Int. J. Exerc. Sci. 10, 1009–1017. doi: 10.70252/XLOM8139,

PubMed Abstract | Crossref Full Text | Google Scholar

Chang, C.-C., Pan, M.-C., and Shu, S.-T. (2019). Relationship between self-presentation, exercise dependence, and perceived body image. Int. J. Sport Exerc. Psychol. 17, 617–629. doi: 10.1080/1612197X.2018.1444078

Crossref Full Text | Google Scholar

Chen, P., Li, F., and Harmer, P. (2019). Healthy China 2030: moving from blueprint to action with a new focus on public health. Lancet Public Health 4:e447. doi: 10.1016/S2468-2667(19)30160-4

Crossref Full Text | Google Scholar

Chen, T., and Guestrin, C. (2016). XGBoost: a scalable tree boosting system. Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. 785–794

Google Scholar

Collins, G. S., Reitsma, J. B., Altman, D. G., and Moons, K. G. (2015). Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): the TRIPOD statement. Br. J. Surg. 102, 148–158. doi: 10.1002/bjs.9736

Crossref Full Text | Google Scholar

de Coverley Veale, D. M. W. (1987). Exercise dependence. Br. J. Addict. 82, 735–740. doi: 10.1111/j.1360-0443.1987.tb01539.x,

PubMed Abstract | Crossref Full Text | Google Scholar

Dvorak, R. D., Pearson, M. R., Neighbors, C., Martens, M. P., Stevenson, B. L., and Kuvaas, N. J. (2016). A road paved with safe intentions: increasing intentions to use alcohol protective behavioral strategies via deviance regulation theory. Health Psychol. 35, 604–613. doi: 10.1037/hea0000327,

PubMed Abstract | Crossref Full Text | Google Scholar

Dvorak, R. D., Raeder, C. A., Kramer, M. P., Sargent, E., Stevenson, B. L., and Helmy, M. (2018). Using deviance regulation theory to target marijuana use intentions among college students. Exp. Clin. Psychopharmacol. 26, 29–35. doi: 10.1037/pha0000159,

PubMed Abstract | Crossref Full Text | Google Scholar

Ferrer, R. A., Dillard, A. J., and Klein, W. M. P. (2011). Projection, conformity and deviance regulation: a prospective study of alcohol use. Psychol. Health 27, 688–703. doi: 10.1080/08870446.2011.620106,

PubMed Abstract | Crossref Full Text | Google Scholar

Forbush, K. T., Wildes, J. E., Pollack, L. O., Dunbar, D., Luo, J., Patterson, K., et al. (2013). Development and validation of the Eating Pathology Symptoms Inventory (EPSI). Psychol. Assess. 25, 859–878. doi: 10.1037/a0032639,

PubMed Abstract | Crossref Full Text | Google Scholar

Gallagher, K. M., and Updegraff, J. A. (2012). Health message framing effects on attitudes, intentions, and behavior: a meta-analytic review. Ann. Behav. Med. 43, 101–116. doi: 10.1007/s12160-011-9308-7,

PubMed Abstract | Crossref Full Text | Google Scholar

Garman, J. F., Hayduk, D. M., Crider, D. A., and Hodel, M. M. (2004). Occurrence of exercise dependence in a college-aged population. J. Am. Coll. Heal. 52, 221–228. doi: 10.3200/JACH.52.5.221-228,

PubMed Abstract | Crossref Full Text | Google Scholar

Goh, Y. S., Yong, J. Q. Y. O., Chee, B. Q. H., Kuek, J. H. L., and Ho, C. S. H. (2022). Machine learning in health promotion and behavioral change: scoping review. J. Med. Internet Res. 24:e35831. doi: 10.2196/35831,

PubMed Abstract | Crossref Full Text | Google Scholar

Gong, Y., Yang, H., Bai, X., Wang, Y., and An, J. (2023). The relationship between physical exercise and smartphone addiction among Chinese college students: the mediating effect of core self-evaluation. Behav. Sci. 13:647. doi: 10.3390/bs13080647,

PubMed Abstract | Crossref Full Text | Google Scholar

Griffiths, M. D., Szabo, A., and Terry, A. (2005). The exercise addiction inventory: a quick and easy screening tool for health practitioners. Br. J. Sports Med. 39:e30. doi: 10.1136/bjsm.2004.017020,

PubMed Abstract | Crossref Full Text | Google Scholar

Gulhane, M., and Sajana, T. (2021). Human behavior prediction and analysis using machine learning—a review. Turk. J. Comput. Math. Educ. 12, 870–876. doi: 10.17762/turcomat.v12i5.1499

Crossref Full Text | Google Scholar

Hair, J. F. Jr., Black, W. C., Babin, B. J., and Anderson, R. E. (2009). Multivariate data analysis. Upper Saddle River, NJ: Prentice Hall.

Google Scholar

Hausenblas, H. A., and Symons Downs, D. (2002). Exercise dependence: a systematic review. Psychol. Sport Exerc. 3, 89–123. doi: 10.1016/S1469-0292(00)00015-7

Crossref Full Text | Google Scholar

Hong, Y., Rong, X., and Liu, W. (2024). Construction of influencing factor segmentation and intelligent prediction model of college students’ cell phone addiction model based on machine learning algorithm. Heliyon 10:e29245. doi: 10.1016/j.heliyon.2024.e29245

Crossref Full Text | Google Scholar

Hu, L., and Bentler, P. M. (1999). Cutoff criteria for fit indexes in covariance structure analysis: conventional criteria versus new alternatives. Struct. Equ. Model. 6, 1–55. doi: 10.1080/10705519909540118

Crossref Full Text | Google Scholar

Khare, S. K., Blanes-Vidal, V., Nadimi, E. S., and Acharya, U. R. (2024). Emotion recognition and artificial intelligence: a systematic review (2014–2023) and research recommendations. Inf. Fusion 102:102019. doi: 10.1016/j.inffus.2023.102019

Crossref Full Text | Google Scholar

Köpetz, C. E., Lejuez, C. W., Wiers, R. W., and Kruglanski, A. W. (2013). Motivation and self-regulation in addiction: a call for convergence. Perspect. Psychol. Sci. 8, 3–24. doi: 10.1177/1745691612457575,

PubMed Abstract | Crossref Full Text | Google Scholar

Lichtenstein, M. B., Christiansen, E., Elklit, A., Bilenberg, N., and Støving, R. K. (2014a). Exercise addiction: a study of eating disorder symptoms, quality of life, personality traits and attachment styles. Psychiatry Res. 215, 410–416. doi: 10.1016/j.psychres.2013.11.010,

PubMed Abstract | Crossref Full Text | Google Scholar

Lichtenstein, M. B., Larsen, K. S., Christiansen, E., Støving, R. K., and Bredahl, T. V. G. (2014b). Exercise addiction in team sport and individual sport: prevalences and validation of the exercise addiction inventory. Addict. Res. Theory 22, 431–437. doi: 10.3109/16066359.2013.875537

Crossref Full Text | Google Scholar

Lichtenstein, M. B., Nielsen, R. O., Gudex, C., Hinze, C. J., and Jørgensen, U. (2018). Exercise addiction is associated with emotional distress in injured and non-injured regular exercisers. Addict. Behav. Rep. 8, 33–39. doi: 10.1016/j.abrep.2018.06.001,

PubMed Abstract | Crossref Full Text | Google Scholar

Li, M., Yang, J., and Ren, Y. (2019). Test of Chinese exercise identity scale in college students. Chin. J. Clin. Psychol. 27, 63–66. doi: 10.16128/j.cnki.1005-3611.2019.01.013

Crossref Full Text | Google Scholar

Livieris, I. E., Kotsilieris, T., Dimopoulos, I., and Pintelas, P. (2018). Decision support software for forecasting patient’s length of stay. Algorithms 11:199. doi: 10.3390/a11120199

Crossref Full Text | Google Scholar

Lundberg, S. M., and Lee, S.-I. (2017). A unified approach to interpreting model predictions. Advances in Neural Information Processing Systems. 30

Google Scholar

Lu, Z. Q. J. (2010). The elements of statistical learning: data mining, inference, and prediction. Oxford: Oxford University Press.

Google Scholar

Mahindru, A., Patil, P., and Agrawal, V. (2023). Role of physical activity on mental health and well-being: a review. Cureus 15:e33475. doi: 10.7759/cureus.33475

Crossref Full Text | Google Scholar

McIver, J., and Carmines, E. G. (1981). Unidimensional scaling. SAGE, Thousand Oaks, CA.

Google Scholar

Miller, K. J., and Mesagno, C. (2014). Personality traits and exercise dependence: exploring the role of narcissism and perfectionism. Int. J. Sport Exerc. Psychol. 12, 368–381. doi: 10.1080/1612197X.2014.932821

Crossref Full Text | Google Scholar

Mulaik, S. A., James, L. R., Van Alstine, J., Bennett, N., Lind, S., and Stilwell, C. D. (1989). Evaluation of goodness-of-fit indices for structural equation models. Psychol. Bull. 105, 430–445. doi: 10.1037/0033-2909.105.3.430

Crossref Full Text | Google Scholar

Nogueira, A., Molinero, O., Salguero, A., and Márquez, S. (2018). Exercise addiction in practitioners of endurance sports: a literature review. Front. Psychol. 9:1484. doi: 10.3389/fpsyg.2018.01484S

Crossref Full Text | Google Scholar

Palermo, M., Staples, C., and Rancourt, D. (2021). Examining the impact of weight bias on the association between exercise identity and maladaptive exercise behaviors. Eat. Behav. 41:101503. doi: 10.1016/j.eatbeh.2021.101503,

PubMed Abstract | Crossref Full Text | Google Scholar

Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., et al. (2011). Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12, 2825–2830.

Google Scholar

Puhl, R., and Suh, Y. (2015). Health consequences of weight stigma: implications for obesity prevention and treatment. Curr. Obes. Rep. 4, 182–190. doi: 10.1007/s13679-015-0153-z,

PubMed Abstract | Crossref Full Text | Google Scholar

Ruegsegger, G. N., and Booth, F. W. (2018). Health benefits of exercise. Cold Spring Harb. Perspect. Med. 8:a029694. doi: 10.1101/cshperspect.a029694,

PubMed Abstract | Crossref Full Text | Google Scholar

Strachan, S. M., Brawley, L. R., Spink, K. S., and Jung, M. E. (2011). Strength of exercise identity and identity‐related social cognitions as predictors of physical activity behavior. Journal of Sport and Exercise Psychology, 33, 671–690. doi: 10.1123/jsep.33.5.671

Crossref Full Text | Google Scholar

Strobl, C., Malley, J., and Tutz, G. (2009). An introduction to recursive partitioning: rationale, application, and characteristics of classification and regression trees, bagging, and random forests. Psychol. Methods 14, 323–348. doi: 10.1037/a0016973,

PubMed Abstract | Crossref Full Text | Google Scholar

Symons Downs, D., MacIntyre, R. I., and Heron, K. E. (2019). “Exercise addiction and dependence” in APA handbook of sport and exercise psychology: exercise psychology (Washington, DC: American Psychological Association).

Google Scholar

Szabo, A., and Demetrovics, Z. (2022). Passion and addiction in sports and exercise. New York: Routledge.

Google Scholar

Szabo, A., Griffiths, M. D., de La Vega Marcos, R., Mervó, B., and Demetrovics, Z. (2015). Methodological and conceptual limitations in exercise addiction research. Yale J. Biol. Med. 88:303,

PubMed Abstract | Google Scholar

Szabo, A., Vega, R. D. L., Ruiz-BarquÍn, R., and Rivera, O. (2013). Exercise addiction in Spanish athletes: investigation of the roles of gender, social context and level of involvement. J. Behav. Addict. 2, 249–252. doi: 10.1556/JBA.2.2013.4.9,

PubMed Abstract | Crossref Full Text | Google Scholar

Terry, A., Szabo, A., and Griffiths, M. (2004). The exercise addiction inventory: a new brief screening tool. Addict. Res. Theory 12, 489–499. doi: 10.1080/16066350310001637363

Crossref Full Text | Google Scholar

Tucker, L. R., and Lewis, C. (1973). A reliability coefficient for maximum likelihood factor analysis. Psychometrika 38, 1–10. doi: 10.1007/BF02291170

Crossref Full Text | Google Scholar

Varma, S., and Simon, R. (2006). Bias in error estimation when using cross-validation for model selection. BMC Bioinformatics 7:91. doi: 10.1186/1471-2105-7-91,

PubMed Abstract | Crossref Full Text | Google Scholar

Wang, X., Lu, C., and Niu, L. (2023). Body image construction and mental health levels among college students: a data survey of Chinese university students. Front. Public Health 11:1268775. doi: 10.3389/fpubh.2023.1268775

Crossref Full Text | Google Scholar

Yang, P., Yu, Q., Montag, C., Becker, B., Cheval, B., Herold, F., et al. (2021). Validation of the Chinese version of the Exercise Dependence Scale-Revised (EDS-R). Int. J. Ment. Health Addict. 21, 1200–1219. doi: 10.1007/s11469-021-00654-4

Crossref Full Text | Google Scholar

Yarkoni, T., and Westfall, J. (2017). Choosing prediction over explanation in psychology: lessons from machine learning. Perspect. Psychol. Sci. 12, 1100–1122. doi: 10.1177/1745691617693393,

PubMed Abstract | Crossref Full Text | Google Scholar

Zhang, S. (2025). The effect of physical exercise on Chinese college students’ mental sub-health: the mediating role of mental resilience and the moderating role of self-efficacy. Front. Psychol. 16:1572974. doi: 10.3389/fpsyg.2025.1572974,

PubMed Abstract | Crossref Full Text | Google Scholar

Živkov, I., Živkov, A., Živkov, M., and Škrlec, I. (2022). The prevalence of exercise dependence among students. J. Phys. Educ. Sport 22, 1708–1715. doi: 10.7752/jpes.2022.07214

Crossref Full Text | Google Scholar

Keywords: college student, ensemble learning (EN), exercise dependence behavior, machine learning, risk prediction

Citation: Deng Y, Lan W, Si M and Ren YL (2026) Predicting college students’ exercise dependence: a machine learning approach. Front. Psychol. 17:1743725. doi: 10.3389/fpsyg.2026.1743725

Received: 11 November 2025; Revised: 15 December 2025; Accepted: 09 January 2026;
Published: 29 January 2026.

Edited by:

David Manzano Sánchez, University of Almeria, Spain

Reviewed by:

Yuer Yang, The University of Hong Kong, Hong Kong SAR, China
Seungbak Lee, State University of New York at Fredonia, United States

Copyright © 2026 Deng, Lan, Si and Ren. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Yi Lin Ren, eWlsaW5yZW5Aam51LmVkdS5jbg==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.