ORIGINAL RESEARCH article

Front. Psychiatry, 05 February 2024

Sec. Computational Psychiatry

Volume 15 - 2024 | https://doi.org/10.3389/fpsyt.2024.1280935

Diagnosing and tracking depression based on eye movement in response to virtual reality

  • 1. School of Information and Communication Engineering, Hainan University, Haikou, China

  • 2. School of Information Engineering, Hainan Vocational University of Science and Technology, Haikou, China

  • 3. The First Affiliated Hospital of Hainan Medical University, Haikou, China

  • 4. Department of Psychology, University of Chinese Academy of Sciences, Beijing, China

  • 5. School of Electronic Science and Technology, Hainan University, Haikou, China

Article metrics

View details

9

Citations

6,2k

Views

2,1k

Downloads

Abstract

Introduction:

Depression is a prevalent mental illness that is primarily diagnosed using psychological and behavioral assessments. However, these assessments lack objective and quantitative indices, making rapid and objective detection challenging. In this study, we propose a novel method for depression detection based on eye movement data captured in response to virtual reality (VR).

Methods:

Eye movement data was collected and used to establish high-performance classification and prediction models. Four machine learning algorithms, namely eXtreme Gradient Boosting (XGBoost), multilayer perceptron (MLP), Support Vector Machine (SVM), and Random Forest, were employed. The models were evaluated using five-fold cross-validation, and performance metrics including accuracy, precision, recall, area under the curve (AUC), and F1-score were assessed. The predicted error for the Patient Health Questionnaire-9 (PHQ-9) score was also determined.

Results:

The XGBoost model achieved a mean accuracy of 76%, precision of 94%, recall of 73%, and AUC of 82%, with an F1-score of 78%. The MLP model achieved a classification accuracy of 86%, precision of 96%, recall of 91%, and AUC of 86%, with an F1-score of 92%. The predicted error for the PHQ-9 score ranged from -0.6 to 0.6.To investigate the role of computerized cognitive behavioral therapy (CCBT) in treating depression, participants were divided into intervention and control groups. The intervention group received CCBT, while the control group received no treatment. After five CCBT sessions, significant changes were observed in the eye movement indices of fixation and saccade, as well as in the PHQ-9 scores. These two indices played significant roles in the predictive model, indicating their potential as biomarkers for detecting depression symptoms.

Discussion:

The results suggest that eye movement indices obtained using a VR eye tracker can serve as useful biomarkers for detecting depression symptoms. Specifically, the fixation and saccade indices showed promise in predicting depression. Furthermore, CCBT demonstrated effectiveness in treating depression, as evidenced by the observed changes in eye movement indices and PHQ-9 scores. In conclusion, this study presents a novel approach for depression detection using eye movement data captured in VR. The findings highlight the potential of eye movement indices as biomarkers and underscore the effectiveness of CCBT in treating depression.

1 Introduction

The “World Mental Health Report 2022” highlights a concerning trend: the number of people with mental illness has surpassed one billion worldwide. The coronavirus disease 2019 pandemic has exacerbated this issue, with a 25% increase in the number of people experiencing anxiety and depression in the first year of the pandemic Organization et al. (1). Depression alone is responsible for over 800,000 deaths annually Cai et al. (2), and its prevalence has increased by 18% over the past decade Mahato and Paul (3).

Currently, the clinical diagnosis of depression primarily relies on psychological behavioral scales, such as the Hamilton Depression Scale Hamilton (4), Beck Depression Inventory Beck et al. (5), and Self-Rating Depression Scale Zung (6). However, these methods are limited by subjectivity and lack objective and quantitative evaluation criteria, increasing the risk of various subjective biases Sung et al. (7). Moreover, completing these questionnaires is time consuming, with each form requiring at least 30 minutes. Therefore, finding an objective and efficient diagnostic method is an urgent problem that must be addressed Strawbridge et al. (8); de Aguiar Neto and Rosa (9); Tavakolizadeh et al. (10); Chojnowska et al. (11).

The phrase “the eyes are the windows to the soul” holds true in psychology; changes in pupil size are related to psychological states, reflecting an individual’s higher cognitive processes Hess and Polt (12). Pupil size measurements have been used to evaluate stress Yamanaka and Kawakami (13). Eye movement indices such as fixation duration, saccade count, and pupil size are direct reflections of brain information processing and can quantitatively characterize emotional perception. Previous studies have established a connection between eye movement indices and depression. Free-viewing, fixation stability, and smooth pursuit tests were conducted, resulting in 35 eye movement measurements. The results revealed that in the free viewing test, individuals with Major Depressive Disorder (MDD) exhibited significantly shorter scanpath length. In the smooth pursuit test, the duration of saccades was significantly shorter, and the peak saccade velocity was significantly lower in individuals with MDD Takahashi et al. (14).The prosaccade task and the antisaccade task were performed using iViewX RED 500 eye-tracking instruments. The results showed significant differences between the depression group and the control group in the antisaccade task, specifically in the correct rate (t = 3.219, p = 0.002) and mean velocity (F = 3.253, p< 0.05) Gao et al. (15). Eyelink 1000 recorder was used to collect eye movement data. Features related to fixation, saccades, pupil size, and depression preference were extracted from the data. Five classifiers, namely k-Nearest-Neighbor (kNN), Naïve Bayes (NB), Logistic Regression (LR), Support Vector Machine (SVM), and Random Forest (RF), were employed to classify a group of students into depressed and normal categories. By utilizing eye movement features during free viewing tasks, an accuracy of 80.1% was achieved using Random Forest to differentiate between depressed and non-depressed subjects Li et al. (16).Eye movement metrics can also be combined with other indicators. Facial expressions and eye movement data were integrated, and Classification Accuracy using cross-validation (within-study replication) achieved a level of 79% (sensitivity 76%, specificity 82%).Stolicyn et al. (17). Eye movement data and resting-state EEG signals were combined, and an ensemble method was employed. The study utilized free viewing eye tracking and resting-state EEG data, and the results demonstrated that the combined approach, known as CBEM, achieved accuracies of 82.5%.Zhu et al. (18).Eye movement indices can effectively be used to distinguish depression from other mental illnesses and serve as biomarkers to aid in the diagnosis and evaluation of depression Bae et al. (19); Pan et al. (20); Carvalho et al. (21); Wen et al. (22); Wang et al. (23); Carvalho et al. (24). These findings are supported by several studies.

An eye tracker is an essential instrument for measuring eye movement indices and trajectories. We used a virtual reality (VR) eye tracking system comprised of an aSee VR (Beijing 7invensun Technology Co., Ltd., Beijing, China) and HTC Vive Pro (HTC Corporation, Taoyuan City, Taiwan), as shown in Figure 1. This eye tracking system differs from commonly used series such as the EyeLink or Tobii. This VR eye tracking system provides a more immersive and realistic experience because of its ability to create rich virtual environments. Moreover, psychotherapy and interventions based on VR have shown promise for the treatment of patients with anxiety and depression Li et al. (25); Jiede et al. (26); Suwanjatuporn and Chintakovid (27); Torous et al. (28).

Figure 1

Machine learning algorithms can analyze data, thereby enabling the prediction and classification of practical problems Islam et al. (29); Chiong et al. (30); AlSagri and Ykhlef (31); Narayanrao and Kumari (32). Recently, the use of physiological indices such as speech Herniman et al. (33); Cohen et al. (34); Rapcan et al. (35) and electroencephalogram signals de Aguiar Neto and Rosa (9); Keren et al. (36); Acharya et al. (37) has facilitated the collection and analysis of data for the purpose of diagnosing depression and distinguishing individuals with depression from healthy individuals. These are fast, effective, and efficient auxiliary methods for detecting depression.

Computerized cognitive behavioral therapy (CCBT) is an online tool that enables patients to receive professional cognitive behavior therapy and has gained significant popularity in recent years Wickersham et al. (38); Wright et al. (39). CCBT is not bound by time or location, effectively solving the issue of insufficient medical resources Wright et al. (40); Pfeiffer et al. (41).

Based on the reviewed literature, we aimed to utilize machine learning methods to detect depression by analyzing eye movement data obtained from VR scenes. In addition, this study investigated the effectiveness of CCBT in treating depression.

2 Materials and methods

2.1 Participant information

The Participants were recruited from Hainan Medical University and assessed using the PHQ-9 scale Kroenke et al. (42) based on the inclusion and exclusion criteria. The PHQ-9 scale, developed by Kroenke et al. in 2001, is based on diagnostic criteria for depressive episodes. A total of 167 participants with PHQ-9 scores of 5 or higher, indicating significant depressive symptoms, were selected for the study. Subsequently, 60 participants were selected based on the inclusion criteria and divided into a depression intervention group and a depression control group, with 30 participants in each group. In addition, 30 students were selected as the healthy control group from 124 participants who scored 4 points or lower on the PHQ-9 scale, indicating no or minimal depressive symptoms, according to the inclusion criteria. The inclusion and exclusion criteria are presented in Table 1.

Table 1

Participant with depressive emotionhealthy controls
Inclusion criteria1 Between the age of 16-25 years old, male or female
2 All participants are university students
3 The PHQ-9 score of the subject was greater than or equal to 5.
4 No psychotropic drug treatment having been performed in the last two weeks
1 Between the age of 16-25 years old, male or female
2 All participant are university students
3 No history of mental illness
4 PHQ-9 score of subject was lower than 4
exclusion criteria1 History of significant physical illness
2 have suicidal tendency
3 have bipolar affective disorder and psychotic symptoms
1 Individual or family with mental illness
2 with other serve physical illness
3 Were dependent on alcohol or psychotropic drugs in the past year

The inclusion and exclusion criteria of eye tracker experiments for depression detection.

In the study, the participants were categorized based on their PHQ-9 scores. Among the participants, 13 individuals were classified as having mild depression (scores of 5-9), 33 individuals were classified as having moderate depression (scores of 10-14), 8 individuals were classified as having moderately severe depression (scores of 15-19), and 6 individuals were classified as having severe depression (scores of 20-27).During the experiment, there was participant attrition, and a total of 69 individuals completed the experiment. Among them, 24 individuals were in the healthy group, and 45 individuals were in the depressive emotion group. This study was reviewed and approved by the Ethics Committee of The First Affiliated Hospital of Hainan Medical University. All participants provided written informed consent.

2.2 Experimental design

The participants were divided into three groups: depression intervention, depression control, and healthy control, with 30 individuals in each group. Next, the participants completed the PHQ-9 and their eye movement data was collected using the HTC Vive Eye Pro eye-tracking device. The data was then processed, classified, and used to predict PHQ-9 scores using XGBoost Chen and Guestrin (43) and MLP Fatima et al. (44) models. The performance of the two models was compared. After completing five sessions of group-based CCBT, we readministered the PHQ-9 and repeated the eye movement experiment in the intervention group. The results were compared with those obtained before the CCBT intervention to assess the changes in PHQ-9 scores and eye movement data. The experimental flowchart is shown in Figure 2.

Figure 2

2.3 Feature extraction

The definition of eye movement indices can be found in reference Holmqvist et al. (45). For this experiment, eight eye-movement indices were extracted as input parameters for model training using a VR eye tracker. These indices included fixation count, fixation duration, mean fixation duration, saccade count, saccade amplitude, mean saccade amplitude, and the average pupil sizes of both eyes.

2.4 Model introduction

2.4.1 XGBoost model

In this study, we analyzed the collected data separately using the XGBoost and MLP models. The XGBoost model was used for classification and prediction. XGBoost is an upgraded version of the gradient-boosted decision tree algorithm Wang et al. (46) that uses a gradient-boosting algorithm, similar to the boosting method used in ensemble learning. Ensemble learning algorithms such as XGBoost construct multiple weak learners and aggregate the modeling results of all weak learners to obtain better classification or regression performance than that of a single model. XGBoost is known for its efficiency, scalability, and accuracy, making it a popular choice for machine learning applications, including those related to mental health.

The modeling process involves building a tree to form an evaluator that is gradually iterated. During each iteration, the evaluator focuses on easily misclassified samples and gradually integrates strong evaluators.

This iterative process enables the XGBoost model to improve its performance over time, leading to increasingly accurate predictions. Figure 3 illustrates the modeling process of the XGBoost algorithm.

Figure 3

Multiple decision trees were used during the process of building the XGBoost model, and the regression result of the ensemble model is the sum of the predicted scores for all trees. If K decision trees exist in the ensemble model, the predicted result of the entire model for sample I can be expressed as follows (Equation 1):

The objective function of the XGBoost model can be expressed as a combination of the loss function and model complexity as follows (Equation 2):

2.4.2 Multilayer perceptron model

MLP is a fundamental model that yields high accuracy when working with small datasets. MLP is an artificial neural network consisting of multiple layers of interconnected nodes, with each node performing a simple mathematical operation. The MLP model uses the backpropagation of errors and continuously updates the weights using the gradient descent algorithm to find the minimum error value and calculate the weights of each node, as shown in Figure 4.

Figure 4

In this study, we employed a 3-layer neural network with the following parameters: (8,512), (512,1024), and (1024,1). The activation functions used were ReLU and Sigmoid. We utilized the Adam optimization method and the BCELoss as the loss function. The programming environment for this study was PyTorch version 1.13.1. Since the dataset size was small, the hardware requirements were not high, and training could be performed using a CPU.

3 Results

3.1 Classification and predication results

Five-fold cross-validation was performed during the classification training process, as illustrated in Figure 5. This process involved dividing the data into five parts: one part for validation data and the remaining four parts for training data. The final result was obtained by averaging the results of the five-fold cross-validation runs.

Figure 5

After training, the classification performance of each of the four classifiers was evaluated, and the results are presented in Table 2. The confusion matrix and receiver operating characteristic curve are shown in Figures 6, 7, respectively. The average areas under the receiver operating characteristic curve after fivefold cross-validation were 0.86 and 0.82 for the MLP and XGBoost models, respectively. However, in terms of predicting PHQ-9 scores based on eye movement data, the XGBoost model exhibited superior performance compared with that of the MLP model. The XGBoost and MLP models predicted PHQ-9 scores with mean absolute error values ranging from -0.6 to 0.6 and from -1 to 1, respectively, indicating good predictive performance.

Table 2

Fold1Fold2Fold3Fold4Fold5Mean-Value
(A)XGBoost
Accuaray(%)717965798576
Precision(%)851001008710094
Recall(%)676773787873
F1-score(%)758065828878
AUC(%)768675829282
(B)MLP
Accuaray(%)859285789286
Precision(%)8890888710096
Recall(%)8810088778891
F1-score(%)889488829492
AUC(%)849084789486
(C)SVM
Accuaray(%)719271786174
Precision(%)779077877080
Recall(%)7710077777782
F1-score(%)779477827381
AUC(%)689068785172
(D)Random Forest
Accuracy(%)786464854667
Precision(%)877575886077
Recall(%)776666886673
F1-score(%)827070886374
AUC(%)786363843365

The performance of the four classifier models.

Figure 6

Figure 7

3.2 Analysis of model feature importance

The importance of each model feature was evaluated using the model’s “feature_importances_function.” The saccade and fixation indices had relatively large proportions in the model. Saccades and fixations accounted for 39% and 34% of the model’s feature importance, respectively. Together, these two indices accounted for 73% of the model’s feature importance. The feature importance of the model is illustrated in Figure 8, which provides a visual representation of the relative importance of each feature in the model.

Figure 8

3.3 Comparison before and after intervention

3.3.1 Changes in indices

After five group-based CCBT sessions, significant changes were observed in the three eye movement indices. Fixation duration (p=0.035), saccade amplitude (p=0.0193), and mean saccade amplitude (p=0.0132) exhibited significant changes(as illustrated in Figure 9), indicating that group-based CCBT affected the eye movement patterns in individuals with depression.

Figure 9

3.3.2 Changes in PHQ-9 scores before and after intervention

The data obtained after five group-based CCBT sessions were analyzed using Prism 8 (GraphPad Software, San Diego, CA, USA). The data were found to be normally distributed, and paired t-tests were subsequently performed to assess the significance of the changes in PHQ-9 scores before and after the intervention. The results showed a significant decrease in the average PHQ-9 score from 11.57 to 7.8 after the intervention (p=0.0158) (as shown in Figure 10). Additionally, the depression scores for the control group remained consistent, with measurements of 12 and 11.6, indicating no change over time. These findings suggest that group-based CCBT effectively reduces depressive symptoms in individuals with depression.

Figure 10

4 Discussion

We developed a model for classifying and predicting depression that differs from those reported in prior studies, which focused solely on classification. Unlike previous studies, our model includes both classification and prediction, enabling a more comprehensive analysis of depressive symptoms.

4.1 Advantages of using virtual reality eye-tracking devices in research

The use of eye tracking technology in response to VR has been investigated in several studies. For example, Imaoka et al. (47) investigated the characteristics and data accuracy of eye-tracking technology using a VIVE Pro Eye device and found it to be an effective tool for evaluating eye tracking. Similarly, Alghowinem et al. (48) found that participants who used VR eye-tracking devices were more easily immersed in a scene and less likely to be disturbed by external environmental factors compared with viewing pictures.

Our findings are consistent with those of the aforementioned studies, as we also observed that using a VR eye-tracking device to watch videos provides a more realistic experience and effectively reduces attentional bias in the participants. The use of VR technology can create a more immersive and engaging experience for participants, leading to more reliable data.

4.2 Eye movement indices can be used to classify depression

Our study provides further evidence of the effectiveness of eye movement indices in the classification of depression, with results consistent with those of previous studies in this field. For example, Li et al. (49) analyzed eye movement data while viewing emotionally charged images from 48 individuals with depressive symptoms and 48 normal individuals. The learning machine classification method achieved the highest classification accuracy at 84.21%. In our study, the MLP model achieved a classification accuracy of 86%. These findings underscore the potential of eye movement indices as effective tools for depression classification and offer opportunities for the development of noninvasive and objective methods for diagnosing and treating depression.

Alghowinem et al. (50) used an interview method to classify eye movement data from 30 patients with depression and 30 healthy individuals and achieved an accuracy rate of 75% using a support vector machine classifier. Similarly, Zhang et al. (51) reported the highest classification accuracy at 80.1% using a random forest classifier based on data obtained when participants freely browsed pictures of emotional faces, using pupil size, gaze position, and gaze duration as classification features.

Compared with the aforementioned studies, the XGBoost and MLP models used in our study both achieved a higher classification performance. Furthermore, our study is unique in that we were able to predict PHQ-9 scores based on eye movement data, with the maximum mean absolute error ranging from -0.6 to 0.6, which is a significant improvement from that reported in previous studies.

This article utilizes four machine learning methods to establish models based on eye-tracking data for classification and prediction. The results demonstrate high levels of accuracy in both classification and prediction tasks, indicating that it can serve as an effective means for depression detection. These four machine learning methods include ensemble learning models and support vector machine models. The experiments show that both categories of models achieve significant results in utilizing eye-tracking data for depression classification and prediction, highlighting the effectiveness of these models in these tasks. Compared to the commonly used questionnaires at present, this method is quick, convenient, and objective.

4.3 Eye movement indices serve as biomarkers for detecting depression

Prior studies have reported that during a fixation task, patients with depression have higher fixation counts and shorter fixation durations than those of healthy controls, which is consistent with our findings. Yu et al. (52) reported similar results, highlighting the potential of eye movement data as biomarkers for the diagnosis of depression.

In our study, we observed significant differences in fixation duration (p=0.035), saccade amplitude (p=0.0193), and mean saccade amplitude (p=0.0132) after CCBT. Analysis of the predictive model parameters revealed that fixation and saccade were important influencing factors, with the importance of these two indices accounting for more than 70% of the model. These findings highlight the potential of fixation duration and saccade as biomarkers for diagnosing depression and underscore their significance in clinical diagnosis.

5 Conclusion

The XGBoost and MLP models were effective in classifying and predicting depression and achieved high levels of accuracy. These results suggest that machine learning techniques hold great promise for improving the diagnosis and treatment of depression, and have the potential to become valuable tools for clinicians in the future.

The analysis revealed that fixation durations and saccade counts were the two primary indices of model importance and showed significant changes following CCBT. These results suggest that fixation durations and saccade counts could serve as important biomarkers for the diagnosis of depression.

After undergoing CCBT, significant changes were observed in both eye movement indices as well as PHQ-9 scores among the participants in the control group with depressive symptoms. These results suggest that CCBT is an effective approach for managing depression.

6 Limitations and directions for future studies

We intend to increase the sample sizes in future studies to obtain more stable data and improve the reliability of our findings. In addition, we aim to investigate other potential biomarkers and explore the underlying mechanisms of CCBT for depression.

The number of CCBT sessions could also be increased in future studies to facilitate the collection of intermediate treatment data. Furthermore, qualitative studies could be conducted to explore the relationships between treatment duration, eye movement data, and PHQ-9 scores.

In this study, we used neutral scenes in the eye movement analysis. Future studies could also explore the use of positive and negative scenes and compare the differences in depression diagnoses.

Statements

Data availability statement

The original contributions presented in the study are included in the article/Supplementary Material. Further inquiries can be directed to the corresponding author.

Ethics statement

The studies involving humans were approved by The First Affiliated Hospital of Hainan Medical University, Haikou, China. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

ZZ: Data curation, Formal Analysis, Methodology, Writing – original draft. LL: Conceptualization, Formal Analysis, Funding acquisition, Investigation, Writing – original draft. XL: Conceptualization, Data curation, Investigation, Writing – original draft. JC: Software, Validation, Visualization, Writing – original draft. ML: Formal Analysis, Visualization, Writing – original draft. GW: Supervision, Validation, Writing – review & editing. CX: Supervision, Validation, Writing – review & editing.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This study was supported by the Hainan Province Natural Science Foundation of China (project number 821RC700), which focuses on the diagnosis of depression.

Acknowledgments

The authors express their gratitude to the First Affiliated Hospital of Hainan Medical University for their assistance with this project and to all the participants for their participation.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyt.2024.1280935/full#supplementary-material

References

  • 1

    Organization, W. HAbellIAbdiKBabbNDababnahSWagnerJWalkerHet al. World mental health report: transforming mental health for all. (2022).

  • 2

    CaiHHanJChenYShaXWangZHuBet al. A pervasive approach to eeg-based depression detection. Complexity (2018) 2018:113. doi: 10.1155/2018/5238028

  • 3

    MahatoSPaulS. (2019). Electroencephalogram (eeg) signal analysis for diagnosis of major depressive disorder (mdd): a review, in: Nanoelectronics, Circuits and Communication Systems: Proceeding of NCCS 2017, . pp. 323–35.

  • 4

    HamiltonM. A rating scale for depression. J Neurol Neurosurg Psychiatry (1960) 23:56. doi: 10.1136/jnnp.23.1.56

  • 5

    BeckATSteerRABallRRanieriWF. Comparison of beck depression inventories-ia and-ii in psychiatric outpatients. J Pers Assess (1996) 67:588–97. doi: 10.1207/s15327752jpa6703_13

  • 6

    ZungWW. A self-rating depression scale. Arch Gen Psychiatry (1965) 12:6370. doi: 10.1001/archpsyc.1965.01720310065008

  • 7

    SungMMarciCPentlandA. Objective physiological and behavioral measures for identifying and tracking depression state in clinically depressed patients. Cambridge, MA: Massachusetts Institute of Technology Media Laboratory (2005).

  • 8

    StrawbridgeRYoungAHCleareAJ. Biomarkers for depression: recent insights, current challenges and future prospects. Neuropsychiatr Dis Treat (2017) 13:1245–62. doi: 10.2147/NDT.S114542

  • 9

    de Aguiar NetoFSRosaJLG. Depression biomarkers using non-invasive eeg: A review. Neurosci Biobehav Rev (2019) 105:8393. doi: 10.1016/j.neubiorev.2019.07.021

  • 10

    TavakolizadehJRoshanaeiKSalmaninejadAYariRNahandJSSarkariziHKet al. Micrornas and exosomes in depression: potential diagnostic biomarkers. J Cell Biochem (2018) 119:3783–97. doi: 10.1002/jcb.26599

  • 11

    ChojnowskaSPtaszyńska-SarosiekIKepkaAKnaśMWaszkiewiczN. Salivary biomarkers of stress, anxiety and depression. J Clin Med (2021) 10:517. doi: 10.3390/jcm10030517

  • 12

    HessEPoltJ. Pupil size in relation to mental activity during simple problem-solving. science (1964) 143(whole no. 3611):1190–2. doi: 10.1126/science.143.3611.1190

  • 13

    YamanakaKKawakamiM. Convenient evaluation of mental stress with pupil diameter. Int J Occup Saf Ergonomics (2009) 15:447–50. doi: 10.1080/10803548.2009.11076824

  • 14

    TakahashiJHiranoYMiuraKMoritaKFujimotoMYamamoriHet al. Eye movement abnormalities in major depressive disorder. Front Psychiatry (2021) 12:673443. doi: 10.3389/fpsyt.2021.673443

  • 15

    GaoMXinRWangQGaoDWangJYuY. Abnormal eye movement features in patients with depression: Preliminary findings based on eye tracking technology. Gen Hosp Psychiatry (2023) 84:2530. doi: 10.1016/j.genhosppsych.2023.04.010

  • 16

    LiXCaoTSunSHuBRatcliffeM. (2016). Classification study on eye movement data: Towards a new approach in depression detection. In: 2016 IEEE Congress on Evolutionary Computation (CEC), Vancouver, BC, Canada: IEEE. pp. 1227–32.

  • 17

    StolicynASteeleJDSerièsP. Prediction of depression symptoms in individual subjects with face and eye movement tracking. psychol Med (2022) 52:1784–92. doi: 10.1017/S0033291720003608

  • 18

    ZhuJWangZGongTZengSLiXHuBet al. An improved classification model for depression detection using eeg and eye tracking data. IEEE Trans Nanobiosci (2020) 19:527–37. doi: 10.1109/TNB.2020.2990690

  • 19

    BaeHKimDParkYC. Eye movement desensitization and reprocessing for adolescent depression. Psychiatry Invest (2008) 5:60. doi: 10.4306/pi.2008.5.1.60

  • 20

    PanZMaHZhangLWangY. (2019). Depression detection based on reaction time and eye movement, in: 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan: IEEE. pp. 2184–8.

  • 21

    CarvalhoNLaurentENoiretNChopardGHaffenEBennabiDet al. Eye movement in unipolar and bipolar depression: A systematic review of the literature. Front Psychol (2015) 6:1809. doi: 10.3389/fpsyg.2015.01809

  • 22

    WenMDongZZhangLLiBZhangYLiK. Depression and cognitive impairment: Current understanding of its neurobiology and diagnosis. Neuropsychiatr Dis Treat (2022) 18:2783–94. doi: 10.2147/NDT.S383093

  • 23

    WangYLyuH-LTianX-HLangBWangX-YSt ClairDet al. The similar eye movement dysfunction between major depressive disorder, bipolar depression and bipolar mania. World J Biol Psychiatry (2022) 23:689702. doi: 10.1080/15622975.2022.2025616

  • 24

    CarvalhoNNoiretNVandelPMonninJChopardGLaurentE. Saccadic eye movements in depressed elderly patients. PloS One (2014) 9:e105355. doi: 10.1371/journal.pone.0105355

  • 25

    LiHDongWWangZChenNWuJWangGet al. Effect of a virtual reality-based restorative environment on the emotional and cognitive recovery of individuals with mild-to-moderate anxiety and depression. Int J Environ Res Public Health (2021) 18:9053. doi: 10.3390/ijerph18179053

  • 26

    JiedeWLiFXiaHJiashunX. (2020). Research status of depression related to virtual reality technology in China. In: 2020 International Conference on Intelligent Computing and Human-Computer Interaction (ICHCI), Sanya, China: IEEE. pp. 71–4.

  • 27

    SuwanjatupornAChintakovidT. (2019). Using a virtual reality system to improve quality of life of the elderly people with depression. In: 2019 IEEE International Conference on Consumer Electronics-Asia (ICCE-Asia), Bangkok, Thailand: IEEE. pp. 153–6.

  • 28

    TorousJBucciSBellIHKessingLVFaurholt-JepsenMWhelanPet al. The growing field of digital psychiatry: current evidence and the future of apps, social media, chatbots, and virtual reality. World Psychiatry (2021) 20:318–35. doi: 10.1002/wps.20883

  • 29

    IslamMRKabirMAAhmedAKamalARMWangHUlhaqA. Depression detection from social network data using machine learning techniques. Health Inf Sci Syst (2018) 6:112. doi: 10.1007/s13755-018-0046-0

  • 30

    ChiongRBudhiGSDhakalSChiongF. A textual-based featuring approach for depression detection using machine learning classifiers and social media texts. Comput Biol Med (2021) 135:104499. doi: 10.1016/j.compbiomed.2021.104499

  • 31

    AlSagriHSYkhlefM. Machine learning-based approach for depression detection in twitter using content and activity features. IEICE Trans Inf Syst (2020) 103:1825–32. doi: 10.1587/transinf.2020EDP7023

  • 32

    NarayanraoPVKumariPLS. (2020). Analysis of machine learning algorithms for predicting depression, in: 2020 International Conference on Computer Science, Engineering and Applications (ICCSEA), Gunupur, India: IEEE. pp. 14.

  • 33

    HernimanSEAllottKAKillackeyEHesterRCottonSM. The effect of comorbid depression on facial and prosody emotion recognition in first-episode schizophrenia spectrum. J Affect Disord (2017) 208:223–9. doi: 10.1016/j.jad.2016.08.068

  • 34

    CohenASMcGovernJEDinzeoTJCovingtonMA. Speech deficits in serious mental illness: a cognitive resource issue? Schizophr Res (2014) 160:173–9. doi: 10.1016/j.schres.2014.10.032

  • 35

    RapcanVD’ArcySYeapSAfzalNThakoreJReillyRB. Acoustic and temporal analysis of speech: A potential biomarker for schizophrenia. Med Eng Phys (2010) 32:1074–9. doi: 10.1016/j.medengphy.2010.07.013

  • 36

    KerenHO’CallaghanGVidal-RibasPBuzzellGABrotmanMALeibenluftEet al. Reward processing in depression: a conceptual and meta-analytic review across fmri and eeg studies. Am J Psychiatry (2018) 175:1111–20. doi: 10.1176/appi.ajp.2018.17101124

  • 37

    AcharyaUROhSLHagiwaraYTanJHAdeliHSubhaDP. Automated eeg-based screening of depression using deep convolutional neural network. Comput Methods Programs Biomed (2018) 161:103–13. doi: 10.1016/j.cmpb.2018.04.012

  • 38

    WickershamABarackTCrossLDownsJ. Computerized cognitive behavioral therapy for treatment of depression and anxiety in adolescents: systematic review and meta-analysis. J Med Internet Res (2022) 24:e29842. doi: 10.2196/29842

  • 39

    WrightJHOwenJJRichardsDEellsTDRichardsonTBrownGKet al. Computer-assisted cognitive-behavior therapy for depression: a systematic review and meta-analysis. J Clin Psychiatry (2019) 80:3573. doi: 10.4088/JCP.18r12188

  • 40

    WrightJHMishkindMEellsTDChanSR. Computer-assisted cognitive-behavior therapy and mobile apps for depression and anxiety. Curr Psychiatry Rep (2019) 21:19. doi: 10.1007/s11920-019-1031-2

  • 41

    PfeifferPNPopeBHouckMBenn-BurtonWZivinKGanoczyDet al. Effectiveness of peer-supported computer-based cbt for depression among veterans in primary care. Psychiatr Serv (2020) 71:256–62. doi: 10.1176/appi.ps.201900283

  • 42

    KroenkeKSpitzerRLWilliamsJB. The phq-9: validity of a brief depression severity measure. J Gen Internal Med (2001) 16:606–13. doi: 10.1046/j.1525-1497.2001.016009606.x

  • 43

    ChenTGuestrinC. (2016). Xgboost: A scalable tree boosting system, in: Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining, . pp. 785–94.

  • 44

    FatimaIAbbasiBUDKhanSAl-SaeedMAhmadHFMumtazR. Prediction of postpartum depression using machine learning techniques from social media text. Expert Syst (2019) 36:e12409. doi: 10.1111/exsy.12409

  • 45

    HolmqvistKNyströmMAnderssonRDewhurstRJarodzkaHVan de WeijerJ. Eye tracking: A comprehensive guide to methods and measures. Oxford: OUP (2011).

  • 46

    WangBHanXZhaoZWangNZhaoPLiMet al. Eeg-driven prediction model of oxcarbazepine treatment outcomes in patients with newly-diagnosed focal epilepsy. Front Med (2022) 8:781937. doi: 10.3389/fmed.2021.781937

  • 47

    ImaokaYFluryADe BruinED. Assessing saccadic eye movements with head-mounted display virtual reality technology. Front Psychiatry (2020) 11:572938. doi: 10.3389/fpsyt.2020.572938

  • 48

    AlghowinemSAlShehriMGoeckeRWagnerM. (2014). Exploring eye activity as an indication of emotional states using an eye-tracking sensor, in: Intelligent Systems for Science and Information: Extended and Selected Results from the Science and Information Conference 2013, London, UK: Springer. pp. 261–76.

  • 49

    LiMCaoLZhaiQLiPLiuSLiRet al. Method of depression classification based on behavioral and physiological signals of eye movement. Complexity (2020) 2020:19. doi: 10.1155/2020/8882813

  • 50

    AlghowinemSGoeckeRWagnerMParkerGBreakspearM. (2013). Eye movement analysis for depression detection, in: 2013 IEEE International Conference on Image Processing, Melbourne, VIC, Australia: IEEE. pp. 4220–4.

  • 51

    ZhangBZhouWCaiHSuYWangJZhangZet al. Ubiquitous depression detection of sleep physiological data by using combination learning and functional networks. IEEE Access (2020) 8:94220–35. doi: 10.1109/ACCESS.2020.2994985

  • 52

    YuLYangyangXMengqingXZhangTJunjieWXuLet al. Eye movement indices in the study of depressive disorder. Shanghai Arch Psychiatry (2016) 28:326. doi: 10.11919/j.issn.1002-0829.216078

Summary

Keywords

depression, diagnose, XGBoost, MLP, CCBT

Citation

Zheng Z, Liang L, Luo X, Chen J, Lin M, Wang G and Xue C (2024) Diagnosing and tracking depression based on eye movement in response to virtual reality. Front. Psychiatry 15:1280935. doi: 10.3389/fpsyt.2024.1280935

Received

21 August 2023

Accepted

16 January 2024

Published

05 February 2024

Volume

15 - 2024

Edited by

Ming Huang, University of Texas Health Science Center at Houston, United States

Reviewed by

Hao Liu, Montclair State University, United States

Nan Huo, Mayo Clinic, United States

Mingquan Lin, Cornell University, United States

Updates

Copyright

*Correspondence: Guanjun Wang, ; Chenyang Xue,

†These authors have contributed equally to this work and share first authorship

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics