- 1Beijing Key Laboratory of Applied Experimental Psychology, National Demonstration Center for Experimental Psychology Education (Beijing Normal University), Faculty of Psychology, Beijing Normal University, Beijing, China
- 2School of Information Management, Wuhan University, Wuhan, China
- 3Institute of Biomedical Engineering, University of Oxford, Oxford, United Kingdom
- 4School of Economics and Management, Beijing Jiaotong University, Beijing, China
Introduction: The exploration of life’s meaning has been a key topic across disciplines, and artificial intelligence is now beginning to investigate it.
Methods: This study leveraged social media to assess meaning in life (MIL) and its associated factors at individual and group levels. We compiled a diverse dataset consisting of microblog posts (N = 7,588,597) and responses from user surveys (N = 448), annotated using a combination of self-assessment, expert opinions, and ChatGPT-generated insights. Our methodology examined MIL in three ways: (1) developing deep learning models to assess MIL components, (2) applying semantic dependency graph algorithms to identify MIL associated factors, and (3) constructing eight subnetworks to analyze factors, their interrelations, and MIL differences.
Results: We validated these methods and bridged two foundational MIL theories, highlighting their interconnections.
Discussion: By identifying psychological risk factors, our work may provide clues to mental health issues and inform possible intervention.
1 Introduction
Meaning in life (MIL) has long been recognized as the sense that one’s life has purpose, coherence, and significance (1). Building on this view, later research typically distinguishes the Presence of Meaning in Life (POM) and the Search for Meaning in Life (SFM) (2). From a Self-Determination Theory perspective, MIL has been linked to autonomy, competence, and relatedness needs (3). Higher MIL was related to greater life satisfaction (4, 5) and better health indicators (2). It was found to protect against psychological distress (6), depression (7–9), anxiety (67), and loneliness (10). It also related to mortality-related concerns, including death anxiety and suicide risk (10–12). Beyond mental health, MIL was linked to better sleep quality (13), lower risk of eating disorders (14), and enhanced coping among individuals with chronic conditions such as cancer and HIV (15, 16).
Building on the importance of MIL, our research question concerns how MIL can be assessed. Existing assessments have relied on validated questionnaires developed from diverse theoretical perspectives. These included the Purpose in Life questionnaire (17), the Meaning in Life Questionnaire with Presence and Search subscales (2), and the Meaningful Life Measure (18). Subsequently, the Multidimensional Existential Meaning Scale conceptualized comprehension, purpose, and mattering (19). Most recently, the Three-Dimensional Model of Meaning conceptualized coherence, purpose, and significance (20). Although these tools have advanced our understanding of MIL’s structure and correlates, two research gaps remain. First, because MIL is abstract and multifaceted, there is no universally accepted structured definition, and the relations between competing frameworks such as two-dimensional and three-dimensional models require further clarification. Second, while questionnaires capture subjective perceptions well, they are limited in scalability, timeliness, and adaptability for large-scale assessment in naturalistic settings.
This study introduces a novel approach to automatically assess an individual’s MIL from social media expressions and examines differences in linguistic patterns across POM and SFM levels. Three lines of evidence support the feasibility of this approach. First, MIL could emerge from everyday experiences such as establishing daily routines (21, 22), and online interactions often mirrored real-world social dynamics (23, 24). Second, social media content provided a window into existential reflection (25, 26). Moreover, social media engagement was closely associated with processes of meaning making (23). More active engagement such as posting photos was related to a stronger sense of purpose in life (25, 27), whereas passive browsing without interaction mediated the relation between meaning and self-esteem in stressful contexts (28). Third, on the technical side, prior studies inferred implicit psychological variables from user-generated content (29), including anxiety (30), insomnia (31, 32), stress (33–35) and stressors (36), and suicide risk (37).
Building on these foundations, we moved from usage patterns to language-based, large-scale assessment of MIL in social media text. We then examined how levels of POM and SFM relate to associated factors at the group level. This study aims to address two central research questions:
RQ1: The relationship between POM and SFM is complex and sometimes inconsistent. How can we model these two components from real-world social media data?
RQ2: The associated factors that shape MIL are varied and often interrelated (e.g., achievements, security, spirituality, health, family life). How can we accurately extract these factors from large-scale textual data and characterize their interconnections?
Specifically, this study assessed MIL within the sleep context for three reasons. First, on theoretical grounds, prior work showed that rumination and self-evaluation occurred more often at night (38). Moreover, rumination was closely associated with MIL (39). Ge (31) further showed that POM significantly predicted sleep quality via mediators such as depression. These studies supported sleep as a theoretically dense window on MIL. Second, from a methodological perspective, constraining the analysis of textual expressions of MIL to the sleep context reduced noninformative noise in open social media data. It increased the density and retrievability of MIL-relevant signals, which made the linguistic features easier to capture and model. Third, from a reproducibility perspective, using sleep as a contextual starting point facilitated replication across platforms. The methods could be reapplied with matched time windows in follow-up studies.
Focusing on the sleep context, we collected over 7,500,000 microblogs, of which 189,213 contained MIL-related keywords. We conducted manual annotation and data augmentation on sampled posts, constructing a labeled dataset comprising 3,000 MIL-relevance labels, 1,600 POM (High vs. Low) labels, and 1,600 SFM (High vs. Low) labels. The three-part framework is shown in Figure 1. In Study 1, we developed three binary deep learning models to assess MIL relatedness, POM level, and SFM level. We then applied these models to segment the large-scale microblog dataset into eight subgroups reflecting different MIL states. Study 2 employed semantic dependency graphs to extract associated factors and their semantic relations. In Study 3, we constructed eight subnetworks to support downstream analyses of associated-factor patterns, including their effects on MIL, their interconnections, and differences across subnetworks.
This study makes contributions both methodologically and theoretically. Methodologically, we introduce a framework for automated, timely assessment of MIL and its associated factors. Beyond scalable analysis of social media content, the framework has practical implications. It supports early detection of mental health risks. For example, it can monitor loss of meaning as a precursor to depression or insomnia. It also helps tailor interventions in educational or occupational settings and informs the design of digital platforms that enhance social support and well-being. Theoretically, we classify the associated factors into five areas: factor frequency, influential factors, factor relationships, significant factor differences, and clustering trends, and explore their implications for the two-component model of MIL, offering new insights into the complexities of meaning.
2 Methods
2.1 Dataset
To address the absence of publicly available datasets for MIL, we started by constructing a reliable corpus comprising two parts: a large-scale Weibo dataset and an empirical dataset collected through participant recruitment (see Figure 2).
2.1.1 Microblogs
We executed a Python program on Sina Weibo, a leading social media platform in China, between December 2021 and January 2022. It collected 7,588,597 sleep-related microblogs posted from January 15 to July 15, 2021 (see Figure 2). All collected posts were originally written in Chinese. The sleep-related seed keywords/expressions were chosen based on two criteria: (1) terms related to sleep derived from the Pittsburgh Sleep Quality Index (40) and (2) synonyms of these terms identified in the microblog corpus. The seed list of sleep-related keywords/expressions included insomnia, staying up late, having many dreams, nightmare, waking up startled, sleepy, early morning, having a dream, unable to fall asleep, sleep, easy to wake up, and dreaming about. The criteria for selecting seed keywords/expressions related to MIL were as follows: (1) identifying terms associated with MIL from the Meaning in Life Questionnaire (16); and (2) selecting synonyms of these keywords/expressions found in the microblog text. The MIL seed list included meaning, purpose, value, faith, ideal, aspiration, future, pursuit, quest, seek, establish, presence, and exploration.
In this study, posts were retained in their original Chinese form throughout the entire processing pipeline, including preprocessing (data cleaning and augmentation), word embeddings, and modeling, without intermediate translation that might introduce semantic distortion. An initial review revealed substantial noise, such as advertisements and http links. We employed a stop-word list containing 469 entries to filter out noisy posts. Examples were provided in Supplementary Table A1. The complete list is available at the “Dataset and Code” link (see Data Availability). This list comprised common Chinese function words, punctuation marks, and high-frequency non-informative terms identified in the corpus [e.g., terms related to advertising and fan-engagement super-topic hashtags (#)]. In addition, we removed duplicate texts longer than 10 Chinese characters. After cleaning, the dataset comprised 3,410,469 posts.
This dataset was then screened using the MIL seed keywords/expressions, yielding 189,213 posts that constituted the high-density MIL dataset. From this subset, we randomly selected 2,000 posts, divided them into 10 files, and conducted cross-annotation with the assistance of 10 psychology undergraduates. For each post, the annotator performed three binary classification tasks: (1) determine if the post is “Related” or “Not Related” to MIL or “Unable to Judge”; (2) for MIL-related posts, assess as “High POM” or “Low POM” or “Unable to Judge”; and (3) for MIL-related posts, assess as “High SFM” or “Low SFM” or “Unable to Judge.” Detailed examples and annotation guidelines are provided in Supplementary Tables B1, B2. Specifically, for MIL we excluded 532 posts in total (184 with inconsistent labels and 348 labeled as Unable to Judge), leaving 1,468 consistently labeled posts (68.8% Related, 31.2% Not Related). Within the 1,010 MIL-related posts, we excluded 176 with inconsistent SFM labels and 154 labeled as Unable to Judge, leaving 680 posts for SFM (63.4% High, 36.6% Low). For POM, we excluded 207 with inconsistent labels and 181 labeled as Unable to Judge, leaving 622 posts for POM (49.4% High, 50.6% Low). Cohen’s kappa was 0.84 for MIL, 0.71 for SFM, and 0.68 for POM (all p < 0.01).
Data augmentation was then applied to the consistently labeled dataset (excluding Unable to Judge cases) using two strategies: random deletion and random synonym replacement. Random deletion involved removing one or two noncritical words (for example, adverbs or intensifiers) without altering the central meaning. Random synonym replacement substituted words with semantically close synonyms (cosine similarity >0.80 using Sentence-BERT embeddings). To ensure semantic validity, a random 10% of the augmented posts was manually checked by two annotators, and items with altered meaning were discarded. After augmentation, the dataset was expanded to 3,000 posts for Model 1 (MIL), and 1,600 posts each for Model 2 (SFM) and Model 3 (POM).
2.1.2 Participants
We recruited Sina Weibo users through two channels: the AiShiyan Participant Recruitment Platform and the Weibo Super Topic “#Questionnaire#.” This resulted in 222 users and 701 users signing up, respectively, for a total of 923 participants. Informed consent was obtained from all participants. They were asked to complete the Meaning in Life Questionnaire (2) and to provide their user ID. Among them, 448 individuals completed the questionnaire. We then defined active users as those who posted more than 10 original microblogs between January 1, 2020, and May 4, 2022. We filtered out these active users and also removed prominent marketing accounts, which were characterized by exceptionally high numbers of comments, likes, and shares per post. After these exclusions, the dataset included 409 users and 67,852 posts. We excluded participants who failed the lie-detection item. The instruction was: “Please select ‘Strongly Agree’ for this question.” We also excluded patterned responses (i.e., selecting the same option for all items). This yielded 315 valid participants and 55,476 microblogs.
Finally, we applied the cleaning rules described in section 2.1.1. Specifically, we filtered posts containing advertisements, http links, or items from the customized stop-word list, and we removed duplicate posts longer than 10 Chinese characters. After cleaning, we retained 289 active users and 30,743 posts, averaging 106.38 posts per user and 44.25 characters per post. Detailed demographics were listed in Supplementary Table E3. The overall process is illustrated in Figure 2.
2.2 Study 1: a three-stage process for structurally assessing MIL from social media text
This study implemented a three-stage pipeline to assess MIL from microblog text. Model 1 detected whether a post was related to MIL. Model 2 classified posts into high or low levels of SFM. Model 3 classified posts into high or low levels of POM. In each stage, we adopted Bidirectional Encoder Representations from Transformers (BERT) (41) as the primary encoder to convert tokenized text into contextual representations. After encoding, a Text Convolutional Neural Network (TextCNN) (42) applied multiple convolution filters to capture the significance of keywords and enhance the models’ ability to identify key semantic cues. The input for all three models was raw post text, and the output was class probabilities and a predicted label.
The three models shared a five-layer architecture, as illustrated in Figure 3. The first layer was the BERT encoder with 12 transformer blocks. Shallow blocks captured lower-level semantics and deeper blocks encode higher-level semantics. For each input, the classification token vector (CLS; i.e., the representation of the special [CLS] token used by BERT) from each block was extracted and stacked to form a matrix that served as the input to the convolutional module. The second layer was the convolutional layer. Three sets of convolution filters with window sizes of 3 × d, 4 × d, and 5 × d were applied, where d denoted the dimensionality of the BERT embeddings (typically 768) and m was the number of filters for each window size. These filters slid over the stacked [CLS] matrix to extract local features and generate one-dimensional feature maps. The third layer was the pooling layer, which applied max pooling to reduce dimensionality and retain the most informative signals. The fourth layer was the fusion layer, which concatenated the pooled outputs from the three window sizes into a single vector. The fifth layer was the output layer, where a fully connected layer followed by the softmax function predicted task-specific class probabilities.
Figure 3. Structures of the three MIL assessment models. Model 1 determines whether a post is MIL-related (Yes/No). If a post is MIL-related, Model 2 evaluates the SFM level (High/Low) and Model 3 evaluates the POM level (High/Low). Detailed symbol definitions are provided in section 2.2.
The annotated microblog dataset was randomly divided into training, validation, and test sets in an 8:1:1 ratio. Each microblog was tokenized and mapped to subword units with special tokens added, and sequences were truncated or padded to a length of 164. Training used a batch size of 32 and a learning rate of , and was performed using a 10-fold cross-validation approach to ensure robust model evaluation and prevent overfitting. During training, prediction errors were calculated, and the parameters were iteratively adjusted using the Adam optimizer (43). Early stopping was applied when validation performance did not improve for 1,000 update steps. Class imbalance was handled with class weights in the loss function. The validation set was used to tune model hyperparameters, and the test set was used for the final evaluation of model effectiveness. Using identical architecture and training settings across the three tasks ensured comparability of results.
2.3 Study 2: identifying associated factors of MIL through semantic dependency graph algorithms
We used the Language Technology Platform (LTP) Python package (44) to construct semantic dependency graphs. These graphs were intended to collect elements associated with MIL and to illustrate potential co-occurrence and directional tendencies among them, not causal pathways. We began from such associations to identify semantic cues that explained why something happened, which was relevant to MIL. We then mapped the extracted elements to Linguistic Inquiry and Word Count (LIWC) categories to enable standardized comparisons across groups.
As input to the LTP package, each post was segmented into linguistic units (i.e., sentences ending with a period, exclamation mark, or question mark), and the output was a set of dependency relations labeled with semantic roles. LTP provided a neural pipeline for Chinese that performed sentence segmentation, part-of-speech tagging, dependency parsing with a graph-based parser, and semantic role labeling. We focused on nine key semantic dependency roles: reason (REAS, i.e., the cause or motivation behind an action), agent (AGT, i.e., the entity performing the action), experiencer (EXP, i.e., the entity that perceives or experiences an event), object (PAT, i.e., the entity that is affected by the action), content (CONT, i.e., the subject matter or information conveyed), dative (DATV, i.e., the recipient of something in a transaction), link (LINK, i.e., a relationship or connection between entities), temporal (TIME, i.e., the time at which an action occurs), and locative (LOC, i.e., the place where an action occurs). Each dependency role was represented as a three-tuple <(wordA, wordB), Role>.
As shown in Figure 4, the outputs of LTP were iteratively expanded from the REAS role (Layer 0) to subsequent layers (Layers 1–3), thereby constructing the semantic dependency graph and identifying relationships between semantic roles. Nodes in the graph were lexical items normalized to surface forms after tokenization. Edges were added when two nodes were linked by any of the nine roles within a sentence. Edge weights were the corpus counts of such links aggregated across sentences and posts, and self-links were removed. For readability and robustness, we filtered stop words and punctuation and pruned edges with very low frequency. This graph served as the foundation for detecting associated factors and mapping them to LIWC categories. A high-level pseudocode of this process was provided in Supplementary Table C1.
Figure 4. Examples of semantic dependency graphs of microblog texts. The color transition from dark to light indicated the relationship between the current role and REAS role, moving from closer to more distant connections. The nine semantic roles are reason (REAS), agent (AGT), experiencer (EXP), object (PAT), content (CONT), dative (DATV), link (LINK), temporal (TIME), and locative (LOC). ROOT represents the virtual root node that anchors the sentence’s semantic head (main predicate).
Next, we reviewed the LIWC Chinese Dictionary, and extracted keywords (associated components) from the semantic dependency graphs that were indicative of MIL, thereby mapping the associated factors to their corresponding LIWC categories. The step-by-step pseudocode and the definitions of variables and symbols were provided in Supplementary Tables C2, C3. We adopted 52 LIWC categories and consolidated them into nine broader groups (see Supplementary Table D1).
2.4 Study 3: multi-level network analysis of associated factors triggering MIL
To examine the associated factors identified in Study 2 at a higher level of community detection, we conducted network analysis based on 295,777 posts containing MIL-related associated factors. All posts were divided into eight sub-datasets based on POM (High/Low) and SFM (High/Low) levels, corresponding to eight subnetworks. Four sub-networks captured single-dimensional variations (High SFM, Low SFM, High POM, Low POM), while the other four reflected the joint distribution of both dimensions (High POM and High SFM, High POM and Low SFM, Low POM and High SFM, Low POM and Low SFM). This design allowed us to analyze how different levels of SFM and POM, individually and jointly, shaped the manifestation of MIL in social media.
In each subnetwork, nodes represented associated factors extracted from semantic dependency graphs in Study 2, while edges represented correlations between these factors. Edge strength was measured using Pearson correlations, and the extended Bayesian information criterion was applied to identify connections and prevent overfitting (45). Centrality index, particularly expected influence, was used to quantify the relative importance of each node by accounting for both the magnitude and direction of connections (46). The bootnet package in R (Version 1.5.0) (46) was used to estimate the network structure, and the qgraph package (Version 1.9.2) (47) was used to visualize the networks and calculate centrality.
Furthermore, we applied the Louvain method (48) to cluster factors within each subnetwork across different SFM and POM levels. In the Louvain method, a community referred to a group of nodes (factors) that were more densely connected within the group than with nodes outside. Density was defined as the ratio of the number of observed edges (E) to the maximum possible number of edges among N nodes, i.e.,
Modularity quantitatively measured the quality of a partition by comparing the density of within-community connections to the density expected under a random graph that preserved node degrees. Its standard definition is:
The procedure comprised three steps: (1) measured link density within communities, (2) clustered nodes to maximize modularity, and (3) merged communities and repeated the process until modularity no longer improved, thereby optimizing the partition of the MIL network.
3 Results
3.1 Performance and applications of MIL assessment models
3.1.1 Performance of MIL assessment models
We integrated BERT with TextCNN and fine-tuned three binary classifiers for the three-stage MIL assessment. Model 1 detected whether a post was related to MIL. Model 2 classified posts into high or low levels of SFM, and Model 3 classified posts into high or low levels of POM. Results (Table 1) indicated that the three models performed well, with consistently high validation and test accuracies (all above 88%). Model 1 achieved a validation accuracy of 93.67% and a test accuracy of 93.67%, demonstrating robust performance in the binary classification of MIL-related content. Model 2 reached a validation accuracy of 88.75% and a test accuracy of 90.62%, reflecting its effectiveness in capturing individuals’ varying levels of SFM. Model 3 also achieved strong results, with a validation accuracy of 95.00% and a test accuracy of 93.12%, confirming its reliability in distinguishing between high and low levels of POM. Taken together, these findings support the feasibility of our modeling approach for assessing MIL and its two core components from social media texts. This performance is conditional on the constructed dataset. We will then evaluate the approach on larger-scale datasets.
We further compared the model performance with two commonly used deep learning models for natural language classification tasks: (1) Long Short-Term Memory (LSTM) and (2) Enhanced Representation through Knowledge Integration (ERNIE). Taking Model 1 as an example, as shown in Table 2, the “BERT + TextCNN” model achieved a precision of 0.92, recall of 0.89, and an F1 score of 0.90, outperforming other models on the same dataset. For instance, standalone BERT had a lower F1 score of 0.85, and “BERT + LSTM” and ERNIE achieved F1 scores of 0.60 and 0.72, respectively.
Table 2. Performance of the MIL assessment model (model 1) compared with baseline deep learning methods.
3.1.2 Applications to a large-scale microblog dataset
We then applied the MIL assessment models to a large corpus of 3,410,469 microblog posts. We constructed a stratified random sampling frame to validate the models using human cross-validation and ChatGPT-based validation (Table 1). First, Model 1 identified 295,777 MIL-related posts (8.67%) from the full corpus; Models 2 and 3 then assessed MIL levels (High vs. Low) for SFM and POM. Posts were categorized into four groups: High SFM and High POM, High SFM and Low POM, Low SFM and High POM, and Low SFM and Low POM. These groups were used for associated factor extraction and MIL network analysis. Second, for these 295,777 MIL-related posts, we built semantic dependency graphs (section 2.3, Study 2) and obtained 43,338 posts (14.65%) with identified associated factors and LIWC categories (Table 3). Third, from this 43,338-post frame, we drew a stratified sample of approximately 1% (n = 462) with at least 100 items per SFM × POM quadrant. This size follows common practice for large-corpus quality checks, yielding an overall 95% CI with a half-width of approximately 5 percentage points and, per quadrant, approximately 9 to 10 percentage points, balancing coverage and annotation cost. The realized composition was High SFM and High POM: 102 (0.88%), High SFM and Low POM: 111 (1.22%), Low SFM and High POM: 134 (2.22%), and Low SFM and Low POM: 115 (0.69%). In the validation sample, the marginal distributions were SFM (High 46.1%, Low 53.9%) and POM (High 51.1%, Low 48.9%), indicating how the two components were represented.
Building on the sampling procedure described above, we then used the random subset of 462 posts from the 43,338 posts with identified associated factors to conduct a manual validation by four psychology students and a ChatGPT-based validation. Specifically, we implemented a Python program that accessed the gpt-3.5-turbo model via the OpenAI API. The detailed prompt was provided in Supplementary Table E1. This prompt used a one-shot prompting strategy. An example post was presented first to instruct the model on the task. Each new post in the out-of-sample subset was then evaluated with the same prompt. Both the manual labels and the assessment results of our models (Models 1, 2, and 3) were blinded to ChatGPT. We reported joint-label accuracy in Table 1, which requires the SFM and POM labels to be correct simultaneously. In manual validation, Model 1 achieved 90.40% accuracy, Model 2 achieved 74.90%, and Model 3 achieved 75.00%. In the ChatGPT-based validation, accuracies were 64.10, 41.00, and 47.50% for Models 1, 2, and 3, respectively.
In addition, the user-survey validation showed a significant correlation between MIL scores and the proportion of MIL-related posts (p = 0.006), further supporting the effectiveness of Model 1 (see Supplementary Table E2). This analysis was conducted among participants who had posted MIL-related content, as identified by Model 1.
3.2 Results of semantic dependency graphs and associated factors in MIL
3.2.1 Performance of the associated-factor extraction algorithm
For the 295,777 MIL-related posts identified by Model 1, we built semantic dependency graphs and extracted 43,338 instances with identified associated factors and corresponding LIWC categories. Each post could contain multiple associated factors across LIWC categories. LIWC categories were ranked by frequency (descending). The highest-ranked was labeled Top 1, and the top three were labeled Top 3. Posts were categorized into four groups based on SFM and POM, with examples shown in Table 3. Manual cross-validation was conducted on the sampled dataset described in section 3.1.2 (n = 462) by four psychology students. Their evaluation assessed whether the Top 1 category matched the MIL associated factor (Cohen’s kappa = 0.412) and whether the Top 3 categories covered the factor (Cohen’s kappa = 0.566). With manual cross-validation, our algorithms achieved Top-1 and Top-3 accuracies of 0.560 and 0.830. With ChatGPT-prompt validation, the corresponding accuracies were 0.519 and 0.619 (Table 3).
3.2.2 Proportions of associated factors across eight MIL groups
We examined the distribution of 52 LIWC Chinese Dictionary categories across the eight groups defined by levels of POM and SFM (Figure 5). Several patterns emerged. First, groups with higher POM and lower SFM showed higher proportions of everyday activity categories. For example, terms related to daily routines (e.g., “ingest” and “eat”). Second, groups with lower POM and lower SFM showed higher proportions of second-person address, reflected in the frequent use of “you.” Third, the Interrogation category was most frequent in SFM High posts, including terms such as “when” and “what.” Finally, death-related content appeared more often in groups with higher POM. Among the eight groups, High SFM and High POM and High POM showed the highest rates for the LIWC Death category at 15.08 and 14.33%, respectively.
3.3 Multi-level network analysis of associated factors in MIL
After dividing MIL-related posts into eight sub-datasets by POM (High/Low) and SFM (High/Low) levels, we first presented a centrality analysis to identify influential factors (section 3.3.1). We then showed pairwise relationships highlighting the strongest co-occurrences between factors (section 3.3.2). Finally, we compared structural patterns across subnetworks (section 3.3.3), focusing on clustering differences and network-wide contrasts.
3.3.1 Centrality analysis of nodes influencing MIL
This section aimed to identify which factors function as key nodes within each subnetwork and to highlight differences in node centrality across subnetworks. The 52 associated factors were further grouped into nine higher-level categories to reveal more concentrated and interpretable regularities (see Figure 6). Node centrality indices in the High SFM and Low POM subnetwork were shown as an example in Figure 7. The “Attitude” category exhibiting the highest node centrality indices (indicating the greatest influence), followed by “Emotion” and “Inner Thoughts.” This suggested that these categories contained numerous factors that served as key bridges in MIL expressions (e.g., “assent” and “compare”). Detailed factors were listed in Supplementary Table D1. Within the “Topic” category, “Bio” emerged as the most influential, underscoring the critical role of physical state in shaping MIL.
Figure 6. Node centrality indices of eight MIL subnetworks. (a) Node centrality indices of 9 groups. (b) Node centrality indices of the “Topic” group. For example, in panel (a), the “Attitude” category shows the highest node centrality values across all subnetworks, suggesting that it contains many factors that play key bridging roles in MIL expressions (e.g., “assent” and “compare”). Additional factors in this category are listed in Supplementary Table D1.
Figure 7. MIL-factor network structure and node centrality: using the high SFM & low POM subnetwork as an example. Each node corresponds to one of the 52 LIWC categories, aggregated into nine groups, each shown in a different color. Thicker edges indicate greater edge strength. For example, the nodes with the highest centrality are “bio,” “drive,” and “perception,” suggesting that these factors play a stronger connective role in shaping individuals’ MIL expressions in this state.
The complete results for the 52 factors across the eight subnetworks were presented in Supplementary Table F1. Key findings included. (1) Across all subnetworks, the nodes with higher centrality were “bio,” “perception,” “drive,” “negative emotion,” and “social.” (2) “Shehe” ranked higher in the High SFM and Low POM subnetwork (node centrality index = 1.13). (3) “Social” and “work” showed their highest centrality in Low SFM and High POM (node centrality index = 1.37). (4) “Positive emotion” and “religion” rank highest in High SFM and High POM (node centrality indices = 1.20 and 0.41, respectively). (5) “Achieve” was highest in High SFM contexts (node centrality index = 0.82). (6) “Compare” showed higher centrality in Low POM subnetworks than in the other subnetworks (node centrality index = 0.96).
3.3.2 Latent relationships among associated factors in MIL
We highlight the strongest pairwise co-occurrences among MIL-associated factors and illustrate how these relationships differ across subnetworks. The top 15 correlations for each subnetwork were listed in Supplementary Tables F2, F3, with the strongest pairs shown in boldface. The full 52 × 52 matrices were available at the “Dataset and Code” link (see Data Availability). Highlighted structures for all subnetworks were shown in Supplementary Figures F1, F2. Each node represented a LIWC category. Thicker edges and shorter inter-node distances indicated stronger correlations between categories. Key observations included. (1) Across all conditions, “space” and “time” were strongly correlated with the “relative” theme in MIL discussions. (2) In the High POM panel of Supplementary Figure F2, the “body” and “bio” nodes were close and connected by a thick edge. Consistent values were observed, with r = 0.779 for Low SFM and High POM and r = 0.810 for High SFM and High POM (both p < 0.001) (Supplementary Tables F2, F3). (3) In the Low POM panel of Supplementary Figure F2, the “power” and “drive” nodes were close with a thick edge. Corresponding values were observed, with r = 0.783 for Low SFM and Low POM and r = 0.790 for High SFM and Low POM (both p < 0.001) (Supplementary Tables F2, F3).
3.3.3 Structural comparison of MIL subnetworks
We contrasted network-wide structures across the eight subnetworks, emphasizing clustering differences as well as global strength and weight invariance. The structural comparisons were presented in Supplementary Figure F3. The results indicated the following. (1) Across MIL levels, “space” and “time” appeared in different clusters: in SFM High and POM Low, “space” and “time” clustered with “focus past,” “focus present,” and “focus future”; in SFM Low and POM Low, they clustered only with “focus past” and “focus present.” (2) In SFM High and POM High, career-related factors (“power,” “money,” “achieve”) clustered with “risk,” as well as with “positive emotion.”
Additionally, we applied the network comparison test to evaluate differences among the eight MIL subnetworks. We summarized each network’s overall edge strength and node proximity using global strength and network weight, and we compared these two indices pairwise across subnetworks. As shown in Figure 8, the largest differences occurred between the SFM Low and POM Low and SFM High and POM High subnetworks, whereas the smallest differences occurred between the SFM Low and POM Low and SFM High and POM Low subnetworks. These patterns suggested that POM exerted a stronger organizing influence on the MIL-related factor structure than SFM. When POM was low, changing SFM produced minimal structural change, while networks diverged most when both POM and SFM shifted from low to high.
Figure 8. Global strength invariance and network weight invariance across subnetworks. Colors indicate the degree of invariance between pairs of subnetworks. Darker shades denote higher invariance. (a) Invariance in global edge strength. (b) Invariance in edge weights. S0, low SFM; S1, high SFM; P0, low POM; P1, high POM.
4 Discussion
This study constructed three-stage deep learning MIL assessment models and extracted associated factors from MIL-related microblogs. The results aligned with Steger’s three-dimensional model (20), which included coherence, purpose, and significance. Coherence referred to making sense of one’s life and integrating experiences over time and context. Purpose referred to having valued goals and a sense of direction. Significance referred to perceiving one’s life as worthwhile and important. (1) Our findings on temporal and spatial components reflected the coherence dimension. Under Low POM, “space” and “time” clustered with “focus future” only in the High SFM and Low POM condition, not in Low SFM and Low POM (see Supplementary Figure F3). This pattern was consistent with the perspective that individuals high in search but lower in presence tend to look ahead and evaluate meaning using future-oriented criteria (49). (2) Results on purpose-related factors were evidenced in two ways. We operationalized purpose-related factors as including “work,” “social,” “achieve,” “power,” “money,” and “risk,” based on the LIWC/associated-factor mapping (Supplementary Table D1). This set was indicative rather than exhaustive, and other categories might also reflect purpose depending on context. First, for both “social” and “work,” centrality was highest in Low SFM and High POM (Supplementary Table F1). Second, career-related factors (“power,” “money,” “achieve”) clustered with risk in the High SFM and High POM subnetwork (Supplementary Figure F3). (3) For the significance dimension, “positive emotion,” “religion,” and “achieve” jointly served as indicative cues of perceived significance. Supplementary Table F1 showed that “positive emotion” and “religion” exhibited the highest centrality in High SFM and High POM, and “achieve” was highest in High SFM contexts. The preliminary convergence between the two-dimensional and three-dimensional MIL frameworks was exploratory rather than confirmatory. Further investigation was needed to provide definitive evidence.
According to the results reported in section 3.2.2 and Figure 5, corpus-level tendencies could be read through Self-Determination Theory (3). Language emphasizing everyday activities in higher POM and lower SFM groups might have indicated a greater focus on autonomy-related routines, whereas the higher use of second person forms in lower POM and lower SFM groups may have reflected a stronger orientation to social-relatedness in this context. The prominence of interrogation terms in High SFM posts was consistent with work connecting exploratory tendencies with active information seeking (50). In addition, prior research reported a negative association between POM and death anxiety (11). While LIWC “Death” reflected factor frequency rather than anxiety, the higher rates of death-related factors in High POM groups might be compatible with an interpretation that individuals with greater POM could approach mortality themes with less anxiety (i.e., more open, approach-oriented processing), though this remains speculative. These interpretations should be treated as tendencies within this corpus and might depend on platform and cultural conventions rather than stable person-level traits.
The centrality patterns reported in section 3.3.1 offered a coherent picture of how different factors may shape MIL (see Supplementary Table F1). The prominence of “attitude,” “emotion,” and “inner thoughts” at the category level suggested that evaluative stance, affective tone, and reflective cognition were central to meaning construction. The high centrality of “biology” was consistent with accounts linking physical condition to MIL, including associations with lower pain, anxiety, and depression, better illness acceptance, and improved quality of life [(e.g., 51–53)]. Likewise, the relatively high values for “social,” “work,” and “leisure” aligned with evidence that social roles and support are positively related to MIL (54). In addition, subnetwork-specific patterns reported were consistent with prior theorizing. Positive emotion showed a lower rank under Low POM. By contrast, it showed a higher rank under High POM and High SFM. This pattern was consistent with the view that diminished POM was accompanied by attempts to reduce negative affect. Meanwhile, higher MIL was associated with cultivating positive affective states (7). The greater centrality of “compare” in Low POM subnetworks echoed research on social comparison as a means of status appraisal under uncertainty (55). The higher centrality of “shehe” in Low POM and High SFM might reflect an outward orientation toward models or referents when search was high but presence was limited (56).
Building on the correlation patterns reported in section 3.3.2 (Supplementary Figures F1, F2; Supplementary Tables F2, F3), we interpreted three descriptive regularities. First, pairs involving “space”/“time” with “relative” recurred among the higher correlations across contexts. This was consistent with the idea that appraisals of change and continuity drew on temporal and spatial comparisons (22). Second, within High POM, the association between “body” and “bio” was among the higher pairs. This aligned with the view that connected bodily states and recovery experiences were linked to meaning. Prior work linked meaning to pain experiences and related higher POM to lower health anxiety (57, 58). Third, within Low POM, the association between “power” and “drive” was also among the higher pairs. This fit perspectives that lower POM could co-occur with compensatory striving (59). These interpretations were intended to contextualize the observed correlations and remained descriptive rather than inferential. No causal claims were intended.
Building on the structural patterns reported in section 3.3.3 (Supplementary Figure F3), we offered two descriptive interpretations. First, under High POM and High SFM, the alignment of risk with career-related factors and positive emotion fit perspectives that more satisfied individuals pursued new achievements even in the face of risk (60). Second, “we” clustered with “temporal” “spatial” and “comparative” factors in Low SFM and Low POM, but with “social” factors in High POM and Low SFM. This pattern was compatible with the view that higher presence related to finding meaning in social roles and positive interactions (61). These interpretations were descriptive rather than inferential.
This study adopted a dual data collection strategy, combining a large-scale corpus of over 3 million publicly available microblogs with a smaller participant dataset that integrated both self-report questionnaires and personal microblogs. The large-scale dataset enabled population-level analysis of MIL, while the participant dataset provided validated ground truth by linking subjective measures to corresponding online expressions. These complementary datasets allowed model validation from three perspectives: human cross-annotation, ChatGPT-based labeling, and user self-report surveys. Each method carried distinct strengths and limitations. Human annotation, often considered the gold standard, offered nuanced interpretations but was resource-intensive and subject to inter-annotator variability. ChatGPT-based labeling was used as a supplemental convergent check of human annotation, as it provided scalability and efficiency (62). However, the outputs of ChatGPT were sensitive to prompts and sometimes prone to hallucination (63). In addition, ChatGPT models pretrained on general corpora had difficulty in accurately identifying and evaluating the construct ambiguity of MIL, as the boundaries between POM and SFM were sometimes overlapping. Therefore, although the judgments of ChatGPT-based validation were largely consistent in direction with human annotations, the overall accuracy was lower. User surveys grounded predictions in participants’ self-reports and enhanced ecological validity, though they remained vulnerable to recall bias, social desirability, and limited sample size. Taken together, the convergences and divergences across these approaches underscored the value of triangulation: manual coding secured high-quality benchmarks, GPT-assisted annotation enabled efficient large-scale analysis, and user self-report anchored computational predictions to psychological ground truth. Future work may benefit from hybrid strategies that combine LLM-assisted pre-annotation with human oversight, alongside triangulation using self-reported measures (64).
The performance of our proposed MIL assessment models on the large-scale dataset varied substantially. Model 1 identified whether a post was related to MIL, while Models 2 and 3 classified the two dimensions of MIL, namely POM and SFM. As dimensions of MIL, POM and SFM were conceptually more complex and theoretically debated. For instance, previous debates have concerned whether the two dimensions overlap or should be separated, as proposed in three-dimensional models (19, 20). This conceptual ambiguity increased task difficulty. Moreover, both dimensions represented latent psychological constructs that were expressed more implicitly in social media text, making them more challenging for both human annotation and model classification. During training data construction, the inter-annotator agreement (Cohen’s kappa = 0.84 for MIL, 0.71 for SFM, and 0.68 for POM; all p < 0.01; see section 2.1.1) reflected this trend. Therefore, the pattern in which Model 1 (≈90%) outperformed Models 2 and 3 (≈75%) was consistent with human annotation reliability. As these estimates were based on a sampled subset, they should be interpreted as exploratory references rather than definitive conclusions.
This study had limitations. First, we did not differentiate by demographics. Because MIL could arise from different sources at different life stages, categorizing participants could facilitate understanding of associated factors within groups. Second, our MIL assessment models focused on the post level rather than on users’ posts over time. Assessing MIL at the user level could address the sparsity of microblog data and add temporal cues, potentially improving model performance. Third, in the multi-level network analysis of MIL-associated factors, no inferential statistics were performed. All statements were descriptive rather than inferential and were not presented as evidence of statistical significance or causality. Future studies will test these differences with formal inferential methods. Fourth, the representativeness and cultural generalizability of our findings were limited. Our text corpus consisted of posts written in Chinese and posted at night on the Sina Weibo platform. Because culture and platform norms may have shaped language use, expressions of MIL associated factors might differ on other social media platforms and in other countries. In addition, the participants in the empirical study were demographically specific (see Supplementary Table E3). For example, 69.9% were under 24 years old and 72.3% were undergraduates, master’s, or PhD students. Socioeconomic status was not collected. Therefore, replication across platforms, languages, cultures, and demographics was needed before broader generalization. In addition, future work could assess MIL in non-sleep contexts to enable comparative analyses and further examine generalization.
Three future research directions were envisioned. First, we will examine the relationship between perceived social support in social networks and users’ MIL. Interpersonal relationships are expected to be a significant source of meaning and a predictor of MIL (65). Clues related to “social” and “friend” were identified in this study, indicating that the category “we” clustered with social categories only under the condition of High SFM and High POM. Second, images in social networks, which contain rich visual information reflecting users’ interests, values, and emotional states (66), will be investigated to determine whether visual information can improve the precision of MIL assessment. Third, this study provided a preliminary exploration of using prompts based on a pretrained large language model to label components and levels of MIL, but the performance was limited. Future studies may attempt fine-tuning large language models on manually annotated MIL corpora to further improve both efficiency and accuracy of automatic MIL assessment.
5 Conclusion
Using social media text, we proposed a structured three-part framework for assessing MIL along SFM and POM and validated the models with multiple approaches. Through graph-based semantic analysis, we identified associated factors from MIL expressions. We then explored their large-scale patterns across groups. These findings outline context-dependent topic and network differences associated with MIL levels and provided a data-driven lens for monitoring MIL-related signals in online populations. While exploratory and descriptive, this work may help prioritize factors for follow-up assessment and inform the design of supportive, evidence-guided interventions.
Data availability statement
The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found in the article/Supplementary material.
Ethics statement
The studies involving humans were approved by School of Psychology, Beijing Normal University. The studies were conducted in accordance with the local legislation and institutional requirements. Written informed consent for participation in this study was provided by the participants. The social media data was accessed and analyzed in accordance with the platform’s terms of use and all relevant institutional/national regulations.
Author contributions
QL: Methodology, Writing – review & editing, Funding acquisition, Writing – original draft, Visualization, Formal analysis, Conceptualization. MW: Writing – original draft, Data curation, Conceptualization, Formal analysis, Methodology. JY: Conceptualization, Writing – original draft. WJ: Methodology, Writing – original draft, Data curation, Conceptualization, Formal analysis. LZ: Visualization, Writing – original draft, Funding acquisition, Conceptualization. XW: Conceptualization, Writing – original draft, Visualization, Data curation. BY: Visualization, Data curation, Writing – original draft. LC: Methodology, Writing – review & editing, Funding acquisition, Conceptualization, Formal analysis.
Funding
The author(s) declare that financial support was received for the research and/or publication of this article. This study was funded by the Beijing Education Science “14th Five-Year Plan” Project (grant No. CCGA25167).
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The authors declare that no Gen AI was used in the creation of this manuscript.
Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpubh.2025.1642085/full#supplementary-material
References
1. Frankl, VE. Man’s search for meaning: An introduction to logotherapy. New York, NY: Washington Square Press (1963).
2. Steger, MF, Frazier, P, Oishi, S, and Kaler, M. The meaning in life questionnaire: assessing the presence of and search for meaning in life. J Counsel Psychol. (2006) 53:80–93. doi: 10.1037/0022-0167.53.1.80
3. Martela, F, Ryan, RM, and Steger, MF. Meaningfulness as satisfaction of autonomy, competence, relatedness, and beneficence: comparing the four satisfactions and positive affect as predictors of meaning in life. J Happiness Stud. (2018) 19:1261–82. doi: 10.1007/s10902-017-9869-7
4. Hallford, DJ, Mellor, D, Cummins, RA, and McCabe, MP. Meaning in life in earlier and later older-adulthood: confirmatory factor analysis and correlates of the meaning in life questionnaire. J Appl Gerontol. (2018) 37:1270–94. doi: 10.1177/0733464816658750
5. Yu, EA, and Chang, EC. Depressive symptoms and life satisfaction in Asian American college students: examining the roles of self-compassion and personal and relational meaning in life. Asian Am J Psychol. (2020) 11:259–68. doi: 10.1037/aap0000214
6. Yu, Y, Yu, Y, and Hu, J. COVID-19 among Chinese high school graduates: psychological distress, growth, meaning in life, and resilience. J Health Psychol. (2022) 27:1057–69. doi: 10.1177/1359105321990819
7. Bergman, YS, Bodner, E, and Haber, Y. The connection between subjective nearness-to-death and depressive symptoms: the mediating role of meaning in life. Psychiatry Res. (2018) 261:269–73. doi: 10.1016/j.psychres.2017.12.078
8. Boreham, ID, and Schutte, NS. The relationship between purpose in life and depression and anxiety: a meta-analysis. J Clin Psychol. (2023) 79:2736–67. doi: 10.1002/jclp.23576
9. Czekierda, K, Banik, A, Park, CL, and Luszczynska, A. Meaning in life and physical health: systematic review and meta-analysis. Health Psychol Rev. (2017) 11:387–418. doi: 10.1080/17437199.2017.1327325
10. Lew, B, Chistopolskaya, K, Osman, A, Huen, JMY, and Leung, ANM. Meaning in life as a protective factor against suicidal tendencies in Chinese university students. BMC Psychiatry. (2020) 20:73. doi: 10.1186/s12888-020-02485-4
11. Jin, Y, Zeng, Q, Cong, X, and Li, H. Impact of death anxiety on mental health during COVID-19: the mediating role of the meaning in life. J Pac Rim Psychol. (2023) 17:1–9. doi: 10.1177/18344909231165187
12. Kleiman, EM, and Beaver, JK. A meaningful life is worth living: meaning in life as a suicide resiliency factor. Psychiatry Res. (2013) 210:934–9. doi: 10.1016/j.psychres.2013.08.002
13. Lu, N. Proactive coping moderates the relationship between meaning of life and sleep quality in university students. Sleep Med. (2013) 14:e187. doi: 10.1016/j.sleep.2013.11.443
14. Shek, DT, Dou, D, Zhu, X, and Chai, W. Positive youth development: current perspectives. Adolesc Health Med Ther. (2019) 10:131–41. doi: 10.2147/AHMT.S179946
15. Du, H, Li, X, Chi, P, Zhao, J, and Zhao, G. Meaning in life, resilience, and psychological well-being among children affected by parental HIV. AIDS Care. (2017) 29:1410–6. doi: 10.1080/09540121.2017.1307923
16. Zhong, M, Zhang, Q, Bao, J, and Xu, W. Relationships between meaning in life, dispositional mindfulness, perceived stress, and psychological symptoms among Chinese patients with gastrointestinal cancer. J Nerv Ment Dis. (2019) 207:34–7. doi: 10.1097/NMD.0000000000000922
17. Crumbaugh, JC, and Maholick, LT. An experimental study in existentialism: the psychometric approach to Frankl’s concept of noogenic neurosis. J Clin Psychol. (1964) 20:200–7. doi: 10.1002/1097-4679(196404)20:2<200::AID-JCLP2270200203>3.0.CO;2-U
18. Morgan, J, and Farsides, T. Measuring meaning in life. J Happiness Stud. (2009) 10:197–214. doi: 10.1007/s10902-007-9075-0
19. George, LS, and Park, CL. Meaning in life as comprehension, purpose, and mattering: toward integration and new research questions. Rev Gen Psychol. (2016) 20:205–20. doi: 10.1037/gpr0000077
20. Martela, F, and Steger, MF. The role of significance relative to the other dimensions of meaning in life – an examination utilizing the three-dimensional meaning in life scale (3DM). J Posit Psychol. (2022) 18:606–26. doi: 10.1080/17439760.2022.2070528
21. Chen, S, Li, X, and Ye, S. Self-concept clarity and meaning in life: a daily diary study in a collectivistic culture. J Happiness Stud. (2024) 25:1–23. doi: 10.1007/s10902-024-00775-2
22. King, LA, and Hicks, JA. The science of meaning in life. Annu Rev Psychol. (2021) 72:561–84. doi: 10.1146/annurev-psych-072420-122921
23. Helm, PJ, Jimenez, T, Galgali, MS, Edwards, ME, Vail, KE, and Arndt, J. Divergent effects of social media use on meaning in life via loneliness and existential isolation during the coronavirus pandemic. J Soc Pers Relat. (2022) 39:1768–93. doi: 10.1177/02654075211066922
24. Kosinski, MV, and Stillwell, D. Facebook as a research tool for the social sciences: opportunities, challenges, ethical considerations, and practical guidelines. Am Psychol. (2015) 70:543–56. doi: 10.1037/a0039210
25. Chen, X, Chen, Y, and Yang, Z. The effect of proactive social networking site use on sense of meaning in life: a moderated mediation model. Chin J Clin Psychol. (2021) 29:236–41. doi: 10.16128/j.cnki.1005-3611.2021.02.004
26. Ermakov, PN, and Belousova, E. The relationship between the strategies of transferring the meanings of information messages and the meaning-of-life orientations of social network users. Int J Cogn Sci Eng. (2021) 9:279–89. doi: 10.23947/2334-8496-2021-9-2-275-289
27. Frison, E, and Eggermont, S. Exploring the relationships between different types of Facebook use, perceived online social support, and adolescents’ depressed mood. Soc Sci Comput Rev. (2015) 34:153–71. doi: 10.1177/0894439314567449
28. Burke, M., Kraut, R., and Marlow, C. (2011). Social capital on Facebook: differentiating uses and users. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. ACM.
29. Settanni, M, Azucar, D, and Marengo, D. Predicting individual characteristics from digital traces on social media: a meta-analysis. Cyberpsychol Behav Soc Netw. (2018) 21:217–28. doi: 10.1089/cyber.2017.0384
30. Gruda, D, and Hasan, S. Feeling anxious? Perceiving anxiety in tweets using machine learning. Comput Hum Behav. (2019) 98:245–55. doi: 10.1016/j.chb.2019.04.020
31. Ge, D. Relationship between the presence of meaning in life and sleep quality: a moderated chain-mediation model. J Health Psychol. (2025) 30:921–35. doi: 10.1177/13591053241249236
32. Tian, X, He, F, Batterham, P, Wang, Z, and Yu, G. An analysis of anxiety-related postings on Sina Weibo. Int J Environ Res Public Health. (2017) 14:775. doi: 10.3390/ijerph14070775
33. Wang, X., Cao, L., Zhang, H., Feng, L., Ding, Y., and Li, N. (2022). A meta-learning based stress category detection framework on social media. In Proceedings of the ACM Web Conference 2022. New York, NY: ACM.
34. Wang, X., Zhang, H., Cao, L., and Feng, L. (2020). Leverage social media for personalized stress detection. In Proceedings of the 28th ACM International Conference on Multimedia. New York, NY: Association for Computing Machinery.
35. Wang, X., Zhang, H., Cao, L., Zeng, K., Li, Q., Li, N., et al. (2023). Contrastive learning of stress-specific word embedding for social media-based stress detection. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. Association for Computing Machinery.
36. Wang, X., Feng, L., Zhang, H., Cao, L., Zeng, K., Li, Q., et al. (2025). MISE: meta-knowledge inheritance for social media-based stressor estimation. In Proceedings of the ACM Web Conference 2025. New York, NY: Association for Computing Machinery.
37. Cao, L, Zhang, H, and Feng, L. Building and using personal knowledge graph to improve suicidal ideation detection on social media. IEEE Trans Multimedia. (2022) 24:87–102. doi: 10.1109/TMM.2020.3046867
38. Thomsen, DK, Mehlsen, MY, and Christensen, S. Rumination—relationship with negative mood and sleep quality. Pers Individ Differ. (2003) 34:1293–301. doi: 10.1016/S0191-8869(02)00120-4
39. Yu, X, and Zhao, J. How rumination influences meaning in life among Chinese high school students: the mediating effects of perceived chronic social adversity and coping style. Front Public Health. (2023) 11:38106898. doi: 10.3389/fpubh.2023.1280961
40. Buysse, DJ, Reynolds, CF, Monk, TH, Berman, SR, and Kupfer, DJ. The Pittsburgh sleep quality index: a new instrument for psychiatric practice and research. Psychiatry Res. (1989) 28:193–213. doi: 10.1016/0165-1781(89)90047-4
41. Devlin, J., Chang, M., Lee, K., and Toutanova, K. (2019). BERT: pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Minneapolis: Association for Computational Linguistics (ACL).
42. Kim, Y. (2014). Convolutional neural networks for sentence classification. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing. Stroudsburg, PA, USA: ACL.
43. Kingma, D. P., and Ba, J. (2014) Adam: a method for stochastic optimization arXiv. Available online at: https://arxiv.org/abs/1412.6980.
44. Che, W., Feng, Y., Qin, L., Liu, P., Wang, Y., and Liu, T. (2021). N-ltp: an open-source neural language technology platform for Chinese. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing: System Demonstrations. Stroudsburg, PA, USA: Association for Computational Linguistics (ACL).
45. Foygel, R., and Drton, M. (2010) Extended Bayesian information criteria for Gaussian graphical models. In Advances in neural information processing systems (pp. 604–612) Curran Associates Inc. Available online at: https://api.semanticscholar.org/CorpusID:6311495
46. Epskamp, S, Borsboom, D, and Fried, EI. Estimating psychological networks and their accuracy: a tutorial paper. Behav Res Methods. (2018) 50:195–212. doi: 10.3758/s13428-017-0862-1
47. Epskamp, S, Cramer, AO, Waldorp, LJ, Schmittmann, VD, and Borsboom, D. Qgraph: network visualizations of relationships in psychometric data. J Stat Softw. (2012) 48:1–18. doi: 10.18637/jss.v048.i04
48. Blondel, VD, Guillaume, JL, Lambiotte, R, and Lefebvre, E. Fast unfolding of communities in large networks. J Stat Mech Theory Exp. (2008) 2008:P10008. doi: 10.1088/1742-5468/2008/10/P10008
49. Hicks, JA, Trent, J, Davis, WE, and King, LA. Positive affect, meaning in life, and future time perspective: an application of socioemotional selectivity theory. Psychol Aging. (2012) 27:181–9. doi: 10.1037/a0023965
50. Kashdan, TB, and Steger, MF. Curiosity and pathways to well-being and meaning in life: traits, states, and everyday behaviors. Motiv Emot. (2007) 31:159–73. doi: 10.1007/s11031-007-9068-7
51. Majid, YA, Abbasi, M, and Abbasi, M. The relationship between meaning of life, perceived social support, spiritual well-being, and pain catastrophizing with quality of life in migraine patients: the mediating role of pain self-efficacy. BMC Psychol. (2023) 11:7. doi: 10.1186/s40359-023-01053-1
52. Quinto, RM, Vincentiis, FD, Campitiello, L, Innamorati, M, Secinti, E, and Iani, L. Meaning in life and the acceptance of cancer: a systematic review. Int J Environ Res Public Health. (2022) 19:5547. doi: 10.3390/ijerph19095547
53. Testoni, I, Sansonetto, G, Ronconi, L, and Zamperini, A. Meaning of life, representation of death, and their association with psychological distress. Palliat Support Care. (2018) 16:511–9. doi: 10.1017/S1478951517000669
54. Steger, MF, Kashdan, TB, and Oishi, S. Being good by doing good: daily eudaimonic activity and well-being. J Res Pers. (2008) 42:22–42. doi: 10.1016/j.jrp.2007.03.004
55. White, JB, Langer, EJ, Yariv, L, and Welch, M. Frequent social comparisons and destructive emotions and behaviors: the dark side of social comparisons. J Adult Dev. (2006) 13:36–44. doi: 10.1007/S10804-006-9005-0
56. Chu, ST, and Fung, HH. Is the search for meaning related to the presence of meaning? Moderators of the longitudinal relationship. J Happiness Stud. (2021) 22:127–45. doi: 10.1007/s10902-020-00222-y
57. Boring, BL, Maffly-Kipp, J, Mathur, VA, and Turk, DC. Meaning in life and pain: the differential effects of coherence, purpose, and mattering on pain severity, frequency, and the development of chronic pain. J Pain Res. (2022) 15:299–314. doi: 10.2147/JPR.S338691
58. Yek, MH, Olendzki, N, Kekecs, Z, and Belsher, BE. Presence of meaning in life and search for meaning in life and relationship to health anxiety. Psychol Rep. (2017) 120:383–90. doi: 10.1177/0033294117697084
59. Church, AT, Katigbak, MS, Locke, KD, Zhang, H, Shen, J, and Alvarez, JM. Need satisfaction and well-being: testing self-determination theory in eight cultures. J Cross-Cult Psychol. (2012) 44:507–34. doi: 10.1177/0022022112466590
60. Martela, F, and Pessi, AB. Significant work is about self-realization and broader purpose: defining the key dimensions of meaningful work. Front Psychol. (2018) 9:363. doi: 10.3389/fpsyg.2018.00363
61. Klein, N. Prosocial behavior increases perceptions of meaning in life. J Posit Psychol. (2017) 12:354–61. doi: 10.1080/17439760.2016.1209541
62. Gilardi, F, Alizadeh, M, and Kubli, M. ChatGPT outperforms crowd workers for text-annotation tasks. Proc Natl Acad Sci. (2023) 120:e2305016120. doi: 10.1073/pnas.2305016120
63. Reiss, M. V. (2023). Testing the reliability of ChatGPT for text annotation and classification: a cautionary remark. arXiv.
64. Sahitaj, A., Sahitaj, P., Solopova, V., Li, J., Möller, S., and Schmitt, V. (2025). Hybrid annotation for propaganda detection: integrating LLM pre-annotations with human intelligence. In Proceedings of the Fourth Workshop on NLP for Positive Impact. Stroudsburg, PA, USA: ACL.
65. Deci, EL, and Ryan, RM. Self-determination theory In: International encyclopedia of the social and behavioral sciences. 2nd ed. Amsterdam, Netherlands: Elsevier (2014)
66. Li, Y, and Xie, Y. Is a picture worth a thousand words? An empirical study of image content and social media engagement. J Mark Res. (2019) 57:1–19. doi: 10.1177/0022243719881113
Keywords: meaning in life, social network, factor extraction, semantic analysis, deep learning
Citation: Li Q, Wang M, Yan J, Jiake W, Zhao L, Wang X, Yao B and Cao L (2025) Exploring meaning in life from social network content in the sleep scenario. Front. Public Health. 13:1642085. doi: 10.3389/fpubh.2025.1642085
Edited by:
Elke Humer, University of Continuing Education Krems, AustriaCopyright © 2025 Li, Wang, Yan, Jiake, Zhao, Wang, Yao and Cao. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Lei Cao, Y2FvbGVpQGJudS5lZHUuY24=
Mengyao Wang1