Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Psychol., 23 September 2025

Sec. Media Psychology

Volume 16 - 2025 | https://doi.org/10.3389/fpsyg.2025.1535533

This article is part of the Research TopicSubjective Well-Being and Human Decision BehaviorsView all 5 articles

Understanding the privacy disclosure behavior and privacy concerns on short-form video platforms: an empirical investigation

  • 1Department of Journalism and Communication, Shaanxi Normal University, Xi’an, China
  • 2Key Laboratory of Folk Song Intelligent Computing and Service Technology, Ministry of Culture and Tourism, Xi’an, China

In recent years, platform user privacy leaks have occurred frequently, heightening users’ privacy concerns, which may directly affect their behavior on short video platforms. Based on communication privacy management theory, this study constructs a conceptual model of factors influencing the privacy disclosure behavior of short video users from the perspectives of both peer privacy concerns and institutional privacy concerns. Data from 302 short video users were collected via an online questionnaire and analyzed using structural equation modeling. The results indicate that perceived peer risk and information sensitivity positively influence peer privacy concerns; the effectiveness of privacy protection technology and privacy policies negatively influence institutional privacy concerns; information sensitivity also positively affects institutional privacy concerns; both institutional and peer privacy concerns negatively affect users’ privacy disclosure behavior; and institutional privacy concerns positively influence peer privacy concerns. The study discusses how these findings extend the application of communication privacy management theory in the context of short video platforms, highlighting the intertwined nature of institutional and peer-induced privacy risks. Practical implications for platform designers and policymakers are offered to enhance user trust and promote responsible privacy management.

1 Introduction

Videos contain richer forms of personal information than text or static images, encompassing facial, physical, and voice features Kang et al. (2022). Consequently, users may disclose more personal information when engaging with short video platforms. The increasingly accurate and intelligent recommendation algorithms of these platforms may heighten users’ institutional privacy concerns. Furthermore, as the social features of short video platforms continue to evolve, privacy threats from peers may amplify users’ concerns, potentially influencing their privacy disclosure behaviors (Ozdemir et al., 2017). The content generated by users is a crucial factor for the survival and development of short video platforms. Users’ concerns regarding peer and institutional privacy may affect both their willingness to engage with the platform and their usage patterns, which in turn impacts the platforms’ growth. Therefore, it is of great theoretical and practical significance to explore the two facets of privacy concerns on the factors influencing the privacy leakage behavior of short video users.

The majority of studies begin with institutional privacy issues, pointing out that users’ security and privacy may be threatened by the information practices of websites, online platforms, and other organizations. Users become concerned about platforms and organizations violating their privacy as a result of this threat (Xu et al., 2011). Such institutional privacy concerns can directly affect users’ privacy disclosure behavior (Xu et al., 2013). A few studies have found that users’ privacy concerns not only originate from institutions but also exist in peer-oriented online interactions. This threat of privacy invasion from online peers is defined as peer privacy concerns (Zhang et al., 2022). The impact of a single privacy concern on users’ privacy disclosure behavior has been extensively studied in previous research, but it is unclear how the two privacy concern perspectives—which refers to the conflict between institutional and peer privacy—affects users’ privacy disclosure behavior on short video platforms and how the two facets of privacy concerns relate to one another. Furthermore, the impact of users’ privacy concerns on privacy disclosure behavior in short video platforms is not included in the current research on users’ privacy disclosure behavior, which focuses on items like social media SNS (Jozani et al., 2020), e-commerce platforms (Dinev and Hart, 2006), online medical platforms (Young et al., 2013), and financial websites (Xu et al., 2011). It is found that short video users face the two privacy concerns from institutions and peers. It is difficult to fully reveal the mechanism of privacy concerns in the privacy disclosure behavior of short video users from the perspective of institutional privacy concerns alone.

Based on the above reality and theoretical background, this paper proposes the following research questions: (1) How do privacy concerns affect users’ privacy disclosure behavior on short video platforms? (2) How do the two facets of privacy concerns work in tandem to impact user’s privacy disclosure behavior? (3) What are the antecedents of institutional and peer privacy concerns? In order to address the above questions, with communication privacy management (CPM) as the theoretical basis, this study examines the facets of concerns and benefits that may affect user’s privacy disclosure behavior on short video platforms.

This study makes two significant contributions to the IS literature. First, it demonstrates the necessity of incorporating the two privacy concerns in short video users’ privacy behavior research. Existing studies predominantly focus on single-dimensional institutional privacy concerns while neglecting the role of peer privacy concerns. This study innovatively introduces a dual privacy concerns framework, thoroughly elucidating the influencing mechanisms of users’ privacy disclosure behaviors based on the unique characteristics of short video platforms, thereby expanding the research perspective on short video users’ privacy behaviors. Second, it clarifies the formation mechanism of dual privacy concerns in the short video context. Current privacy concern research mainly concentrates on social media and e-commerce contexts, with limited exploration of the interrelationships between different privacy concerns. Grounded in CPM, this study examines how three boundary management factors (boundary rules, coordination, and turbulence) influence dual privacy concerns, while also verifying the moderating effect of institutional privacy concerns. These findings broaden the research paradigm of privacy concerns from the theoretical perspective of CPM.

2 Literature review

2.1 Privacy concerns

Privacy concerns reflect Internet users’ anxieties regarding how their personal information is collected and used by websites, highlighting perceived gaps between expected and actual data handling practices (Hong and Thong, 2013). Research primarily distinguishes two dimensions: institutional privacy concerns and peer privacy concerns. (1) Institutional Privacy Concerns: Stemming from user apprehensions about how platforms, governments, and other organizations manage personal data—specifically unauthorized access, secondary use, interception, and sale to third parties (Liu et al., 2022; Wang et al., 2019). Studies explore how these concerns influence user behaviors, including privacy disclosure (Xu et al., 2013), reduced social media engagement (Neves et al., 2022), and the privacy paradox (Livingstone, 2008). For instance, Dinev and Hart (2006) found they significantly impact disclosure willingness, while Xu et al. (2011) examined how perceptions of institutional protection shape these concerns. (2) Peer Privacy Concerns: Arise when users feel unable to maintain personal boundaries online due to peer behavior (Zhang et al., 2022). Zhang et al. conceptualized these across four dimensions: information, psychological, virtual territory, and communication privacy concerns. As social networks evolve, online privacy faces threats not only from institutions but increasingly from peers, heightening the prominence of peer privacy issues (Dinev et al., 2013). Research investigates both antecedents and consequences: Ozdemir et al. (2017) used the APCO framework to show how privacy experiences and awareness intensify peer concerns, reducing disclosure willingness on social media. Similarly, Nguyen et al. (2024) proposed a dual peer privacy calculus model linking these concerns to active and passive social networking site use.

While existing research has explored various factors influencing short video users’ privacy disclosure, it predominantly relies on privacy calculus theory to explain the privacy paradox. Studies on user privacy concerns have largely focused on institutional (platform-level) concerns, overlooking the impact of peer privacy concerns. Although some work suggests peer concerns significantly reduce disclosure on social media, these findings stem from traditional platforms and fail to address the unique privacy implications of visual content and algorithmic recommendations inherent to short video platforms. Furthermore, current research inadequately examines the interaction between these dual concerns—specifically, how institutional concerns may moderate the effect of peer risks on disclosure, a relationship needing careful analysis within short video contexts. To bridge these gaps, this study introduces a dual privacy concerns framework and analyzes short video users’ disclosure behaviors through CPM, focusing on boundary rules, turbulence, and coordination. This approach aims to uncover the mechanisms governing privacy concerns in short video disclosure contexts.

2.2 Users’ privacy disclosure behaviors

User privacy disclosure behavior refers to the behavior of users who voluntarily disclose personal information such as names, preferences, and demographics to Internet platforms (Wang et al., 2021). The conflict between the necessity of privacy disclosure and the risk of privacy leakage is a key factor affecting users’ use or continued use of Internet platforms (Hong and Thong, 2013). The rise of short video platforms has introduced novel privacy challenges. The visual nature of their content format (e.g., exposure of facial features and geolocation data) may amplify risks of sensitive information leakage (Kang et al., 2022), while algorithm-driven precision recommendation mechanisms further blur privacy boundaries.

Most of the related studies are empirical analyses of theoretical model construction based on the theories of privacy calculus theory (Dinev and Hart, 2006), the theory of planned behavior (Xu et al., 2013), social exchange theory (Liu et al., 2016), social exchange theory (Loiacono, 2015), uses and gratifications theory (Hollenbaugh and Ferris, 2014), and social influence process (Min and Kim, 2015). In addition, many scholars have explored the relationship between privacy concerns and privacy disclosure behavior. However, the conclusions of many studies are inconsistent. For example, Nemec Zlatolas et al. (2015) and Wang J. et al. (2024) found that privacy concerns are an important factor influencing users’ self-disclosure behaviors, while Krasnova et al. (2010) found that there is no significant relationship between privacy concerns and privacy disclosure behaviors in their study. Therefore, the relationship between privacy concerns and users’ privacy disclosure behavior still needs to be further explored. Several scholars have investigated the influencing factors of privacy disclosure behaviors among short video users. For instance, Zhu et al. (2022) examined how interpersonal and human-computer interactions affect users’ privacy disclosure intentions through perceived risks and perceived benefits from an interactive perspective. Li et al. (2023) analyzed the impact of user behavior and benefits obtained on privacy concerns and disclosure behavior through user data. Kang et al. (2022) explored the impact of parental intervention on adolescents’ privacy disclosure behaviors and risk perceptions in short video contexts.

Esmaeilzadeh (2024) and Neves et al. (2022) studied privacy disclosure in a dual privacy context. But they focused on health information and the social networking sites. How short video users disclose personal information under dual privacy concerns remains unclear. Furthermore, existing studies have inadequately addressed the interaction between these dual concerns, particularly how institutional privacy concerns may moderate the effect of peer risks on disclosure behaviors (Neves et al., 2022).

2.3 Communication privacy management theory

In 1991, Petronio proposed the Theory of Communication Boundary Management (CBM) and applied it to privacy management between couples (West and Turner, 2014). In 2002, Petronio further elaborated on this theory and formally renamed it the Theory of Communication Privacy Management (CPM). CPM is used to explain the negotiation process individuals undergo when disclosing information and determining whether access to that information poses potential risks (Petronio, 2002). The theory suggests that individuals maintain and coordinate privacy boundaries with various communicating parties by establishing these boundaries based on the delineation between personal and public information. Consequently, individuals form a region of information with well-defined boundaries, determining what information is private. Any attempt to breach these boundaries may be perceived as a threat to privacy, giving rise to privacy concerns (Xu et al., 2011). Meanwhile, when people want or need to disclose private information, they make judgments about to whom they decide to open their privacy boundaries in order to achieve their goal of revealing or concealing their private information (Smith and Brunner, 2017).

Communication Privacy Management theory conceptualizes privacy management through three interdependent elements: boundary rules, coordination, and turbulence (Xu et al., 2011). These elements capture the core processes and challenges individuals face in managing private information. Boundary rules govern when and how individuals disclose personal information (De Wolf et al., 2014). Boundary coordination occurs when private information is shared, involving negotiated rule-setting between parties (e.g., trading privacy control for compensation; Metzger, 2007). Boundary turbulence arises when rules are ambiguous or coordination fails, prompting behavioral adjustments (West and Turner, 2014). CPM is now widely applied to digital privacy contexts, including: Enterprise social networks (Wang Y. et al., 2024), Hotel social robots (Jia et al., 2024), Chatbots (Liu et al., 2024). For instance, Liu et al. (2022) used CPM to explore consumer vulnerability on shared lodging platforms, while Wang et al. (2019) developed a CPM-based model clarifying how privacy violations amplify concerns.

The process of users utilizing short videos constitutes a process of boundary management between users and platforms. In the context of short video platforms, boundary rules refer to users’ self-established conditions for information disclosure (e.g., “not sharing geolocation data”); boundary coordination involves negotiating rule implementation with either the platform or peers (e.g., accepting privacy policies in exchange for service access); and boundary turbulence triggers remediation mechanisms when rules are violated (e.g., adjusting permission settings). Short video users employ the CPM to implement dual privacy boundary management. Regarding peer privacy boundaries, users dynamically adjust their sharing strategies based on feedback from fan interactions. For institutional privacy boundaries, users determine their information disclosure levels according to the transparency of platform privacy policies and the trustworthiness of technological safeguards. Platforms engage in boundary coordination with users through privacy agreements, while technical vulnerabilities or ambiguous policies may trigger boundary turbulence (Xu et al., 2011; Petronio, 2002).

3 Constructs and hypotheses in the model

This paper analyzes the factors influencing the privacy disclosure behavior of short video users from the perspective of institutional privacy concerns and peer privacy concerns, based on CPM. This study investigates the current status of short video users’ usage behavior, refines the factors affecting their privacy concerns, and combines the theoretical foundation of this paper to propose the antecedent variables of their institutional privacy concerns and peer privacy concerns. It constructs a model of the factors affecting short video users’ privacy disclosure behavior, as shown in Figure 1.

Figure 1
Flowchart illustrating relationships between factors affecting privacy concerns and disclosure behavior. Factors include perceived peer risk, information sensitivity, effectiveness of privacy protection technology, and policy. Arrows indicate hypothesized effects, labeled H1+ through H10+, affecting peer and institutional privacy concerns, influencing privacy disclosure behavior, considering age, gender, and education.

Figure 1. Research model.

3.1 Perceived peer risk

Perceived peer risk refers to an individual’s subjective assessment of the potential privacy violations by online peers (Ozdemir et al., 2017). Grounded in Communication Privacy Management (CPM) theory, unauthorized information sharing constitutes a significant privacy threat (Petronio, 2002). Empirical studies demonstrate that while individuals exhibit higher trust toward familiar groups (Foddy et al., 2009), such trust diminishes when peer networks expand and risk perception emerges, subsequently triggering defensive behaviors. Although short-video platforms provide privacy settings for information access control, user data remain vulnerable to secondary dissemination by peers (Nemec Zlatolas et al., 2019). The platforms’ recommendation algorithms, which synthesize multi-dimensional data including contact lists, social connections, and geolocation information, often lead to private information leakage into public domains. When user-disclosed content is redistributed or utilized by online peers without consent, it violates users’ predetermined privacy boundary rules, thereby inducing privacy boundary turbulence. Therefore:

H1: Perceived peer risk positively affects users' peer privacy concerns.

3.2 Information sensitivity

Information sensitivity refers to the degree of psychological discomfort individuals experience when disclosing specific types of information (Dinev et al., 2013), constituting a fundamental determinant in establishing privacy boundary rules (Malhotra et al., 2004). Grounded in CPM, users construct virtual boundaries to regulate access to personal information, wherein information sensitivity governs their disclosure scope to both platforms and peers (Kolotylo-Kulkarni et al., 2021). Empirical studies consistently demonstrate users’ propensity to avoid disclosing sensitive information, with platforms’ requests for highly sensitive data frequently resulting in user opt-out behaviors (Malhotra et al., 2004). When handling highly sensitive information, short-video users exhibit heightened vigilance toward potential rights infringements stemming from privacy breaches (Yang et al., 2008). This heightened risk perception engenders privacy concerns through two distinct mechanisms: (1) peer privacy concerns emerge through social risk diffusion processes inherent in short-video platforms’ design. The platforms’ social visibility features create vulnerability to secondary data exposure, where sensitive information may be unexpectedly disseminated by peers through resharing or screenshotting, thereby inducing relational anxiety and boundary turbulence. (2) Institutional privacy concerns develop through technological trust erosion when platforms request information exceeding users’ sensitivity thresholds or fail to provide adequate transparency regarding data usage. Such practices foster perceptions of control asymmetry and trust deficit, leading users to question both the legitimacy of platforms’ data processing frameworks and the ethical boundaries of algorithmic data extraction and commercial exploitation (Maseeh et al., 2023). These dual pathways ultimately shape users’ comprehensive assessment of privacy threats in the short-video ecosystem. Therefore:

H2: Information sensitivity positively affects users' peer privacy concerns.

H3: Information sensitivity positively affects users' institutional privacy concerns.

3.3 The effectiveness of privacy protection technology

Privacy protection technology effectiveness reflects users’ perceived ability of platform safeguards to secure personal information (Wang et al., 2019). Within Communication Privacy Management (CPM) theory, these technologies serve as boundary management tools. Research confirms they significantly reduce inappropriate data access (Popp and Poindexter, 2006), while evolving standards enable personalized privacy control in mobile contexts (Duan and Wang, 2015). Although most major short-video platforms hold ISO27001 and ISO27701 certifications—indicating mature safeguards—breaches still occur through hacker attacks, eroding user trust. When protection efficacy appears insufficient, users’ concerns about institutional data misuse amplify (Maseeh et al., 2023), intensifying institutional privacy concerns.

Furthermore, effective privacy protection technologies affect not only the user-platform boundary but also privacy management between users and peers. Platform safeguards—like access controls, data encryption, screenshot blocking, and restricted resharing tools—directly limit peers’ ability to access or distribute sensitive user data without authorization. When users perceive robust technical measures preventing peer misuse (e.g., unauthorized forwarding, downloading, or screenshots), their concerns about peer boundary violations and resulting turbulence (i.e., peer privacy concerns) decrease significantly. Ultimately, technological efficacy strengthens users’ perceived control over peer information flows, preventing perceived risks from escalating into substantive peer privacy concerns. Therefore:

H4: The effectiveness of privacy protection technology negatively affects peer privacy concerns.

H5: The effectiveness of privacy protection technology negatively affects institutional privacy concerns.

3.4 The effectiveness of privacy policy

The effectiveness of privacy policy refers to the extent to which users perceive that the published privacy policy actually provides accurate and reliable information about the operator’s information practices (Xu et al., 2011). It is a process of boundary coordination that occurs through the collective control of users’ personal information, furthermore, credible and comprehensive privacy policies help platforms build a positive image among users. In addition, providing a detailed privacy policy is positively associated with users’ trust in the operator (Wu et al., 2012), which mitigates the perceived privacy risk. When users register and log in, links to privacy protection policies—often referred to as “Privacy Policy” and “User Service Agreement”—appear. These links serve to inform users about their rights and how their personal information will be gathered, used, stored, and shared while using the platform. Through the “Privacy Settings” feature, users of short videos can limit the amount of information that is stored and disclosed; however, platforms may obtain unauthorized access to users’ information in order to increase profits, which also puts users’ security and privacy at risk. Users’ institutional privacy concerns may rise when they believe short video services’ privacy policies are less effective.

Clear, transparent, and enforceable privacy policies—particularly those explicitly regulating how user data (especially user-generated content) may be utilized, shared, and disseminated among peers—play a pivotal role in managing privacy boundaries between users. For instance, policies that expressly prohibit unauthorized secondary dissemination (e.g., resharing, downloading, screenshot distribution) without users’ explicit consent, or that clearly delineate users’ control rights over content circulation (such as establishing the legal basis for “followers-only visibility” or “download prohibition” settings alongside platform enforcement commitments), can significantly reduce users’ anticipated risks of peer information misuse. When users trust that platforms will effectively constrain peer behaviors through policy enforcement measures (e.g., penalizing violative accounts, providing accessible infringement reporting mechanisms)—thereby upholding their established privacy boundary rules—their concerns about potential peer-induced boundary turbulence (i.e., peer privacy concerns) are correspondingly mitigated.

H6: The effectiveness of privacy policy negatively affects peer privacy concerns.

H7: The effectiveness of privacy policy negatively affects institutional privacy concerns.

3.5 Peer privacy concerns

Peer privacy concerns stem from worries about potential misuse of personal data shared with online peers, as highlighted by Raynes-Goldie (2010). Specifically, peers might disseminate private communications-including images, texts, emotions, and insights-meant for a restricted audience on social media. Similar to other privacy worries, intensified concerns over peers infringing on one’s privacy boundaries lead users to adopt precautionary measures, like refraining from using social networks or minimizing their exposure (Ozdemir et al., 2017). Consequently, users worried about peers violating their privacy tend to adjust their interactions, seeking to balance communication while reducing privacy risks to tolerable levels. Users with peer privacy concerns actively manage their privacy disclosures, often by limiting peer interactions, to maintain an acceptable level of privacy while using internet platforms (Neves et al., 2022). When users share short videos, this information reaches their peers, posing a potentially greater threat to user privacy (Ozdemir et al., 2017). Online peers could recognize individuals and objects in short videos, pinpoint users’ locations and activities, or reuse posted content without permission. Therefore:

H8: Peer privacy concerns negatively affect users’ privacy disclosure behavior.

3.6 Institutional privacy concerns

Another significant factor that motivates users to lessen their privacy disclosure behavior is institutional privacy concerns, which is defined as the degree of users’ anxiety regarding the exploitation of their personal information by platforms (Neves et al., 2022). According to CPM, users establish institutional trust boundaries through boundary rules when using short video platforms, explicitly defining expectations regarding “what information can be collected” and “how such information may be utilized.” When users perceive institutional behaviors transgressing their established privacy boundaries, boundary turbulence is triggered. To mitigate boundary violation risks, users strategically reduce privacy disclosures to achieve equilibrium, thereby diminishing opportunities for institutional information misuse (Zhang et al., 2022). This equilibrium enables continued platform engagement while allowing users to control or reduce their risk exposure (Neves et al., 2022). Existing research has demonstrated that institutional data collection processes may elicit individual privacy concerns (Maseeh et al., 2023). Therefore:

H9: Institutional privacy concerns negatively affect user privacy disclosure behavior.

Institutional privacy concerns not only directly influence disclosure behaviors but may also indirectly amplify users’ assessment of peer privacy risks by eroding their trust in the platform’s overall privacy protection capabilities. CPM posits that privacy boundary management constitutes a dynamic system where institutional and peer privacy boundaries exhibit synergistic effects (Petronio, 2002). When users experience heightened institutional privacy concerns, they may perceive the platform as incompetent in constraining peer behaviors (e.g., preventing unauthorized resharing or screenshot misuse). This systemic trust deficit lowers users’ tolerance threshold for peer privacy risks—even when the actual perceived peer risk remains unchanged, their pessimistic expectations about platform safeguards intensify peer privacy concerns.

H10: Institutional privacy concerns positively affect peer privacy concerns.

4 Methods

4.1 Resource identification initiative

TikTok is selected as the platform on which to test our model because it is one of the most widely used short video platforms in China. Up until March of 2021, the user market shares for TikTok was 59.88% which far surpassed those of short video applications. We developed an English questionnaire from the literature and then translated it into Chinese. Before distributing the questionnaire on a large scale, 10 heavy users of TikTok were first interviewed. Interviews yielded core findings. Users expressed worries that information shared with peers might be misinterpreted or misused (e.g., personal preferences taken out of context), which helped refine PPC measurements to fit short video scenarios. They distinguished between highly sensitive data (e.g., bank information) and low sensitivity data (e.g., username), aiding adjustments to IS items based on existing scales. Mentions of encryption tools’ practical impact on daily use guided wording tweaks for PPT items. The questionnaire items, originally referenced from established scales, were revised via these interviews to better align with the research context. After modification, 30 valid questionnaires were distributed on a small scale and recovered. Pre-test results showed more details. Items for PP needed specific adjustments (e.g., adding references to short video platform features). Responses clarified that PPR should focus on peers’ improper information handling, prompting wording refinements. Ambiguity in distinguishing platform vs. regulatory concerns in IPC items (based on prior scales) led to context-specific clarifications. Finally, the questionnaire items were modified and improved again according to the pre-test results. The measurement scale was finally generated (Table 1). All these constructs were measured using 1–5 Likert scale (1 = “strongly disagree,” 5 = “strongly agree”).

Table 1
www.frontiersin.org

Table 1. Measurement instrument.

4.2 Sample and data collection

Our data were collected through wjx.cn, an online questionnaire collection platform. At the beginning of the questionnaire, we set a screening question asking the respondents whether they had any experience of using TikTok as users. Only those who chose “yes” could continue filling out the questionnaire. The sampling method used was convenience sampling. We targeted individuals who had prior experience with TikTok. This approach allowed us to efficiently gather responses from a relevant sample.

A total of 360 questionnaires were collected in this study, and after excluding invalid questionnaires, 302 valid questionnaires were finally collected. The statistical results show that the proportion of men and women is relatively balanced, and the proportion of female respondents is slightly higher than that of men; the age of 18–45 years old, the number of samples with education in the specialty and below is larger, which is basically in line with the status quo of the 20–39 years old group and the group of college-educated with the highest rate of short video usage, so the survey samples have a high degree of representativeness, and the results of the descriptive statistics of the surveyed respondents are shown in Table 2.

Table 2
www.frontiersin.org

Table 2. Demographic statistics.

5 Analysis and results

In this study, SmartPLS3.0 was used for structural equation modeling and to validate the relationships between these variables. The reasons for choosing the PLS-SEM method to analyze sample data are as follows: (1) PLS is suitable for testing models with many constructs and complex relationships, and it can avoid infeasible solutions and factor uncertainty. (2) PLS has relatively low requirements for the sample size. In this study, the sample size is 302, which is more than 10 times the total number of measurement items (Hair et al., 2012). Additionally, control variables such as gender, age, and education level were incorporated into the PLS-SEM analysis. This addresses potential moderating effects. The impact is on relationships between PPC, IPC, and PDB. These control variables help eliminate potential confounding biases in the observed relationships.

5.1 Evaluation of measurement model

We analyse construct reliability and discriminant validity, by assessing the lowest indicator loading, the crossing loading, the average variance extracted (AVE), the heterotrait-monotrait ratio (HTMT), and the Fornell–Larcker criterion. Table 3 presents descriptive statistics. It also reports results of the Fornell–Larcker criterion. Table 4 shows discriminant validity. This is based on the cross-loading criterion. Table 5 provides values for Cronbach’s Alpha. It includes the lowest indicator loading. It also reports AVE and CR. Table 6 reports discriminant validity based on the HTMT criterion. To summarise, all Cronbach’s Alpha values and CR for all constructs are higher than 0.7, ensuring indicator reliability and convergent validity. All the AVE are higher than 0.5. All HTMT criterion scores but the relationship between PPC and IPC are below the threshold of 0.9, and the square root of the AVE of all constructs is higher than any correlation with another construct (Fornell-Larker criterion), indicating the presence of discriminant validity.

Table 3
www.frontiersin.org

Table 3. Descriptive statistics and Fornell–Larcker criterion for model.

Table 4
www.frontiersin.org

Table 4. Cross loadings, VIF, and method loadings for model.

Table 5
www.frontiersin.org

Table 5. Cronbach’s Alpha, lowest indicator loading, AVE, CR.

Table 6
www.frontiersin.org

Table 6. Discriminant validity based on HTMT criterion for model.

5.2 Common method bias

To address potential common method bias, multiple strategies were employed in this study. First, at the beginning of the questionnaire, respondents were explicitly informed that all responses would be kept anonymous and confidential, with clear instructions encouraging honest feedback to minimize response bias. Second, Harman’s single-factor test was conducted through exploratory factor analysis; the results showed that the first unrotated factor explained 44.77% of the total variance, which is below the 50% threshold (Podsakoff and Organ, 1986). Third, a common latent factor was incorporated into the measurement model. All items were converted into single-item constructs, and the main constructs were treated as second-order constructs. The results revealed that most loadings of the method factor were either non-significant, small, or negative (Table 4). Therefore, common method bias did not systematically interfere with the measurements.

5.3 Evaluation of structural model

The SRMR for model is 0.06, which is below the threshold of 0.08. The NFI is 0.831. Although it is slightly below the 0.9 threshold, it is close to this value. This suggests that the model exhibits good fit. The model passed the most of hypothesis tests, see Figure 2 and Table 7. Peer privacy concerns and institutional privacy concerns have a significant negative effect on users’ privacy disclosure behavior. The effects of variables affecting institutional privacy concerns are as follows: information sensitivity has a significant positive effect, the effectiveness of privacy protection technology and the effectiveness of privacy protection policies has a significant negative effect. The effects of variables affecting peer privacy concerns are as follows: perceived peer risk has a significant positive effect, information sensitivity has a significant positive effect. In addition, institutional privacy concerns positively influence peer privacy concerns.

Figure 2
Diagram illustrating relationships between factors influencing privacy concerns and disclosure behavior. Arrows show the impact of perceived peer risk, information sensitivity, and effectiveness of privacy measures on peer and institutional privacy concerns, which affect privacy disclosure behavior. Influences are quantified by coefficients such as 0.157***, 0.240***, and others. Demographic factors like age, gender, and education directly affect privacy disclosure behavior.

Figure 2. Model results. ***p < 0.001, **p < 0.01, *p < 0.05.

Table 7
www.frontiersin.org

Table 7. Summary of hypothesis test results.

Table 8 presents the results of total indirect effects and specific indirect effects. The total indirect effect analysis shows that PPR and IS have significant negative total indirect effects on PDB. PPT and PP have significant positive total indirect effects on PDB. For specific indirect effects, PPR has a significant negative indirect effect on PDB through PPC. IS has significant negative indirect effects on PDB through both PPC and IPC. PPT and PP each have significant positive indirect effects on PDB through IPC.

Table 8
www.frontiersin.org

Table 8. Summary of mediation analysis test results.

Among the control variables, gender was found to exert a significant negative effect on PPC (β = −0.079, p = 0.016). In contrast, gender, age, and education level showed no significant impacts on IPC and PDB. Additionally, age and education level had no significant influence on PPC.

6 Discussion

6.1 Perceived peer risk positively influences peer privacy concerns

Perceived peer risk positively and significantly affects peer privacy concerns and indirectly affects users’ privacy disclosure behavior through peer privacy concerns. This finding is consistent with the findings of Ozdemir et al. (2017) and Quan-Haase and Ho (2020). Higher peer risk increases the level of concern about peer threats, and the user’s perceived privacy risk increases further, which in turn leads to higher peer privacy concerns and ultimately affects their privacy disclosure behavior. It has also been shown that although most students are aware of the possible consequences of providing personally identifiable information on Facebook, they are not concerned about it (Govani and Pashley, 2007), which is inconsistent with the findings of this paper. Peer privacy concerns are correspondingly lower when the perceived benefits of online peers outweigh the perceived risks, such benefits may be self-presentation fulfillment, the need to maintain realistic social relationships, and higher online status.

Users exhibit varying levels of concern regarding different types of online peer privacy threats. In the context of short videos, users can categorize peers based on perceived risk, which in turn influences their decisions on content visibility. Selective posting according to peer group classification narrows the scope of information dissemination in short videos, fostering the formation of tighter, interest-based communities. Short videos may be followed by peers in posting and dissemination to form groups or communities centered on blood, geography, or interest, etc. Grouping peers can allow information to be disseminated in a smaller, more reliable circle. Platforms like WeChat allow users to specify who can see their posts, ranging from everyone to no one, thus enabling users to manage their privacy more effectively. When posting a short video, users can select options such as “Publicly visible,” “Not to be seen by anyone,” “Visible to friends,” “Partially visible,” and “Visible only to myself” to restrict the access rights of their peers. Short video platforms should enhance their grouping functionalities to allow for more granular control over content visibility based on factors such as geography, age, and gender. While users delegate some privacy rights to platforms, it is imperative for these platforms to safeguard user privacy by limiting the potential for data misuse and ensuring responsible handling of personal information.

6.2 Information sensitivity positively affects peer privacy concerns, institutional privacy concerns

Information sensitivity positively and significantly affects peer privacy concerns and institutional privacy concerns and indirectly affects users’ privacy disclosure behavior through peer privacy concerns and institutional privacy concerns. This finding is consistent with the findings of Malhotra et al. (2004) and Xu et al. (2013) that information sensitivity affects privacy concerns and that users will be more reluctant to disclose sensitive information compared to less sensitive information. When users perceive the requested information as sensitive, their perceived risk increases, which in turn affects privacy concerns and privacy disclosure behavior.

In general, the more personal information involved and the more sensitive it is, the higher the user’s concern about its privacy leakage, and the higher the user’s institutional privacy concern and peer privacy concern. From the user’s perspective, the personal information involved in short video platforms, such as identity card number, bank card information, home address, work information, etc., are all sensitive information, of which identity card and bank card information are the most sensitive to the user, and many users will not take the initiative to carry out real-name authentication when using short videos, and they will fill in the form only when they have no other choice, such as purchasing commodities in live broadcasting rooms or receiving red packets for cash withdrawal, and will fill in the Personal identification information. Conversely, many users tend to overlook the sensitivity of requests for access to device storage, address books, or geographical locations, often consenting without hesitation. In 2015, the U.S. government proposed the Consumer Privacy Bill of Rights, defining sensitive consumer information as data linked to an individual or their commonly used device within specific contexts. This draft legislation aims to enhance consumer control over personal information, increase operators’ liability for misuse, and strengthen penalties for violations. Therefore, short video platforms should strictly comply with relevant laws and regulations, handle users’ sensitive personal information in accordance with relevant regulations, and provide prompts to users when obtaining sensitive information, explaining the importance of the information and protecting users’ privacy.

6.3 The effectiveness of privacy protection technology and privacy policy negatively affects institutional privacy concerns

Both privacy protection technology and policy effectiveness significantly reduce institutional privacy concerns, indirectly influencing users’ disclosure behaviors through this pathway. However, neither significantly impacts peer privacy concerns. Supporting this, Wang and Herrando (2019) found greater technological efficacy lowers perceived information intrusion. Similarly, Xu et al. (2011) demonstrated effective policies enhance perceived privacy control and reduce perceived risk—aligning with our findings. Collectively, effective technology and policy help regulate user-platform boundary turbulence, facilitate boundary management, and lower perceived privacy risks (Wang et al., 2019). Crucially, peer-related information privacy concerns differ from platform-centric concerns (Zhang et al., 2022). Peer abuse risk depends heavily on social relationships. In high-uncertainty social environments, peer privacy concerns thus center on relationship governance costs (e.g., monitoring fan behavior) rather than platform trust. Privacy technologies and policies primarily govern the user-platform boundary, whereas peer privacy turbulence stems from trust deficits within social relationships. This necessitates relational governance mechanisms—such as layered visibility controls—rather than direct technological or policy solutions.

However, users do not understand the privacy protection technology and privacy protection policy of the platform in the process of using short videos, and may even question it. Therefore, the platform should continuously improve the effectiveness of privacy protection technology and privacy policy. Platforms should continuously upgrade their privacy protection technology to reduce the probability of user information leakage and build the first line of defense with effective privacy protection technology. At the same time, platforms often claim that they use advanced technology to protect users’ privacy, but users do not know which technology is used and how that technology works. Therefore, platforms can use plain and simple language to explain the mechanics of privacy-protecting technologies. In addition, the responsibilities of both parties in ensuring that personal data is protected should be clearly set out in a privacy policy. While privacy policies are obviously placed on the platform’s homepage, privacy policies in short videos are usually presented in the form of links and contain specialized legal terms. Users can simply click “I agree to the terms and conditions” and skip reading the privacy policy. In this regard, platforms should use various forms of digital media (e.g., using images, videos, and animations) to attract users’ attention, promote them to read the full privacy policy (Wang et al., 2019), and popularize the content of the privacy policy.

6.4 Peer privacy concerns and institutional privacy concerns negatively influence user privacy disclosure behavior

Both peer and institutional privacy concerns negatively affect users’ privacy disclosure behavior on short-form video platforms. Our finding that heightened institutional privacy concerns reduce disclosure aligns with prior research: users with strong institutional concerns typically disclose less personal information and do so less actively, engagedly, and promptly (Zhang et al., 2022). Similarly, the impact of peer privacy concerns on disclosure behavior is consistent with earlier peer concern studies (Ozdemir et al., 2017). However, this contrasts with Govani and Pashley's (2007) finding that privacy concerns do not significantly affect disclosure—a result often attributed to the privacy paradox. This discrepancy may stem from sample characteristics (student cohorts exhibiting higher risk tolerance) or the amplification of privacy concerns by the strong social attributes inherent to short-video platforms.

Peer and institutional privacy concerns hinge on users’ awareness of privacy issues, highlighting the necessity for improved privacy literacy. Privacy literacy development is facilitated not only by educational institutions and governments but also by short video platforms. These platforms should provide engaging content that enhances users’ understanding of privacy risks and motivates them to protect their data. Moreover, research indicates that smart, personalized advertisements may intensify privacy concerns, potentially causing users to shun recommended ads on social media (Jung, 2017). To mitigate this, short video platforms should employ big data analytics judiciously to obscure user profiles and restrict access to sensitive information, ensuring users’ data is used responsibly. Additionally, analyzing data such as user interactions and video content can help platforms discern users’ preferences for privacy and their tendencies to limit data sharing. Consequently, platforms must embed privacy protection measures throughout all phases of research, development, design, and operations, fostering a secure environment that encourages continued engagement and content creation.

6.5 Institutional privacy concerns positively influence peer privacy concerns

The institutional privacy concerns had a positive and significant effect on peer privacy concerns. This is consistent with the findings of Neves et al. (2022) and Liu et al. (2022): when users perceive a high likelihood of institutional data leakage, it also affects their judgment of peer privacy concerns. This discovery reveals a dynamic link between dual privacy concerns in short video contexts: users’ negative assessments of platform data management (institutional privacy concerns) generalize to reevaluate peer behavioral risks, amplifying peer privacy concerns. These results strongly support the boundary synergy perspective in CPM (Petronio, 2002), indicating that institutional and peer privacy boundaries are interdependent and interact via platform-facilitated systemic trust linkages. Moreover, lower institutional privacy concerns—reflecting greater trust in platforms’ protective capacity—reduce users’ defensive vigilance against technological power and mitigate peer interaction risk expectations, ultimately enhancing information disclosure intentions.

Thus, platforms must disrupt the dual concern transmission chain through integrated technology and policy design. Proactive intervention technologies can target peer abuse risks. Forwarding watermarking (e.g., dynamic ID watermarks, non-removable on-screen floats) embeds tracking identifiers in user content, enhancing perceived controllability over dissemination (alleviating peer concerns) while demonstrating platform capabilities to prevent data leaks through technical visibility (mitigating institutional concerns). For real-time sensitive information masking, AI can identify critical elements (e.g., documents, house numbers) and trigger automatic blurring/coding, reducing potential peer misuse at the source. Concurrently, privacy policies must clearly delineate boundaries between “institutional data management” and “peer interaction rules.” Short-video platforms could implement a “graded visibility policy” with refined fan data usage terms (e.g., “fan-visible content prohibits downloading/forwarding,” “intimate circle data excluded from algorithmic recommendations”). This rule transparency reduces users’ anticipated risk of peer boundary violations (curbing peer concerns).

7 Conclusion

7.1 Theoretical implications

With above findings, this study has significant theoretical implications for research on user privacy disclosure behavior and privacy concern. Firstly, this study enriches the privacy concern literature by including the two facets of privacy concerns (i.e., institutional and peer privacy concerns) into the research on disclosure behavior in information systems. Despite the considerable academic attention received by Internet privacy concerns since their emergence, the majority of studies have primarily centered on controlling personal information disclosure and online engagements with service providers (Dinev and Hart, 2006; Dinev et al., 2013; De Wolf et al., 2014; Xu et al., 2011), neglecting the exploration of peer privacy concerns. Indeed, following a thorough literature review, Zhang et al. (2022) identified this as a prevalent trend in Internet privacy research and advocated for further investigation into peer privacy concerns. In response to their researches, this study not only empirically examines the antecedents of users’ privacy concerns but also establishes users’ institutional and peer privacy concerns as mediators influencing their privacy disclosure behavior. Thus, we contend that this study contributes significantly to a more holistic understanding of user behavior on short-form video platforms. Such contexts are abundant owing to the proliferation of social interaction elements, particularly contemporary technological advancements. Considering that individuals frequently need to share personal information with others (e.g., on online dating platforms, social media, and multiplayer gaming environments), social factors and interactions can potentially evoke peer privacy concerns. Our research underscores the significance of taking into account user privacy risks stemming from peers, as opposed to solely concentrating on threats from website operators.

Second, an evident progression of this study would be to examine privacy concerns as a mediating factor within a broader nomological framework, wherein privacy concerns catalyze behaviors aimed at minimizing privacy disclosure, a process potentially influenced by institutional and peer-related factors. Indeed, several scholars have already explored aspects of this extended model (Bansal et al., 2015; Jozani et al., 2020; Neves et al., 2022; Ozdemir et al., 2017; Wang et al., 2019; Xu et al., 2011). Nevertheless, we posit that a more expansive and holistic model, encompassing not just the precursors and outcomes of privacy concerns, but also explores the dependencies arising from them, would constitute a significant contribution to scholarly literature. Our preliminary findings suggest that both institutional practices and individuals’ perceptions of their peers’ privacy norms induce individuals’ privacy concerns, which in turn reduce users’ privacy disclosure behavior. However, in platforms with social attributes, there may be an interaction between individuals’ perceptions of institutions and individuals’ perceptions of their peers’ behaviors that is complex beyond the threshold of existing theories, and future research is necessary to delve deeper into the interconnections and synergies between these two aspects.

7.2 Practical implications

For internet users, privacy may seem psychologically distant and could potentially negatively impact their decisions (Bandara and Perera, 2020). While previous research has primarily focused on the influence of institutional privacy on disclosure behaviors, the threat of peer privacy, which could potentially cause more concern for users (Ozdemir et al., 2017), needs to be considered in the context of users’ self-disclosure behaviors. Our study highlights the significance of institutional and peer privacy concerns in deterring users’ privacy disclosure tendencies on short video platforms, thereby urging platform managers to prioritize the administration of user data and address both institutional and peer privacy concerns. More specifically, it is advocated that managers must safeguard users’ privacy not merely from the short video platforms and associated third-party parties, but also from users’ peers, encompassing individual creators, fans, and visitors. To relieve peer privacy concerns, managers should prioritize enhancing their platforms’ data management systems, stringently restricting access to users’ personal information. Concrete actions could involve refraining from sharing users’ identification data with advertisers, fully empowering users to determine the visibility of their profiles, and anonymizing user reviews, among other measures.

Based on the research findings, this study proposes the following practical countermeasures: (1) Regulatory authorities should strengthen privacy regulations and oversight mechanisms by refining compliance boundaries for data collection, storage, and use in video privacy policies; establishing third-party audits of platform privacy technologies with penalties (e.g., fines, blacklisting); and collaborating with platforms to develop micro-courses enhancing users’ privacy risk identification. (2) Short-video platforms must optimize privacy safeguards through real-time blurring of sensitive video content, restructuring policies with transparent clause summaries, and enhancing social functions allowing follower content restrictions like download prohibitions and forwarding authorizations. (3) Users should proactively manage privacy boundaries by reviewing settings regularly, applying platform blurring tools to highly sensitive content (e.g., family scenes, IDs), and immediately reporting suspected data violations.

7.3 Limitations and future research

The conclusions of this study are mainly applicable to algorithm-driven short-video platforms represented by Douyin/TikTok, which indicates that the research has certain limitations. Primarily, it focused exclusively on TikTok users and did not consider those on other short video platforms, where findings might not be universally applicable. Future research should verify the applicability of the model to live-streaming e-commerce platforms (e.g., Taobao Short Videos) and vertical community platforms (e.g., Bilibili), so as to assess the generalizability of the research results across different contexts and user groups.

Additionally, our study did not account for variations in users’ duration of platform use. Longstanding users might exhibit different privacy concerns than newcomers, potentially influencing their privacy disclosure behaviors. Future research should investigate how the duration and frequency of platform use affect privacy concerns and disclosure behaviors, thereby offering a more nuanced understanding of user dynamics over time.

Data availability statement

The data that support the findings of this study are openly available in Science Data Bank at https://doi.org/10.57760/sciencedb.13753.

Ethics statement

The studies involving humans were approved by Department of Journalism and Communication at Shaanxi Normal University. The studies were conducted in accordance with the local legislation and institutional requirements. Written informed consent for participation in this study was provided by the participants’ legal guardians/next of kin.

Author contributions

HL: Conceptualization, Writing – original draft, Writing – review & editing. AW: Methodology, Writing – original draft, Writing – review & editing. XL: Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This work was supported by National Social Science Fund of China for young scholars, project approval number 24CTQ016.

Acknowledgments

We would like to thank our senior editor, associate editor, and reviewers for their constructive feedback throughout the review process.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The authors declare that Gen AI was used in the creation of this manuscript. Generative AI was used for language polishing and translation assistance.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Bandara, R., and Perera, I. (2020). “Unit test code generation tool support for lower level programming languages” In 2020 Moratuwa Engineering Research Conference (MERCon), Moratuwa, Sri Lanka: IEEE. 1–6. doi: 10.1109/MERCon50084.2020.9185378

Crossref Full Text | Google Scholar

Bansal, G., Zahedi, F. M., and Gefen, D. (2015). The role of privacy assurance mechanisms in building trust and the moderating role of privacy concern. Eur. J. Inf. Syst. 24, 624–644. doi: 10.1057/ejis.2014.41

Crossref Full Text | Google Scholar

De Wolf, R., Willaert, K., and Pierson, J. (2014). Managing privacy boundaries together: exploring individual and group privacy management strategies in Facebook. Comput. Hum. Behav. 35, 444–454. doi: 10.1016/j.chb.2014.03.010

Crossref Full Text | Google Scholar

Dinev, T., and Hart, P. (2006). An extended privacy calculus model for e-commerce transactions. Inform. Syst. Res. 17, 61–80. doi: 10.1287/isre.1060.0080

Crossref Full Text | Google Scholar

Dinev, T., Xu, H., Smith, J. H., and Hart, P. (2013). Information privacy and correlates: an empirical attempt to bridge and distinguish privacy-related concepts. Eur. J. Inf. Syst. 22, 295–316. doi: 10.1057/ejis.2012.23

Crossref Full Text | Google Scholar

Duan, X., and Wang, X. (2015). Authentication handover and privacy protection in 5G hetnets using software-defined networking. IEEE Commun. Mag. 53, 28–35. doi: 10.1109/MCOM.2015.7081072

Crossref Full Text | Google Scholar

Esmaeilzadeh, P. (2024). Privacy concerns about sharing general and specific health information on twitter: quantitative study. JMIR Form. Res. 8:e45573. doi: 10.2196/45573

PubMed Abstract | Crossref Full Text | Google Scholar

Foddy, M., Platow, M. J., and Yamagishi, T. (2009). Group-based trust in strangers: the role of stereotypes and expectations. Psychol. Sci. 20, 419–422. doi: 10.1111/j.1467-9280.2009.02312.x

PubMed Abstract | Crossref Full Text | Google Scholar

Govani, T., and Pashley, H. (2007). Student awareness of the privacy implications when using Facebook. Proceedings of The IEEE - PIEEE.

Google Scholar

Hair, J. F., Sarstedt, M., Ringle, C. M., and Mena, J. A. (2012). An assessment of the use of partial least squares structural equation modeling in marketing research. J. Acad. Mark. Sci. 40, 414–433. doi: 10.1007/s11747-011-0261-6

Crossref Full Text | Google Scholar

Hollenbaugh, E. E., and Ferris, A. L. (2014). Facebook self-disclosure: examining the role of traits, social cohesion, and motives. Comput. Hum. Behav. 30, 50–58. doi: 10.1016/j.chb.2013.07.055

Crossref Full Text | Google Scholar

Hong, W., and Thong, J. Y. L. J. M. Q. (2013). Internet privacy concerns: an integrated conceptualization and four empirical studies 37, 275–298. doi: 10.25300/MISQ/2013/37.1.12

Crossref Full Text | Google Scholar

Jia, S. Z., Chi, O. H., and Lu, L. (2024). Social robot privacy concern (SRPC): rethinking privacy concerns within the hospitality domain. Int. J. Hosp. Manag. 122:Article 103853. doi: 10.1016/j.ijhm.2024.103853

Crossref Full Text | Google Scholar

Jozani, M., Ayaburi, E., Ko, M., and Choo, K.-K. R. (2020). Privacy concerns and benefits of engagement with social media-enabled apps: a privacy calculus perspective. Comput. Hum. Behav. 107:106260. doi: 10.1016/j.chb.2020.106260

Crossref Full Text | Google Scholar

Jung, A. R. (2017). The influence of perceived ad relevance on social media advertising: an empirical examination of a mediating role of privacy concern. Comput. Hum. Behav. 70, 303–309. doi: 10.1016/j.chb.2017.01.008

Crossref Full Text | Google Scholar

Kang, H., Shin, W., and Huang, J. (2022). Teens' privacy management on video-sharing social media: the roles of perceived privacy risk and parental mediation. Internet Res. 32, 312–334. doi: 10.1108/INTR-01-2021-0005

Crossref Full Text | Google Scholar

Kolotylo-Kulkarni, M., Xia, W., and Dhillon, G. (2021). Information disclosure in e-commerce: a systematic review and agenda for future research. J. Bus. Res. 126, 221–238. doi: 10.1016/j.jbusres.2020.12.006

Crossref Full Text | Google Scholar

Krasnova, H., Spiekermann, S., Koroleva, K., and Hildebrand, T. (2010). Online social networks: why we disclose. J. Inf. Technol. 25, 109–125. doi: 10.1057/jit.2010.6

Crossref Full Text | Google Scholar

Li, J., Zhang, Y., and Mou, J. (2023). Understanding information disclosures and privacy sensitivity on short-form video platforms: an empirical investigation. J. Retail. Consum. Serv. 72:103292. doi: 10.1016/j.jretconser.2023.103292

Crossref Full Text | Google Scholar

Liu, T., Li, W., and Jia, X. (2022). Consumer data vulnerability, peer privacy concerns and continued usage intention of sharing accommodation platforms: the moderating roles of perceived benefits. Inf. Technol. People 36:2258. doi: 10.1108/itp-05-2021-0365

Crossref Full Text | Google Scholar

Liu, Y. L., Yan, W. J., Hu, B., Lin, Z., and Song, Y. Y. (2024). Chatbots or humans? Effects of agent identity and information sensitivity on users' privacy management and behavioral intentions: a comparative experimental study between China and the United States. Int. J. Hum. Comput. Interac. 40, 5632–5647. doi: 10.1080/10447318.2023.2238974

Crossref Full Text | Google Scholar

Liu, Z., Min, Q., Zhai, Q., and Smyth, R. (2016). Self-disclosure in Chinese micro-blogging: a social exchange theory perspective. Inf. Manag. 53, 53–63. doi: 10.1016/j.im.2015.08.006

Crossref Full Text | Google Scholar

Livingstone, S. (2008). Taking risky opportunities in youthful content creation: teenagers' use of social networking sites for intimacy, privacy and self-expression. New Media Soc. 10, 393–411. doi: 10.1177/1461444808089415

Crossref Full Text | Google Scholar

Loiacono, E. T. (2015). Self-disclosure behavior on social networking web sites. Int. J. Electron. Commer. 19, 66–94. doi: 10.1080/10864415.2015.979479

Crossref Full Text | Google Scholar

Malhotra, N. K., Kim, S. S., and Agarwal, J. (2004). Internet users' information privacy concerns (IUIPC): the construct, the scale, and a causal model. Inf. Syst. Res. 15, 336–355. doi: 10.1287/isre.1040.0032

Crossref Full Text | Google Scholar

Maseeh, H. I., Nahar, S., Jebarajakirthy, C., Ross, M., Arli, D., Das, M., et al. (2023). Exploring the privacy concerns of smartphone app users: a qualitative approach. Mark. Intell. Plann. 41, 945–969. doi: 10.1108/MIP-11-2022-0515

Crossref Full Text | Google Scholar

Metzger, M. J. (2007). Communication privacy management in electronic commerce. J. Comput.-Mediat. Commun. 12, 335–361. doi: 10.1111/j.1083-6101.2007.00328.x

Crossref Full Text | Google Scholar

Min, J., and Kim, B. (2015). How are people enticed to disclose personal information despite privacy concerns in social network sites? The calculus between benefit and cost. J. Assoc. Inf. Sci. Technol. 66, 839–857. doi: 10.1002/asi.23206

Crossref Full Text | Google Scholar

Nemec Zlatolas, L., Welzer, T., Heričko, M., and Hölbl, M. (2015). Privacy antecedents for SNS self-disclosure: the case of Facebook. Comput. Hum. Behav. 45, 158–167. doi: 10.1016/j.chb.2014.12.012

Crossref Full Text | Google Scholar

Nemec Zlatolas, L., Welzer, T., Holbl, M., Hericko, M., and Kamisalic, A. (2019). A model of perception of privacy, trust, and self-disclosure on online social networks. Entropy 21:772. doi: 10.3390/e21080772

PubMed Abstract | Crossref Full Text | Google Scholar

Neves, J., Turel, O., and Oliveira, T. (2022). Sns use reduction: a two-facet privacy concern perspective. Internet Res. 33, 974–993. doi: 10.1108/intr-01-2022-0012

Crossref Full Text | Google Scholar

Nguyen, T. T., Tran, V. T. T., and Hoang, M. T. T. (2024). How peer privacy concerns affect active and passive uses of social networking sites: a dual peer privacy calculus model. Soc. Sci. Comput. Rev. 42, 1440–1463. doi: 10.1177/08944393231224539

Crossref Full Text | Google Scholar

Ozdemir, Z. D., Jeff Smith, H., and Benamati, J. H. (2017). Antecedents and outcomes of information privacy concerns in a peer context: an exploratory study. Eur. J. Inf. Syst. 26, 642–660. doi: 10.1057/s41303-017-0056-z

Crossref Full Text | Google Scholar

Petronio, S. (2002). Boundaries of privacy: dialectics of disclosure. New York: State University of New York Press.

Google Scholar

Podsakoff, P. M., and Organ, D. W. (1986). Self-reports in organizational research: problems and prospects. J. Manage. 12, 531–544. doi: 10.1177/014920638601200408

Crossref Full Text | Google Scholar

Popp, R., and Poindexter, J. (2006). Countering terrorism through information and privacy protection technologies. IEEE Secur. Priv. Mag. 4, 18–27. doi: 10.1109/MSP.2006.147

Crossref Full Text | Google Scholar

Quan-Haase, A., and Ho, D. (2020). Online privacy concerns and privacy protection strategies among older adults in East York. CANA 71, 1089–1102. doi: 10.1002/asi.24364

PubMed Abstract | Crossref Full Text | Google Scholar

Raynes-Goldie, K. (2010). Aliases, creeping, and wall cleaning: understanding privacy in the age of Facebook. First Monday 15. doi: 10.5210/fm.v15i1.2775

Crossref Full Text | Google Scholar

Smith, S. A., and Brunner, S. R. (2017). To reveal or conceal: using communication privacy management theory to understand disclosures in the workplace. Manag. Commun. Q. 31, 429–446. doi: 10.1177/0893318917692896

Crossref Full Text | Google Scholar

Wang, J., Cao, Q., and Zhu, X. (2024). Privacy disclosure on social media: the role of platform features, group effects, trust and privacy concern. Libr. Hi Tech 43, 1035–1059. doi: 10.1108/LHT-06-2023-0253

Crossref Full Text | Google Scholar

Wang, L., Sun, Z., Dai, X., Zhang, Y., and Hu, H.-H. (2019). Retaining users after privacy invasions. Inf. Technol. People 32, 1679–1703. doi: 10.1108/itp-01-2018-0020

Crossref Full Text | Google Scholar

Wang, W.-T., Ou, W.-M., and Chiu, W.-C. (2021). Autonomous motivation and self-disclosure intention. J. Organ. End User Comput. 33, 1–27. doi: 10.4018/joeuc.286808

PubMed Abstract | Crossref Full Text | Google Scholar

Wang, Y., and Herrando, C. (2019). Does privacy assurance on social commerce sites matter to millennials? Int. J. Inf. Manag. 44, 164–177. doi: 10.1016/j.ijinfomgt.2018.10.016

Crossref Full Text | Google Scholar

Wang, Y., Zheng, D. Q., and Fang, Y. L. (2024). Public information sharing in enterprise social networks: a communication privacy management perspective. Internet Res. 34, 610–630. doi: 10.1108/intr-09-2022-0745

Crossref Full Text | Google Scholar

West, R., and Turner, L. H. (2014). Introducing Communication Theory: Analysis and Application (Fifth Edition). New York: McGraw-Hill Education.

Google Scholar

Wu, K.-W., Huang, S. Y., Yen, D. C., and Popova, I. (2012). The effect of online privacy policy on consumer privacy concern and trust. Comput. Hum. Behav. 28, 889–897. doi: 10.1016/j.chb.2011.12.008

Crossref Full Text | Google Scholar

Xu, F., Michael, K., and Chen, X. (2013). Factors affecting privacy disclosure on social network sites: an integrated model. Electron. Commerce Res. 13, 151–168. doi: 10.1007/s10660-013-9111-6

Crossref Full Text | Google Scholar

Xu, H., Dinev, T., Smith, J., and Hart, P. (2011). Information privacy concerns: linking individual perceptions with institutional privacy assurances. J. Assoc. Inf. Syst. 12, 798–824. doi: 10.17705/1jais.00281

Crossref Full Text | Google Scholar

Yang, S., Wang, Y., and Wang, K. (2008). The influence of information sensitivity, compensation on privacy concern and behaviour intention. SIGMIS Database 40, 38–51. doi: 10.1145/1496930.1496937

Crossref Full Text | Google Scholar

Young, R., Willis, E., Cameron, G., and Geana, M. (2013). “Willing but unwilling”: attitudinal barriers to adoption of home-based health information technology among older adults. Health Informatics J. 20, 127–135. doi: 10.1177/1460458213486906

PubMed Abstract | Crossref Full Text | Google Scholar

Zhang, N., Wang, C., Karahanna, E., and Xu, Y. (2022). Peer privacy concerns: conceptualization and measurement. MIS Q. 46, 491–530. doi: 10.25300/misq/2022/14861

Crossref Full Text | Google Scholar

Zhu, X., Cao, Q., and Liu, C. (2022). Mechanism of platform interaction on social media users’ intention to disclose privacy: a case study of TikTok APP. Information 13:461. doi: 10.3390/info13100461

Crossref Full Text | Google Scholar

Keywords: privacy disclosure behavior, short video, institutional privacy concerns, peer privacy concerns, influencing factors

Citation: Lu H, Wei A and Li X (2025) Understanding the privacy disclosure behavior and privacy concerns on short-form video platforms: an empirical investigation. Front. Psychol. 16:1535533. doi: 10.3389/fpsyg.2025.1535533

Received: 27 November 2024; Accepted: 08 September 2025;
Published: 23 September 2025.

Edited by:

Thomas J. Huggins, University of Monterrey, Mexico

Reviewed by:

Junyi Chai, United International College, China
Rafael Octavio Félix Verduzco, Universidad de Monterrey Division de Ciencias de la Salud, Mexico

Copyright © 2025 Lu, Wei and Li. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Heng Lu, bHVoZW5nNjk4QDE2My5jb20=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.