Skip to main content

ORIGINAL RESEARCH article

Front. Psychol., 26 May 2022
Sec. Organizational Psychology

Do Critical Reviews Affect Box Office Revenues Through Community Engagement and User Reviews?

  • 1College of International Business, Zhejiang Yuexiu University, Shaoxing, China
  • 2Shaoxing Key Laboratory for Smart Society Monitoring, Prevention and Control, Shaoxing, China
  • 3School of Finance, Jiangxi University of Finance and Economics, Nanchang, China

With advances in technology and the popularity of the Internet, consumers increasingly rely on various sources of electronic word-of-mouth (eWOM), such as online user reviews and critical reviews, in their decision-making processes. Despite general consensus on the importance of eWOM and the ability of critical reviews to influence product sales, very little is known about the mediation between critical reviews and user reviews. Therefore, we used path analysis to examine how critical reviews and user reviews simultaneously affect box office revenues using eWOM data collected from Metacritic.com and IMDb.com, and box office revenue information collected from BoxOfficeMojo.com. The results showed that critical reviews valence not only directly affects box office revenues but also increases active postings in the community and user reviews volume, thus indirectly leading to greater box office revenues. The study provides strategic guidance and practical implications for eWOM communication management.

Introduction

The popularity of the Internet has clearly affected both the way consumer information is collected and companies’ marketing and communication strategies. Today’s consumers can easily access and use various sources of electronic word-of-mouth (eWOM) information, such as online user reviews (URs) and critical reviews (CRs). In a survey conducted by Podium (2017), an overwhelming 93 percent of consumers agreed that online reviews influence their buying decisions. Accordingly, companies seek to promote their products through online reviews and even obtain information on consumer preferences to develop and improve products. Therefore, online reviews are not only an important reference for consumer decision-making but also an effective marketing communication tool in many industries (Chevalier and Mayzlin, 2006; Liu, 2006; Zhu and Zhang, 2010; Chen et al., 2011; Moe and Trusov, 2011; Zhou and Duan, 2016). The ability to access product evaluation information via the Internet reduces consumers’ time and effort to gather information to compare multiple attributes of a product for purchase decisions (Chen and Xie, 2005; Gu et al., 2012). For companies, online product reviews not only reduce marketing communication costs but also accelerate the dissemination of new product information, which in turn affects product sales (Moon et al., 2010; Moe and Trusov, 2011).

Many popular sites, such as Amazon and CNETD, offer both user reviews and critical reviews of listed products. Online user reviews are spontaneously generated by consumers based on their personal experiences, such as movie reviews on IMDb, book reviews on Amazon, restaurant reviews on Yelp, and hotel reviews on TripAdvisor. Online user reviews are considered a valid proxy for WOM and can influence consumer decision-making (Zhu and Zhang, 2010). Many empirical findings support this idea. For example, Godes and Mayzlin (2004) indicated that eWOM volume increases sales in the context of TV shows. Liu (2006) studied movie reviews and found that the information provided by online user reviews has significant explanatory power for box office revenue. Chevalier and Mayzlin’s (2006) study showed that online book ratings affect consumer purchase behavior. In contrast to consumer-generated reviews, critical reviews mainly come from professionals and well-known newspapers, magazines, and media and seek to demonstrate the quality of the product (Chen et al., 2012), such as pre-release movie reviews and professional reviews of selected products on CNETD. Such recommendation information also has a positive impact on the sales of new products. For example, Eliashberg and Shugan (1997) showed that critical reviews are correlated with late and cumulative box office revenue but not with early box office revenue. Thus, they concluded that film critics are predictors rather than influencers. By contrast, Basuroy et al. (2003) showed that professional reviews both influence and predict box office revenue. Similar results were obtained by Boatwright et al. (2007). Regardless of whether they are written by users or critics, online reviews have a positive effect on sales.

Companies are increasingly devoting more resources to eWOM marketing communication because they believe online WOM is far more effective than offline WOM. In addition, many empirical studies have examined the impact of eWOM on sales and generated practical implications (Basuroy et al., 2003; Godes and Mayzlin, 2004; Chevalier and Mayzlin, 2006; Liu, 2006; Duan et al., 2008; Zhu and Zhang, 2010; Zhou and Duan, 2016). However, few studies have discussed whether critical reviews indirectly influence consumer choice through consumer reviewers. We lack an in-depth understanding of the interplay between online critical reviews and user reviews and its influence on consumers’ choices. A previous study by Liu (2006) explored the relationship between the dynamics of online user reviews and movie box office as well as the effect of critical reviews on the volume of user reviews, but this latter effect was not significant. The lack of an effect may be attributable to differences in the sources of WOM, as critical reviews come from physical magazines (i.e., Variety), while user reviews come from websites (i.e., Yahoo! Movie). As a result, some consumers may not seek out critical reviews. In the context of software downloads, Zhou and Duan (2016) showed that online user reviews mediate the impact of professional reviews on user decisions. This study extends the framework of Zhou and Duan (2016) to investigate the direct impact of critical reviews on the volume of not only user reviews but also postings by community members (i.e., community engagement). In this study, we assume that there is a group of amateur members in the film review community who are passionate about movies and will share their viewing experiences on the platform (not only stat ratings but also text reviews). We expect that the impact of critical reviews on this amateur group of community members will be greater than the impact of critical reviews on the volume of user reviews. In addition, we anticipate that community engagement also indirectly affects box office revenue through the volume of user reviews.

This study collects online review data from IMDb.com, which contains both critical and user reviews, and critical reviews from Metacritic.com, which is a comprehensive source containing a wider range of critical reviews. Data from Metacritic.com has been used in a number of empirical studies (Wiles and Danielova, 2009; Chen et al., 2012; Wang et al., 2015). The purpose of this study is to examine the impact of the mediating role of community engagement in the relationship between critical reviews and user choice. We contribute to the literature in two main ways. First, this study includes two different sources of online reviews (CRs and URs) and examines their direct and indirect effects on consumer choice. Second, an important mediating variable (i.e., community engagement) is proposed in this study. We confirm the mediation model for the effect of CRs and URs on consumer choice, and provide a more complete description of the influence path of eWOM from different sources as well as practical applications. To the best of our knowledge, this work is the first to study the mediating role of community engagement in the relationship between critical reviews and user choices.

Literature Review and Hypotheses Development

In this study, we draw primarily on two main streams of research in information systems and marketing: (1) the generation of eWOM, and (2) the sales impact of CRs and URs. Based on these related studies, we propose our research model.

The Generation of Electronic Word-of-Mouth

There are two main theories: psychological motivation theory and review environment theory. Applying these two theories, we propose that CRs can influence the number of URs. Critics and users differ in terms of their motivation and their target audience when writing reviews. Critics write reviews for professional reasons, attempting to avoid emotional and biased statements. Since they write for a wider audience and have considerable experience with quality judgment (Amblee and Bui, 2007). Users, on the other hand, tend to write reviews for personal and emotional reasons (Hennig-Thurau et al., 2004). Another difference is the different timing of posts. CRs usually remain unchanged and are posted earlier than URs, while URs are posted throughout the period of product availability, and such reviews may be influenced by the opinions of other users (Zhou and Duan, 2016). The first theory argues that, in terms of the psychological motivation for users’ sharing behaviors, users are willing to share their experiences and insights in the community is mainly due to self-enhancement (Hennig-Thurau et al., 2004). Self-enhancement refers to the emotional desire to gain attention and improve image among others (Hennig-Thurau et al., 2004). Hennig-Thurau et al. (2004) showed that self-enhancement affects the number of reviews on online opinion platforms. Thus, psychological motivation theory implies that community members are willing to write product reviews. In other words, community members are motivated to post by an intrinsic drive for self-worth reinforcement (Hussain et al., 2018). In the context of movie reviews, online users can consider CRs as an indicator of the attention or popularity of the corresponding movie. In other words, movies that have been reviewed by critics are more visible, compared to those without critical reviews. By posting reviews of movies that have been reviewed by critics through movie sharing platforms, online users may perceive that others are more likely to read their posts, and thus more likely to project themselves as intelligent consumers. On the other hand, online users may anticipate their reviews would be more valuable to other readers in the absence of CRs, and thus be more motivated to write reviews of these movies. Therefore, we expect to see a significant relationship between the CRs and URs. The second theory: review environment theory also supports this relationship between the CRs and URs. Essentially, this theory suggests that environmental factors (i.e., the opinions of others) influence users’ decision to post (Moe and Schweidel, 2012; Goes et al., 2014). The usefulness of information and related interactions are key factors that influence user behavior (Hussain et al., 2020a,b). Similarly, Goes et al. (2014) showed that online user interaction can increase the number of product reviews. Therefore, an increase in the number of consumers participating in the online movie community will increase the number of user reviews (Wu et al., 2018). In addition, when CR is often provided before or at the same time as URs, they help characterize the user’s review environment. Therefore, the evaluation of an individual product by experts can influence the user’s decision whether to review the product or not.

The Sales Impact of Critical Reviews and User Reviews

Due to the growing popularity of the Web and social media, CRs and URs now constitute a new element in the marketing communications mix that can have a significant impact on sales (see Table 1 for a summary of representative studies). Two common measures of eWOM have been widely discussed: volume and valence. Volume refers to the number of reviews posted by critics or users. A larger volume of URs can better attract users’ attention to the product/service and accordingly leads to more opportunities to choose it. Previous studies have verified its positive impact on sales (Godes and Mayzlin, 2004; Liu, 2006; Duan et al., 2008; Zhou and Duan, 2016). Valence indicates the favorability of the product, which can be measured by the average ratings of critics or users. Previous studies have examined the impact of URs valence on sales, reporting mixed findings. Some researchers have found a positive impact of URs valence on sales (Chevalier and Mayzlin, 2006; Zhu and Zhang, 2010; Chen et al., 2011). Others have found no significant sales impact of URs valence (Liu, 2006). As another eWOM source, CRs are provided by experts and serve as advertisements (Haan et al., 2005). A relationship between CRs valence and sales has been observed (Eliashberg and Shugan, 1997; Basuroy et al., 2003; Reinstein and Snyder, 2005; Boatwright et al., 2007). Recent studies have begun to investigate the roles of both in the same model (Holbrook and Addis, 2007; Zhou and Duan, 2016). Holbrook and Addis (2007) found that URs valence mediates the effect of CRs valence on popular appeal. Zhou and Duan (2016) showed a similar mediating role of URs, but focused on the URs volume. This study extends work by Zhou and Duan (2016) on the indirect effect of professional reviews on online consumer choices through the volume of user reviews. However, the present study argues that some amateurs are more influential in the film community than others, and when these more influential amateurs recommend or share information about a film in the community, other community members adopt their reviews as a reference for decision-making. Thus, we argue that community engagement is an effective mechanism for influencing consumer choice. We propose that critical reviews influence consumer choice both directly and indirectly through community engagement and online user review volume (as shown in Figure 1). Therefore, this study examines how CRs and URs simultaneously influence consumer choice by focusing on the mediating effects of community engagement.

TABLE 1
www.frontiersin.org

Table 1. Literature on the effects of critical reviews (CRs) and user reviews (URs).

FIGURE 1
www.frontiersin.org

Figure 1. Conceptual framework.

Hypotheses Development

Critical Reviews

In recent years, critical reviews have been an important focus of both practical and academic research. Critical reviews provide not only professional reviews to create reputations but also valuable product information for advertising (Haan et al., 2005). Critical reviews often provide a relatively professional critique of the quality of a product and are frequently published before the release of a new product, such as a film. Many studies have shown that critical reviews significantly affect consumer choices and product sales (Eliashberg and Shugan, 1997; Basuroy et al., 2003, 2006; Boatwright et al., 2007). For example, Eliashberg and Shugan (1997) reported that critical reviews are associated with late and cumulative box office revenue but not with early box office revenue. By contrast, Hennig-Thurau et al.’s (2006) study showed that critical reviews have a significant impact on short-term but not long-term box office. In addition, several studies have shown that film critics play dual roles as both influencers and predictors (Basuroy et al., 2003; Boatwright et al., 2007). Reinstein and Snyder (2005) applied a difference-in-difference approach to indicate the impact of movie critics on sales. They found a marginal positive effect of movie critics on demand. Thus, higher critical review valence increases box office revenues (Eliashberg and Shugan, 1997; Basuroy et al., 2003). Therefore, we hypothesize the following:

H1: Critical reviews valence has a positive impact on movie box office.

Critical reviews have been found to influence the amount of online user-generated review behavior, which is usually measured by the number of online user reviews (Liu, 2006; Duan et al., 2008; Zhou and Duan, 2016). There are three possible reasons for this finding. First, with the development of the Internet and information technology, critical reviews have become widely available in the public domain. The rise of social media has further enhanced the impact of critical reviews on user-generated content (Zhou and Duan, 2016). Second, critical reviews have become a valuable source of information for consumers to evaluate product quality (Chen and Xie, 2005), usually from professional and reputable media or experts in the field (Liu, 2006; Chen and Xie, 2008). Finally, consumers perceive more positive critical reviews as a signal expressing higher popularity and are more willing to share reviews of the corresponding products (Holbrook and Addis, 2007). Past empirical research has found a significant positive correlation between critical reviews and the volume of online user reviews (Holbrook and Addis, 2007; Zhou and Duan, 2016). For example, Holbrook and Addis (2007) reported a positive impact of critical reviews on consumer review volume, and Zhou and Duan (2016) also demonstrated a positive effect of professional reviews on user review volume in the software download context. Hence, we hypothesize the following:

H2: Critical reviews valence has a positive impact on user review volume.

While Liu (2006) indicated that critical reviews affect consumers’ film expectations and the volume of user reviews, the empirical results did not support this argument. A potential reason is that the critical reviews were obtained from physical magazines (i.e., Variety), while the user reviews were obtained from websites (i.e., Yahoo! Movie). The use of different data sources may lead to inconsistent results for the impact of critical reviews on the number of user reviews. To overcome this problem, in this study, both critical review and consumer review data were obtained from the website IMDb.com, where consumers can easily access both critical reviews and user reviews of films. In addition, according to review environment theory, consumer comments are influenced by environmental factors, such as the opinions of others (Moe and Schweidel, 2012; Goes et al., 2014). For example, Moe and Schweidel’s (2012) study examined how previously posted product ratings affect an individual’s posting frequency. Their findings show that when the rating environment is positive, consumers are more likely to generate online product reviews after purchase. That is, some amateurs are more willing to engage in the act of posting to the film community if the critical reviews are favorable. Hence, we hypothesize the following:

H3: Critical reviews valence has a positive impact on community engagement.

Online User Reviews

As another eWOM source, online user reviews are product evaluations generated by ordinary users online that mainly provide product quality information based on their purchase experience (Chen and Xie, 2008). There are two types of online user reviews: text product reviews written by customers and numerical star ratings (e.g., between 1 and 5 stars). Both types can be important drivers of product sales and provide consumers with useful product information. In general, text reviews provide richer information than numerical star ratings for consumer decision-making (Salehan and Kim, 2016). In addition, Mudambi and Schuff (2010) suggested that review depth (i.e., the word count of the review) can enhance information diagnosticity and provide increased depth of information to help consumers make decisions with confidence. Chevalier and Mayzlin (2006) pointed out that the length of reviews for a book significantly predicts its sales on Amazon.com. Moreover, the volume of online user reviews (i.e., the total number of user reviews) makes the corresponding product stand out from the crowd and thus gain the attention of consumers (Liu, 2006). Many studies have shown that the number of online user reviews significantly affects consumer choices and product sales (Liu, 2006; Duan et al., 2008; Zhu and Zhang, 2010; Sun, 2012). For example, Liu (2006) showed that online movie review volume has important explanatory power for both total box office and weekly box office revenue. Zhu and Zhang (2010) indicated that online user reviews have a significant positive effect on product sales in the video game industry. Therefore, a large number of user reviews better draws consumers’ attention to the product, increasing the likelihood of choosing that product. Hence, we hypothesize the following:

H4: User reviews volume has a positive impact on movie box office.

Community Engagement

Community engagement is defined as “the consumer’s intrinsic motivations to interact and cooperate with community members” (Algesheimer et al., 2005, p. 21). Consumers share their personal experience, knowledge, and information through the process of actively contributing to the online community, thus highlighting the interactive and bidirectional nature of community participation (Algesheimer et al., 2005). Consequently, many studies suggest that community contribution, such as posting activities, reflects a high level of members’ community participation (Koh et al., 2007) and can be used to measure community engagement (Oestreicher-Singer and Zalmanson, 2013). Community engagement is also related to the desired effect of identification with the brand community on community members. Higher community engagement means higher collaboration and interaction, which in turn will lead to higher positive effects (i.e., community participation or recommendation behavior) (Algesheimer et al., 2005). Recent studies have shown that community engagement is more likely to generate online reviews (Wu et al., 2018), improve online sales (Wu et al., 2015), and enhance customer brand loyalty (Zheng et al., 2015). Hence, we hypothesize the following:

H5: Community engagement has a positive impact on movie box office.

The opinions shared by consumers in the film community can influence the choices of other community members (Moon et al., 2010). Petty and Cacioppo (1984) argued that the greater WOM, the higher the degree of consumer involvement in products; as a result, consumers are more likely to be influenced by others and shape changes in product attitudes themselves. The more WOM spreads on the Internet, that is, the greater the discussion of the product or service, the more likely WOM is to spread again (Brown and Reingen, 1987; Banerjee, 1992). Goes et al. (2014) showed that users tend to produce more objective reviews in order to gain more trust from their peers. In other words, online reviews have social influence on other consumers (Sridhar and Srinivasan, 2012). When consumers are willing to read and compare user reviews in the movie community, they will obtain more information about the movie. We expect that the depth of the text of reviews will enhance consumers’ ability to detect products, attract more consumer attention, and thereby promote consumers to generate online reviews. As a result, an increase in the number of consumers who engage in the online movie community will increase the number of user reviews (Wu et al., 2018). Therefore, we hypothesize the following:

H6: Community engagement has a positive impact on user review volume.

Methodology

Data Collection Process

To examine whether online user reviews mediate the impact of critical reviews on consumer choice, we collected movie characteristic information, critical reviews, user reviews, and box office using web-scraping techniques. We collected data on the top 100 movie releases by United States box office revenue each year in the approximately 6-year period from January 1, 2010 to December 31, 2015. We obtained movie characteristics such as budget, sequel, genre, MPAA rating, and running time from the Internet Movie Database1. We obtained movie box office revenue from Box Office Mojo2. After removing 83 movies with incomplete information, we finally included 517 movies that had been successful at the United States box office. Box office revenue per movie ranged from $21.07 million to $933.66 million, with an average of $101.12 million. The data included specific information on movies from six major genre categories: action or adventure (28%), animation (9%), thriller (20%), comedy or drama (46%), sci-fi or fantasy (12%), and romance (15%). Among the included movies, 19% were sequels, and 21% were rated “general” or “parental guidance suggested.” The average production budget was approximately $69.43 million. The maximum production budget was $260 million, and the minimum was $1.5 million; thus, the highest budget was 173 times greater than the lowest. Detailed descriptive statistics for the sample are reported in Table 2.

TABLE 2
www.frontiersin.org

Table 2. Descriptive statistics of the attributes of the sample movies within the time period 2010–2015.

For each movie, critical reviews were collected form Metacritic3. Metacritic provides the name of the critic, media outlet, and review score for each movie and each review. We used Metacritic as our primary source of critical review data for the following reasons. First, Metacritic is one of the most comprehensive movie critical review sites, and its coverage and summaries of film reviews have been well-received in the film industry. Second, Metacritic presents a summary of each review in the form of a score on a scale between 0 and 100. These scores provide an independent, professionally assessed index of review valence. Third, movie critical reviews from Metacritic have been increasingly used in recent research (Wiles and Danielova, 2009; Chen et al., 2012; Wang et al., 2015). We also collected film review data from a large number of public online users of IMDb.com. This site was chosen because it is one of the most popular online film review sites, with over 57 million independent visitors per month. More importantly, critical review data from Metacritic can also be found on this site.

Variable Measurement and Construction

Dependent Variable

Box Office Revenue

We obtained box office revenue from BoxOfficeMojo.com. Box office revenue (REV) was measured as the total box office revenue of a movie in natural logarithm form (Joshi and Hanssens, 2009; Leenders and Eliashberg, 2011). Since the box office revenues were right-skewed, we used a natural log-transformation of the cumulative values, with subsequent skewness and kurtosis values suggesting a normal distribution.

Independent Variables

Critical Review Valence

The critical review valence (CRVAL) was measured by the average score of critics for each movie posted on the Metacritic.com website (Chen et al., 2012).

User Review Volume

We obtained user reviews from IMDb.com. User review volume (URVOL) was measured by the cumulative number of consumers’ reviews in natural logarithm form (Moe and Trusov, 2011; Sun, 2012).

Community Engagement

Community engagement was defined as users’ prosocial contributions to the online movie community. In this study, we measure community engagement (POSTVOL) as the number of postings for each movie (Oestreicher-Singer and Zalmanson, 2013; Wu et al., 2018). Higher movie review postings in the community indicated higher customer engagement.

Control Variables

We employed some movie characteristics as control variables in this study. First, several studies have shown that the production budget of a movie is significantly related to its box office revenues (Hennig-Thurau et al., 2006; Liu, 2006). Production budget (PB) was measured as the production budgets of movies in natural logarithm form (Hennig-Thurau et al., 2006). Second, star power is widely recognized in the film industry (Ravid, 1999). Star power was measured as the sum of the average box office revenue in natural logarithm form for actors (SPA), directors (SPD), and producers (SPP) in the past 3 years, weighted by their ranking among all participants in the same movie production roles. To operationalize star power, we followed Hennig-Thurau et al.’s (2006) suggested method. Taking actors for example, when an actor was listed in the first place on the movie page on Box Office Mojo, the weight given was 1, followed by 0.5, 0.25, 0.125, and so on. Third, sequel movies are generally found to generate higher box office revenues (Basuroy et al., 2006; Moon et al., 2010). SEQUEL is a binary dummy taking a value of 1 if the movie was the sequel of a previous movie and 0 otherwise (Hennig-Thurau et al., 2006). Fourth, the MPAA rating is considered an important factor in the industry. Ravid (1999) found MPAA ratings to be significant variables in his research. MPAA rating is also a binary dummy variable constructed according to ratings of the appropriateness of movies for different ages as classified by the MPAA: be G (general audiences), PG (parental guidance suggested), PG-13 (parental guidance-13), R (restricted), NC-17 (no one under 17 admitted), and NR (not rated) (Hennig-Thurau et al., 2015). This study categorized G and PG as a group into dummy 1 and PG-13, R, NC-17, and NR as another group into dummy 0. Fifth, genre (GR) is a binary dummy constructed according to the classifications used by IMDb, for which the superscripts are AA, AN, THR, CD, SF and RO, representing action, animation, thriller, comedy and drama, sci-fi and fantasy, and romance, respectively (Lee, 2006; Akdeniz and Talay, 2013). The variables, measures, and data sources are reported in Table 3.

TABLE 3
www.frontiersin.org

Table 3. Variable definitions, measurements and data sources.

Empirical Analysis

Correlation Analysis

Table 4 provides a matrix of correlation coefficients for each variable. Box office revenue (REV) tended to be positively correlated with critical review valence (CRVAL), user review volume (URVOL), and community engagement (POSTVOL), with correlation coefficients of 0.34, 0.56, and 0.51, respectively. Critical review valence (CRVAL) was positively correlated with user review volume (URVOL) and community engagement (POSTVOL), with correlation coefficients of 0.46 and 0.53, respectively. In addition, a high degree of association was apparent between user review volume (URVOL) and community engagement (POSTVOL), with a correlation coefficient of 0.86.

TABLE 4
www.frontiersin.org

Table 4. Correlation matrix.

Path Analysis

This study investigates a mediation model in which community engagement (POSTVOL) and user review volume (URVOL) mediate the impact of critical review valence (CRVAL) on box office revenue (REV). First, as shown in Figure 2, there was a positive effect of critical review valence (CRVAL) on box office revenue (REV), with a coefficient of 0.10 (p < 0.01), supporting H1. Second, critical review valence (CRVAL) had positive effects on user review volume (URVOL) and community engagement (POSTVOL), with coefficients of 0.18 (p < 0.001) and 0.46 (p < 0.001), respectively, supporting H2 and H3 and indicating a greater impact of critical review valence (CRVAL) on community engagement (POSTVOL). Third, the direct effect of user review volume (URVOL) on box office revenue (REV) was significantly positive, with an effect coefficient of 0.22 (p < 0.001), supporting H4. Fourth, community engagement (POSTVOL) had a positive effect on box office revenue (REV), with a coefficient of 0.30 (p < 0.001), supporting H5. Fifth, the effect of community engagement (POSTVOL) on user review volume (URVOL) was significantly positive, with a coefficient of 0.78 (p < 0.001), supporting H6. Thus, the more frequent the interaction between community members, the more topics and attention they generate, which in turn creates higher review volume. Finally, at a significance level of 5%, the effect of production budget (PB) on box office revenue (REV) was significantly positive, echoing findings in the literature (Basuroy et al., 2003; Elberse and Eliashberg, 2003). In addition, critical review valence (CRVAL) indirectly affected box office through user review volume (URVOL) and community engagement (POSTVOL), with an indirect effect coefficient of 0.264.

FIGURE 2
www.frontiersin.org

Figure 2. Results of path analysis for the proposed model. ***, ** and * denote significance at the 0.001, 0.01 and 0.05 levels, respectively. PB denotes production budget. SPA denotes star power of actors. SPD denotes star power of directors. SPP denotes star power of producers. MPAA rating denotes whether the movie is rated for age. Genre has six categories: action (GRAA), animation (GRAN), thriller (GRTHC), comedy and drama (GRCD), sci-fi and fantasy (GRSF), and romance (base category).

Conclusion

This study deepens the understanding of online rating behavior in marketing theory (Chevalier and Mayzlin, 2006; Liu, 2006; Moe and Trusov, 2011; Moe and Schweidel, 2012) by further clarifying how the relationship between CRs and URs affects consumer choice.

Empirical Findings and Managerial Implications

This study examined whether CRs valence affects product sales through URs processes. The results show that CRs valence may perform better through user review processes such as community engagement and URs volume. These findings further clarify the mechanism by which URs impact the relationship between CRs valence and product sales. This study contributes to the development of theory related to the effect of eWOM on product sales. First, this study shows that CRs valence have a positive effect on sales, consistent with previous findings (Basuroy et al., 2003; Boatwright et al., 2007). We also show that CRs valence has a positive effect on URs volume and community engagement. The present study extends Zhou and Duan (2016) to examine how URs and CRs simultaneously influence online user choices by focusing on the mediation relationship. They pointed out that CRs valence indirectly affect online user choices through URs volume. Consistent with this line of thought, the present study demonstrates that in addition to its direct correlation with URs volume, CRs valence is indirectly related to URs volume via community engagement (i.e., postings in the community). The finding is important, as previous research (e.g., Liu, 2006) suggested that CRs valence can affect online users’ expectations of a film and thus influence the URs of the film that are generated. However, this impact may be insignificant if online users do not actually read professional reviews. Conversely, Zhou and Duan (2016) showed that in the software download context, CRs valence has a positive impact on URs volumes. The use of different data sources (i.e., online or offline) may lead to inconsistent results for the impact of CRs valence on URs volumes. To overcome this problem, this study collected both critical and consumer review data from the website IMDb, where consumers can easily obtain both CRs and URs of films. Our study found that CRs valence may also be indirectly related to URs volume through users’ community engagement. Furthermore, we found that the indirect effect of CRs valence on URs volume was greater than its direct effect. This indirect relationship might be due to the fact that users often actively participate in posting in the film community to increase their self-enhancement (Hennig-Thurau et al., 2004). Hennig-Thurau et al. (2004) has shown that self-enhancement has a significant impact on the number of comments on online opinion platforms. Therefore, psychological motivation theory implies that users are willing to write reviews on products.

In addition, we found a marginally significant direct positive relationship between CRs valence and box office revenue when controlling for community engagement and URs volume as intervening variables. The β value of 0.10 indicated partial rather than complete mediation, and the multiplicative indirect effect (0.26) reached significance. Thus, community engagement and URs volume remain important antecedents of box office revenue. The results of previous empirical studies of the effects of CRs valence on box office revenue are inconsistent. Eliashberg and Shugan (1997) showed that professional reviews are related to cumulative box office revenue, whereas Hennig-Thurau et al. (2006) reported that CRs valence have a non-significant impact on long-term box office. The causes of these discrepancies remain to be identified. Our results imply that, due to dilution by mediation via community engagement and URs volume, the overall correlation between CRs valence and box office revenue is weak. This finding has implications for the eWOM literature. For example, Holbrook and Addis (2007) concluded that in the case of movie, the relationship between evaluative expert judgment (evaluations of excellence by reviewers) and popular appeal (WOM or market performance) tends to be significant but weak, typically accounting for less than 10% of the variance in popularity or market performance. This study adds to the literature and shows that user posting behavior in the community, in the form of enhanced interaction, is positively associated with box office revenue (β = 0.30, p < 0.001; see Figure 2). User posting behavior, together with URs volume and the direct path of CRs valence, accounted for 60% of the variance in box office revenue.

The eWOM plays an important role in the consumer decision making process, and many firms allocate significant marketing budgets to manage their eWOM. Based on our findings, we offer practical implication in social media marketing strategy for movie distributors and studios, as well as provide recommendations for online review platforms. Since we have demonstrated that CRs and URs have different impact effects on sales, those distributors or studios can more strategically prioritize their investments. They may provide incentives for critics or users to focus on different forms of comments and suggestions. In addition, a distributor or studio who hope to increase box office performance could take actions to enhance film community interaction. Potential strategies include enhancing the significance of member participation by ensuring that users understand that their posts could significantly affect the purchase choices of others (Oh et al., 2017). In addition, we strongly urge managers to pay more attention to online communities as an additional monitoring tool for film promotion. By facilitating interactions and communication between potential consumers and existing consumers, user community engagement behavior can increase sales.

Another important practical implication of this study is for the online review platform, our main finding is that in an online world where CRs and URs coexist. CRs can influence URs and further contribute to movie revenues. However, many online review platforms currently rely on user-generated content, and expert reviews are often ignored. In recent years, firms in product categories such as automobiles and electronics have often utilized favorable reviews from professionals and experts on their business information (Eliashberg and Shugan, 1997). In addition, many video game developers use professional reviews to design new products, make sales forecasts, and develop business strategies (Chen et al., 2012). Recent research also claims that CRs remain an influential factor on movie revenues (Basuroy et al., 2020). Therefore, these platforms can design their websites to obtain both types of opinions (i.e., CRs and URs) on the same platform to more effectively influence consumer decisions.

Limitations and Future Research Directions

This research is subject to limitations that also provide possible directions for future studies. First, this study only considered the volume of online user reviews and did not take into account their valence. The main reason for this choice was that there are no consistent results on the relationship between online user review valence and consumer choice; some studies have shown a positive influence of consumer rating scores on other consumers’ choices (Chevalier and Mayzlin, 2006; Zhu and Zhang, 2010; Chen et al., 2011), whereas others have shown that consumer rating scores do not have a significant effect on consumer choice (Liu, 2006; Baum and Spann, 2014). Some scholars have recently suggested that this discrepancy may be due to the study of the moderation effects of contextual factors (Zhu and Zhang, 2010; Ma et al., 2013). Future discussions and in-depth comparisons could examine whether the impact of consumer ratings varies depending on product type (e.g., books, music) (Zhu et al., 2017) or product popularity (Zhu and Zhang, 2010). Second, we included only eWOM volume and did not analyze review content. It would be worthwhile to explore whether sentiment analysis using a text mining procedure or natural language processing techniques can produce richer consumer information and suggest business strategies (Ludwig et al., 2013; Moon and Song, 2015; Zhu et al., 2017). Textual content can provide additional emotional information about reviews, and different emotional content will have different impact effects (Salehan and Kim, 2016; Ullah et al., 2016), which suggests that integration with textual content can make the eWOM literature more complete and richer. Third, all of the empirical eWOM data were collected from IMDb.com and Metacritic.com. Both websites are focused on the United States, which means that our findings may only reflect the impact of United States movie eWOM on consumer choice. Including data from more websites would allow quantitative measurement and increase the generalizability of the research results.

Data Availability Statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author Contributions

All authors listed have made a substantial, direct, and intellectual contribution to the work, and approved it for publication.

Funding

This work was supported by the National Natural Science Foundation of China (NSFC No. 72103083).

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Acknowledgments

We would like to thank the editor and four reviewers for their helpful and valuable comments that greatly improved the manuscript.

Footnotes

  1. ^ www.imdb.com
  2. ^ www.boxofficemojo.com
  3. ^ www.metacritic.com
  4. ^ Indirect effect = 0.46 × 0.30 + 0.46 × 0.78 × 0.22 + 0.18 × 0.22 = 0.26.

References

Akdeniz, M. B., and Talay, M. B. (2013). Cultural variations in the use of marketing signals: a multilevel analysis of the motion picture industry. J. Acad. Mark. Sci. 41, 601–624. doi: 10.1007/s11747-013-0338-5

CrossRef Full Text | Google Scholar

Algesheimer, R., Dholakia, U. M., and Herrmann, A. (2005). The social influence of brand community: evidence from European car clubs. J. Mark. 69, 19–34. doi: 10.1509/jmkg.69.3.19.66363

PubMed Abstract | CrossRef Full Text | Google Scholar

Amblee, N., and Bui, T. (2007). “Freeware downloads: an empirical investigation into the impact of expert and user reviews on demand for digital goods,” in Proceedings of the AMCIS 2007, Keystone, CO.

Google Scholar

Banerjee, A. V. (1992). A simple model of herd behavior. Q. J. Econ. 107, 797–817. doi: 10.2307/2118364

CrossRef Full Text | Google Scholar

Basuroy, S., Abraham Ravid, S., Gretz, R. T., and Allen, B. J. (2020). Is everybody an expert? An investigation into the impact of professional versus user reviews on movie revenues. J. Cult. Econ. 44, 57–96. doi: 10.1007/s10824-019-09350-7

CrossRef Full Text | Google Scholar

Basuroy, S., Chatterjee, S., and Ravid, S. A. (2003). How critical are critical reviews? The box office effects of film critics, star power, and budgets. J. Mark. 67, 103–117. doi: 10.1509/jmkg.67.4.103.18692

PubMed Abstract | CrossRef Full Text | Google Scholar

Basuroy, S., Desai, K. K., and Talukdar, D. (2006). An empirical investigation of signaling in the motion picture industry. J. Mark. Res. 43, 287–295. doi: 10.1509/jmkr.43.2.287

PubMed Abstract | CrossRef Full Text | Google Scholar

Baum, D., and Spann, M. (2014). The interplay between online consumer reviews and recommender systems: an experimental analysis. Int. J. Electron. Commerce 19, 129–162. doi: 10.2753/jec1086-4415190104

CrossRef Full Text | Google Scholar

Boatwright, P., Basuroy, S., and Kamakura, W. (2007). Reviewing the reviewers: the impact of individual film critics on box office performance. Quant. Mark. Econ. 5, 401–425. doi: 10.1007/s11129-007-9029-1

CrossRef Full Text | Google Scholar

Brown, J. J., and Reingen, P. H. (1987). Social ties and word-of-mouth referral behavior. J. Consum. Res. 14, 350–362. doi: 10.1086/209118

CrossRef Full Text | Google Scholar

Chen, Y., Liu, Y., and Zhang, J. (2012). When do third-party product reviews affect firm value and what can firms do? The case of media critics and professional movie reviews. J. Mark. 76, 116–134. doi: 10.1509/jm.09.0034

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, Y., Wang, Q., and Xie, J. (2011). Online social interactions: a natural experiment on word of mouth versus observational learning. J. Mark. Res. 48, 238–254. doi: 10.1509/jmkr.48.2.238

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, Y., and Xie, J. (2005). Third-party product review and firm marketing strategy. Mark. Sci. 24, 218–240. doi: 10.1287/mksc.1040.0089

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, Y., and Xie, J. (2008). Online consumer review: word-of-mouth as a new element of marketing communication mix. Manage. Sci. 54, 477–491. doi: 10.1287/mnsc.1070.0810

PubMed Abstract | CrossRef Full Text | Google Scholar

Chevalier, J. A., and Mayzlin, D. (2006). The effect of word of mouth on sales: online book reviews. J. Mark. Res. 43, 345–354. doi: 10.1509/jmkr.43.3.345

PubMed Abstract | CrossRef Full Text | Google Scholar

Duan, W., Gu, B., and Whinston, A. B. (2008). The dynamics of online word-of-mouth and product sales—an empirical investigation of the movie industry. J. Retail. 84, 233–242. doi: 10.1016/j.jretai.2008.04.005

CrossRef Full Text | Google Scholar

Elberse, A., and Eliashberg, J. (2003). Demand and supply dynamics for sequentially released products in international markets: the case of motion pictures. Mark. Sci. 22, 329–354. doi: 10.1287/mksc.22.3.329.17740

PubMed Abstract | CrossRef Full Text | Google Scholar

Eliashberg, J., and Shugan, S. M. (1997). Film critics: influencers or predictors? J. Mark. 61, 68–78. doi: 10.2307/1251831

CrossRef Full Text | Google Scholar

Godes, D., and Mayzlin, D. (2004). Using online conversations to study word-of-mouth communication. Mark. Sci. 23, 545–560. doi: 10.1287/mksc.1040.0071

PubMed Abstract | CrossRef Full Text | Google Scholar

Goes, P. B., Lin, M., and Au Yeung, C.-M. (2014). “Popularity effect” in user-generated content: evidence from online product reviews. Inf. Syst. Res. 25, 222–238. doi: 10.1016/j.ijmedinf.2019.02.005

PubMed Abstract | CrossRef Full Text | Google Scholar

Gu, B., Park, J., and Konana, P. (2012). Research note—the impact of external word-of-mouth sources on retailer sales of high-involvement products. Inf. Syst. Res. 23, 182–196. doi: 10.1287/isre.1100.0343

CrossRef Full Text | Google Scholar

Haan, M. A., Dijkstra, S. G., and Dijkstra, P. T. (2005). Expert judgment versus public opinion–evidence from the eurovision song contest. J. Cult. Econ. 29, 59–78. doi: 10.1007/s10824-005-6830-0

CrossRef Full Text | Google Scholar

Hennig-Thurau, T., Gwinner, K. P., Walsh, G., and Gremler, D. D. (2004). Electronic word-of-mouth via consumer-opinion platforms: what motivates consumers to articulate themselves on the internet? J. Interact. Mark. 18, 38–52. doi: 10.1002/dir.10073

CrossRef Full Text | Google Scholar

Hennig-Thurau, T., Houston, M. B., and Walsh, G. (2006). The differing roles of success drivers across sequential channels: an application to the motion picture industry. J. Acad. Mark. Sci. 34, 559–575. doi: 10.1177/0092070306286935

CrossRef Full Text | Google Scholar

Hennig-Thurau, T., Wiertz, C., and Feldhaus, F. (2015). Does Twitter matter? The impact of microblogging word of mouth on consumers’ adoption of new movies. J. Acad. Mark. Sci. 43, 375–394. doi: 10.1007/s11747-014-0388-3

CrossRef Full Text | Google Scholar

Holbrook, M. B., and Addis, M. (2007). Taste versus the market: an extension of research on the consumption of popular culture. J. Consum. Res. 34, 415–424. doi: 10.1086/518529

CrossRef Full Text | Google Scholar

Hussain, S., Guangju, W., Jafar, R. M. S., Ilyas, Z., Mustafa, G., and Jianzhou, Y. (2018). Consumers’ online information adoption behavior: motives and antecedents of electronic word of mouth communications. Comput. Hummer Behav. 80, 22–32. doi: 10.1016/j.chb.2017.09.019

CrossRef Full Text | Google Scholar

Hussain, S., Huang, K., Ilyas, Z., and Niu, B. (2020a). Exploring the novel input attributes affecting eWOM. Front. Psychol. 11:2017. doi: 10.3389/fpsyg.2020.02017

PubMed Abstract | CrossRef Full Text | Google Scholar

Hussain, S., Song, X., and Niu, B. (2020b). Consumers’ motivational involvement in eWOM for information adoption: the mediating role of organizational motives. Front. Psychol. 10:3055. doi: 10.3389/fpsyg.2019.03055

PubMed Abstract | CrossRef Full Text | Google Scholar

Joshi, A. M., and Hanssens, D. M. (2009). Movie advertising and the stock market valuation of studios: a case of “great expectations?” Mark. Sci. 28, 239–250. doi: 10.1287/mksc.1080.0392

PubMed Abstract | CrossRef Full Text | Google Scholar

Koh, J., Kim, Y.-G., Butler, B., and Bock, G.-W. (2007). Encouraging participation in virtual communities. Commun. ACM 50, 69–73. doi: 10.1145/1216016.12160

CrossRef Full Text | Google Scholar

Lee, F. L. (2006). Cultural discount and cross-culture predictability: examining the box office performance of American movies in Hong Kong. J. Media Econ. 19, 259–278. doi: 10.1207/s15327736me1904_3

CrossRef Full Text | Google Scholar

Leenders, M. A., and Eliashberg, J. (2011). The antecedents and consequences of restrictive age-based ratings in the global motion picture industry. Int. J. Res. Mark. 28, 367–377. doi: 10.1016/j.ijresmar.2011.06.001

CrossRef Full Text | Google Scholar

Liu, Y. (2006). Word of mouth for movies: its dynamics and impact on box office revenue. J. Mark. 70, 74–89. doi: 10.1509/jmkg.70.3.074

PubMed Abstract | CrossRef Full Text | Google Scholar

Ludwig, S., De Ruyter, K., Friedman, M., and Brüggen, E. C. (2013). More than words: the influence of affective content and linguistic style matches in online reviews on conversion rates. J. Mark. 77, 87–103. doi: 10.1509/jm.11.0560

PubMed Abstract | CrossRef Full Text | Google Scholar

Ma, X., Khansa, L., Deng, Y., and Kim, S. S. (2013). Impact of prior reviews on the subsequent review process in reputation systems. J. Manage. Inf. Syst. 30, 279–310.

Google Scholar

Moe, W. W., and Schweidel, D. A. (2012). Online product opinions: incidence, evaluation, and evolution. Mark. Sci. 31, 372–386. doi: 10.1287/mksc.1110.0662

PubMed Abstract | CrossRef Full Text | Google Scholar

Moe, W. W., and Trusov, M. (2011). The value of social dynamics in online product ratings forums. J. Mark. Res. 48, 444–456. doi: 10.1509/jmkr.48.3.444

PubMed Abstract | CrossRef Full Text | Google Scholar

Moon, S., Bergey, P. K., and Iacobucci, D. (2010). Dynamic effects among movie ratings, movie revenues, and viewer satisfaction. J. Mark. 74, 108–121. doi: 10.1509/jmkg.74.1.108

PubMed Abstract | CrossRef Full Text | Google Scholar

Moon, S., and Song, R. (2015). The roles of cultural elements in international retailing of cultural products: an application to the motion picture industry. J. Retail. 91, 154–170. doi: 10.1016/j.jretai.2014.12.002

CrossRef Full Text | Google Scholar

Mudambi, S. M., and Schuff, D. (2010). What makes a helpful online review? A study of customer reviews on amazon.com. MIS Q. 34, 185–200. doi: 10.1089/cyber.2012.0417

PubMed Abstract | CrossRef Full Text | Google Scholar

Oestreicher-Singer, G., and Zalmanson, L. (2013). Content or community? A digital business strategy for content providers in the social age. MIS Q. 37, 591–616. doi: 10.25300/misq/2013/37.2.12

CrossRef Full Text | Google Scholar

Oh, C., Roumani, Y., Nwankpa, J. K., and Hu, H.-F. (2017). Beyond likes and tweets: consumer engagement behavior and movie box office in social media. Inf. Manage. 54, 25–37. doi: 10.1016/j.im.2016.03.004

CrossRef Full Text | Google Scholar

Petty, R. E., and Cacioppo, J. T. (1984). The effects of involvement on responses to argument quantity and quality: central and peripheral routes to persuasion. J. Pers. Soc. Psychol. 46, 69–81. doi: 10.1037/0022-3514.46.1.69

CrossRef Full Text | Google Scholar

Podium (2017). State of Online Reviews. Available online at: http://learn.podium.com/rs/841-BRM-380/images/Podium-2017-State-of-Online-Reviews.pdf (accessed May 2020).

Google Scholar

Ravid, S. A. (1999). Information, blockbusters, and stars: a study of the film industry. J. Bus. 72, 463–492. doi: 10.1086/209624

CrossRef Full Text | Google Scholar

Reinstein, D. A., and Snyder, C. M. (2005). The influence of expert reviews on consumer demand for experience goods: a case study of movie critics. J. Ind. Econ. 53, 27–51. doi: 10.1111/j.0022-1821.2005.00244.x

CrossRef Full Text | Google Scholar

Salehan, M., and Kim, D. J. (2016). Predicting the performance of online consumer reviews: a sentiment mining approach to big data analytics. Decis. Support Syst. 81, 30–40. doi: 10.1016/j.dss.2015.10.006

CrossRef Full Text | Google Scholar

Sridhar, S., and Srinivasan, R. (2012). Social influence effects in online product ratings. J. Mark. 76, 70–88. doi: 10.1509/jm.10.0377

PubMed Abstract | CrossRef Full Text | Google Scholar

Sun, M. (2012). How does the variance of product ratings matter? Manage. Sci. 58, 696–707. doi: 10.1287/mnsc.1110.1458

PubMed Abstract | CrossRef Full Text | Google Scholar

Ullah, R., Amblee, N., Kim, W., and Lee, H. (2016). From valence to emotions: exploring the distribution of emotions in online product reviews. Decis. Support Syst. 81, 41–53.

Google Scholar

Wang, F., Liu, X., and Fang, E. E. (2015). User reviews variance, critic reviews variance, and product sales: an exploration of customer breadth and depth effects. J. Retail. 91, 372–389.

Google Scholar

Wiles, M. A., and Danielova, A. (2009). The worth of product placement in successful films: an event study analysis. J. Mark. 73, 44–63. doi: 10.1509/jmkg.73.4.044

PubMed Abstract | CrossRef Full Text | Google Scholar

Wu, J., Fan, S., and Zhao, J. L. (2018). Community engagement and online word of mouth: an empirical investigation. Inf. Manage. 55, 258–270.

Google Scholar

Wu, J., Huang, L., Zhao, J. L., and Hua, Z. (2015). The deeper, the better? Effect of online brand community activity on customer purchase frequency. Inf. Manage. 52, 813–823. doi: 10.1016/j.im.2015.06.001

CrossRef Full Text | Google Scholar

Zheng, X., Cheung, C. M., Lee, M. K., and Liang, L. (2015). Building brand loyalty through user engagement in online brand communities in social networking sites. Inf. Technol. People 28, 90–106. doi: 10.1108/ITP-08-2013-0144

CrossRef Full Text | Google Scholar

Zhou, W., and Duan, W. (2016). Do professional reviews affect online user choices through user reviews? An empirical study. J. Manage. Inf. Syst. 33, 202–228.

Google Scholar

Zhu, D. H., Ye, Z. Q., and Chang, Y. P. (2017). Understanding the textual content of online customer reviews in B2C websites: a cross-cultural comparison between the US and China. Comput. Hum. Behav. 76, 483–493. doi: 10.1016/j.chb.2017.07.045

CrossRef Full Text | Google Scholar

Zhu, F., and Zhang, X. (2010). Impact of online consumer reviews on sales: the moderating role of product and consumer characteristics. J. Mark. 74, 133–148. doi: 10.1509/jm.74.2.133

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: critical reviews, online user reviews, community engagement, box office revenues, electronic word-of-mouth

Citation: Chiu Y-L, Du J, Sun Y and Wang J-N (2022) Do Critical Reviews Affect Box Office Revenues Through Community Engagement and User Reviews? Front. Psychol. 13:900360. doi: 10.3389/fpsyg.2022.900360

Received: 20 March 2022; Accepted: 25 April 2022;
Published: 26 May 2022.

Edited by:

Kittisak Jermsittiparsert, University of City Island, Cyprus

Reviewed by:

Safdar Hussain, Pir Mehr Ali Shah Arid Agriculture University, Pakistan
Hassaporn Thongdaeng, Sukhothai Thammathirat Open University, Thailand
Adipon Euajarusphan, Thammasat University, Thailand
Warat Karuchit, National Institute of Development Administration, Thailand

Copyright © 2022 Chiu, Du, Sun and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Jiangze Du, jiangze.du@hotmail.com

These authors have contributed equally to this work

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.