# Generalized Structured Component Analysis with Uniqueness Terms for Accommodating Measurement Error

^{1}Department of Psychology, McGill University, Montreal, QC, Canada^{2}Department of Psychology, University of Victoria, BC, Canada^{3}Department of Educational Psychology and Leadership, Texas Tech University, Lubbock, TX, United States

Generalized structured component analysis (GSCA) is a component-based approach to structural equation modeling (SEM), where latent variables are approximated by weighted composites of indicators. It has no formal mechanism to incorporate errors in indicators, which in turn renders components prone to the errors as well. We propose to extend GSCA to account for errors in indicators explicitly. This extension, called GSCA_{M}, considers both common and unique parts of indicators, as postulated in common factor analysis, and estimates a weighted composite of indicators with their unique parts removed. Adding such unique parts or uniqueness terms serves to account for measurement errors in indicators in a manner similar to common factor analysis. Simulation studies are conducted to compare parameter recovery of GSCA_{M} and existing methods. These methods are also applied to fit a substantively well-established model to real data.

## Introduction

Structural equation modeling (SEM) involves the specification and testing of the relationships between variables that are observed (indicators) and unobserved (latent variables). Two approaches have been proposed for SEM: Factor-based vs. component-based (e.g., Fornell and Bookstein, 1982; Jöreskog and Wold, 1982; Tenenhaus, 2008; Rigdon, 2012). The former includes covariance structure analysis (CSA; Jöreskog, 1970, 1973), and the latter includes partial least squares path modeling (PLSPM; Wold, 1966, 1973, 1982; Lohmöller, 1989) and generalized structured component analysis (GSCA; Hwang and Takane, 2004, 2014). As their names imply, the two approaches become divergent in how they approximate latent variables in these sub-models. That is, factor-based SEM assumes that common factors may approximate latent variables as in common factor analysis, whereas component-based SEM posits that weighted composites of indicators may serve as proxies for latent variables as in principal component analysis. In this regard, the two approaches are conceptually different, and choices on them would likely be application-dependent and should be theoretically decided in advance, considering how to conceptualize latent variables in the application as well as how to validate the relation between latent variables and their proxies (factors or components; Rigdon, 2012).

Nonetheless, in comparison with factor-based SEM, perhaps the most common criticism of component-based SEM has been that it has no mechanism to formally take into account errors in indicators, which appear practically inevitable in the social sciences (e.g., Bentler and Huang, 2014). This also leads components to take over these errors to some extent, although extracting a weighted composite from a set of indicators can play a role in reducing the errors implicitly (Gleason, 1973). It is well-known that errors in independent (observed or latent) variables will likely result in biased parameter estimates (e.g., Bollen, 1989, Chapter 5).

To deal with this problem, a bias-correction method, called consistent partial least squares (PLSc; Dijkstra, 2010; Dijkstra and Henseler, 2015), has been proposed in the context of PLSPM. The basic premise of PLSc is that the true measurement model is the so-called basic design (Wold, 1982), representing a unidimensional confirmatory factor analytic model, where each latent variable/factor underlies at least two indicators with each indicator loading on one and only one latent variable. Under this assumption, PLSc begins to apply PLSPM to estimate component weights for a set of indicators per latent variable and then obtains a correction constant for the set of indicators based on their component weights and sample correlations. The loadings for the indicators are estimated by multiplying their component weight estimates by the correction constant. The correlations among latent variables are also estimated using the correction constants for all sets of indicators, which are subsequently used for estimating path coefficients. Conceptually, it is somewhat arbitrary whether PLSc falls into component-based SEM because it has little interest in the specification and estimation of components *per se*, and simply utilizes their weight estimates to obtain parameter estimates of factor-based SEM. In practice, the assumption of the basic design can be restrictive, leading to the exclusion of cross loadings that have been well-accepted in numerous structural equation models (Asparouhov and Muthén, 2009). For example, a classical model involving cross loadings is a multitrait-multimethod model, where trait and method latent variables underlie each indicator (Campbell and Fiske, 1959). Another example is latent growth curve models (Meredith and Tisak, 1990; Duncan et al., 2006), where indicators are typically assumed to load on multiple latent variables, each of which captures a different trajectory of change over time.

To our knowledge, no attempts have been made to incorporate errors in indicators or develop a bias-correction strategy in the context of GSCA. Thus, in this paper, we propose to extend GSCA to explicitly account for errors in indicators. Specifically, we aim to combine a unique part of each indicator into GSCA. As postulated in common factor analysis or factor-based SEM, adding such a unique part may be seen as accounting for measurement error in each indicator. We shall call this proposed extension “GSCA_{M},” standing for GSCA with measurement errors incorporated. GSCA_{M} will provide parameter estimates comparable to those from factor-based SEM. Whereas PLSc involves two separate estimation steps, GSCA_{M} has a single estimation procedure where a least squares criterion is consistently minimized to estimate all model parameters. In addition, GSCA_{M} does not require the basic design assumption in model specification and parameter estimation.

The paper is organized as follows. Section Method discusses the technical underpinnings of GSCA_{M}, including model specification and parameter estimation. Section Simulation Studies conducts a simulation study to evaluate the performance of GSCA_{M} and two existing methods—CSA and PLSc. Section An Empirical Application presents an application to show the empirical usefulness of GSCA_{M} as compared to the existing methods. The final section summarizes the implications of the proposed method.

## Method

### Model

As with GSCA, GSCA_{M} involves three sub-models—measurement, structural, and weighted relation. The measurement model is used to specify the relationships between indicators and latent variables, whereas the structural model is to specify the relationships among latent variables. The weighted relation model is used to express a latent variable as a weighted composite of indicators. Unlike GSCA, however, GSCA_{M} contemplates both common and unique parts of each indicator in the measurement model, and expresses a latent variable as a weighted composite of indicators with their unique parts removed in the weighted relation model.

Let **Z** = [**z**_{1},…, **z**_{J}] denote an *N* by *J* matrix of indicators, where *N* is the number of observations, and **z**_{j} is the *j*th indicator (*j* = 1,…, *J*). Let **Γ** = [**γ**_{1},…, **γ**_{P}] denote an *N* by *P* matrix of latent variables, where **γ**_{p} is the *p*th latent variable (*p* = 1,…, *P*). Assume that all indicators and latent variables are normalized such that their lengths are equal to one (i.e., ${z}_{j}{}^{\prime}{z}_{j}={\gamma}_{p}{}^{\prime}{\gamma}_{p}=1$). Let **C** denote a *P* by *J* matrix of loadings relating latent variables to indicators. Let **U** denote an *N* by *J* matrix of unique variables. Let **D** denote a *J* by *J* diagonal matrix of unique loadings. Let **E**_{1} denote an *N* by *J* matrix of residuals for indicators. Let **B** denote a *P* by *P* matrix of path coefficients connecting latent variables among themselves, and **E**_{2} denote an *N* by *P* matrix of residuals for latent variables. The three sub-models of GSCA_{M} are given as follows.

In the measurement model (1), **ΓC** and **UD** represent common and unique parts of indicators, respectively. We assume that **Γ** is uncorrelated with **U** (**U**′**Γ** = **Γ**′**U** = **0**) and **U** is orthonormalized ($U\left({U}^{\prime}\mathbf{\Gamma}={\mathbf{\Gamma}}^{\prime}U=0\right)$, where **I**_{J} is the identity matrix of order *J*). In the measurement model, the **C** matrix contains fixed values (e.g., zeros) to accommodate hypothesized relationships between indicators and their latent variables as in confirmatory factor analytic models. If this matrix has no fixed values, (1) may be seen as the fixed exploratory factor analytic model (Young, 1941; de Leeuw, 2004, 2008). The structural model (2) remains the same as that for GSCA or the reticular action model (McArdle and McDonald, 1984). The weighted relation model (3) shows that a latent variable is defined as a weighted composite of indicators with their unique parts eliminated.

GSCA_{M} integrates the sub-models into a single equation, as follows.

where **Ψ** = [**Z**, **Γ**], **A** = [**C**, **B**], **S** = [**UD**, **0**], and **E** = [**E**_{1}, **E**_{2}]. This is called the GSCA_{M} model.

### Parameter Estimation

The parameters of GSCA_{M} (**Γ**, **A**, **U**, and **D**) are estimated by minimizing the following least squares criterion

subject to ${\mathbf{\gamma}}_{p}^{\mathbf{\prime}}{\gamma}_{p}=1$or equivalently diag(**Γ**′**Γ**) = **I**_{P}, **U**′**Γ** = **0**, and $\left({\Gamma}^{\prime}\mathbf{\Gamma}\right)={I}_{P},{U}^{\prime}\mathbf{\Gamma}=0$, where SS(**X**) = tr(**X**′**X**), and **I**_{P} is the identity matrix of order *P*.

A simple iterative algorithm is developed to minimize (5). This algorithm begins by assigning initial values to the parameters. Then, it alternates several steps until convergence, each of which updates a set of parameters in a least squares sense, with the other sets fixed. A detailed description of the algorithm is provided in the Appendix in Supplementary Material.

We can apply GSCA to obtain initial values for **Γ**, **C**, and **B**, although any other initial values can be considered. Then, those for **U** and **D** may be obtained as described in Steps 3 and 4 in the Appendix in Supplementary Material. We can employ the bootstrap method (Efron, 1979) to estimate the standard errors or confidence intervals of the parameter estimates without resorting to a distributional assumption such as multivariate normality of indicators. When the number of observations is smaller than that of indicators (*N* < *J*), rank(**U**) < *J* and the constraint ${U}^{\prime}U={I}_{J}$ cannot be fulfilled. In this situation, we may apply Unkel and Trendafilov's (2013) algorithm to update **U**, subject to the new constraint **U**′**UD** = **D** (also see Trendafilov and Unkel, 2011). We assume that all indicators and latent variables are normalized, which still results in standardized parameter estimates except for latent variable scores that are normalized. The standardized latent variable scores are obtained by multiplying the normalized scores by $\sqrt{N}$.

GSCA_{M} can provide a measure of overall model fit, called FIT. The FIT indicates the total variance of all variables explained by a particular model specification. It is given by

The values of the FIT range from 0 to 1. The larger this value, the more variance in the variables is accounted for by the specified model. Moreover, it can provide separate model fit measures for the measurement and structural models, as follows.

The FIT_{M} shows how much the variance of indicators is explained by a measurement model, whereas the FIT_{S} indicates how much the variance of latent variables is accounted for by a structural model. Both measures range from 0 to 1 and can be interpreted in a manner similar to the FIT.

## Simulation Studies

We conducted simulation studies to evaluate the performance of GSCA_{M} as compared to existing methods, including GSCA, PLSc, and CSA. In particular, we focused on comparing GSCA_{M} to these methods in parameter recovery.

### Simulation Study 1

We specified a structural equation model that consisted of three latent variables and three indicators per latent variable. Figure 1 displays the specified model along with their unstandardized and standardized parameter values.

**Figure 1**. The structural equation model specified for the first simulation study. Standardized parameters are given in parentheses.

For this study, we considered four levels of sample size: *N* = 100, 200, 500, and 1,000. At each sample size, we generated 1,000 random samples from a multivariate normal distribution with zero means and the covariance matrix implied by the unstandardized parameters of the correct model, based on a CSA formulation. We used the maximum likelihood method for CSA. For PLSc, we used Mode A and the path weighting scheme that is preferred over the other schemes (centroid and factorial) in estimating component weights (Esposito Vinzi et al., 2010). We used the same random seed for all the methods for each sample to have them run with the same initial values.

Table 1 exhibits the relative biases (expressed as percentages), standard deviations, and mean square errors of the standardized loadings and path coefficients estimated from the four methods over the different sample sizes. In the calculations of these properties, we removed any sample involving non-convergence within 300 iterations or convergence to improper solutions. CSA encountered such convergence problems across all the sample sizes. The number of the samples omitted with the convergence problems under CSA was as follows: 66 (*N* = 100), 46 (*N* = 200), 31 (*N* = 500), and 31 (*N* = 1,000). PLSc was faced with the problems only when *N* ≤ 200. Specifically, 23 and 2 samples were omitted when *N* = 100 and 200, respectively. Conversely, GSCA and GSCA_{M} did not encounter non-convergence or the occurrence of improper solutions across all the sample sizes.

**Table 1**. Relative biases expressed as percentages (RB(%)), standard deviations (SD), and mean square errors (MSE) of standardized loadings and path coefficients obtained from GSCA, GSCA_{M}, PLSc, and CSA over different sample sizes.

We regarded relative bias >10% in absolute value as indicative of an unacceptable degree of bias (e.g., Bollen et al., 2007; Lei and Wu, 2012). As shown in Table 1, GSCA provided positively biased loading estimates and negatively biased path coefficient estimates across all the sample sizes. This is expected because the simulated data were generated based on a CSA formulation, assuming that a latent variable was equivalent to a common factor. In this case, component-based approaches to SEM, such as GSCA and PLSPM, are known to overestimate loadings and underestimate path coefficients (e.g., Velicer and Jackson, 1990; Dijkstra, 2010; Hwang et al., 2010; Sarstedt et al., 2016). Conversely, GSCA_{M}, PLSc, and CSA tended to result in unbiased estimates of both loadings and path coefficients across the sample sizes. When *N* = 100, however, the path coefficient estimates of GSCA_{M} showed larger relative biases (4–5%) than those from PLSc and CSA, although they decreased rapidly with the sample size, approaching essentially zero when *N* = 1,000.

The standard deviations of the estimates from the four methods became smaller with the sample size. However, GSCA provided smaller standard deviations than the other methods. This was particularly salient when the sample size was small (e.g., *N* = 100), which is consistent with the literature (Hwang et al., 2010). When the sample size was small, all the methods tended to show similar mean square errors of the loading estimates, whereas GSCA tended to provide larger mean square errors of the path coefficient estimates. As the sample size increased, the mean square errors of all the parameter estimates obtained from GSCA_{M}, PLSc, and CSA approached zero, while those from GSCA remained slightly larger.

### Simulation Study 2

The first simulation study was useful to evaluate how GSCA_{M} performed as compared to different methods. Nonetheless, this study considered a model with equal loadings, which might be too ideal in reality. Thus, we conducted another simulation study to compare the performance of the methods under a model with the same structure but unequal loadings varying from 0.4 to 0.8. We also compared their performance given a misspecification of the model. Figure 2 displays both correct and misspecified models along with their unstandardized and standardized parameter values. The misspecified model involved two cross loadings and an additional path coefficient, as indicated by dashed arrows in Figure 2.

**Figure 2**. The structural equation model specified for the second simulation study. Standardized parameters are given in parentheses. A model misspecification involves addition of two cross loadings and a path coefficient as indicated by dashed arrows.

For both specifications, we considered the same four levels of sample size, at each of which 1,000 random samples were generated from a multivariate normal distribution with zero means and the covariance matrix implied by the unstandardized parameters of the correct model, based on a CSA formulation. Again, we used the maximum likelihood method for CSA, and Mode A and the path weighting scheme for PLSc. We applied all the four methods to estimate the parameters of the correct model, whereas applied only GSCA, GSCA_{M}, and CSA to estimate the parameters of the misspecified model because as stated earlier, PLSc was not designed for models involving cross loadings.

Table 2 presents the relative biases (expressed as percentages), standard deviations, and mean square errors of the standardized loadings and path coefficients estimated from the four methods under the correct model specification. GSCA did not encounter non-convergence or the occurrence of improper solutions across all the sample sizes. Conversely, the numbers of the samples omitted with the convergence problems under CSA at the different sample sizes were 72 (*N* = 100), 41 (*N* = 200), 30 (*N* = 500), and 23 (*N* = 1000), whereas those under PLSc were 38 (*N* = 100) and 3 (*N* = 200). One sample was removed under GSCA_{M} only when *N* = 100.

**Table 2**. Relative biases expressed as percentages (RB(%)), standard deviations (SD), and mean square errors (MSE) of standardized loadings and path coefficients obtained from GSCA, GSCA_{M}, PLSc, and CSA over different sample sizes.

As expected, all the loading estimates from GSCA except those of two high loadings (0.8) were positively biased, whereas all the path coefficient estimates were negatively biased, regardless of the sample sizes. Conversely, overall, GSCA_{M}, PLSc, and CSA tended to result in unbiased estimates of a majority of loadings and all path coefficients across the sample sizes. However, the estimates of two low loadings (0.4 or 0.5) under these methods remained similar in magnitude and positively biased even when *N* = 1,000. As in the first simulation study, when *N* = 100, the path coefficient estimates of GSCA_{M} showed larger relative biases than those from PLSc and CSA, although they decreased rapidly with the sample size, approaching zero when *N* = 1,000.

The standard deviations of the estimates from the four methods became smaller with the sample size. GSCA provided smaller standard deviations than the other methods. The mean square errors of all the parameter estimates obtained from GSCA_{M}, PLSc, and CSA remained similar in magnitude across the sample sizes and most of them, except for those for the estimates of the two loadings, approached zero when the sample size increased. On the other hand, those from GSCA remained larger and only a few approached zero, although they gradually decreased with the sample size.

Table 3 shows the relative biases (expressed as percentages), standard deviations, and mean square errors of the standardized loadings and path coefficients estimated from GSCA, GSCA_{M}, and CSA under the incorrect model specification. CSA suffered severely from non-convergence or convergence to improper solutions across all the sample sizes. The numbers of the samples omitted under CSA were 371 (*N* = 100), 216 (*N* = 200), 131 (*N* = 500), and 100 (*N* = 1,000), whereas those under GSCA_{M} were 54 (*N* = 100), 12 (*N* = 200), 1 (*N* = 500), and 1 (*N* = 1,000). Again, GSCA had no such problems across all the sample sizes.

**Table 3**. Relative biases expressed as percentages (RB(%)), standard deviations (SD), and mean square errors (MSE) of standardized loadings and path coefficients obtained from GSCA, GSCA_{M}, PLSc, and CSA over different sample sizes.

Overall, GSCA_{M} and CSA tended to produce unbiased estimates of most of the loadings and all the path coefficients across the sample sizes. However, their estimates of two low loadings (0.4 or 0.5) remained similar in magnitude and positively biased even when *N* = 1,000. GSCA resulted in positively biased loading estimates except those of two high loadings (0.8) regardless of the sample sizes. Conversely, it produced path coefficient estimates with an acceptably small amount of bias, although the amount of bias on average remained unchanged over the sample sizes.

The standard deviations of the estimates from the three methods became smaller with the sample size. Again, GSCA provided smaller standard deviations than the other methods. The mean square errors of all the parameter estimates obtained from GSCA_{M} and CSA were similar in magnitude across the sample sizes and most of them, except for those for the estimates of the two loadings, approached zero when the sample size increased. On the other hand, the mean square errors of the loading estimates from GSCA remained larger and only a few approached zero, although they gradually decreased with the sample size. The mean square errors of the path coefficient estimates from GSCA were comparable to those from GSCA_{M} and CSA across the sample sizes.

To summarize, GSCA_{M} was found to recover the parameters equally well to CSA in both simulation studies that generated data within the factor-analytic framework. When the basic design held for the specified model, PLSc also performed equally well to GSCA_{M} and CSA. In general, GSCA_{M} was less likely to suffer from non-convergence or the occurrence of improper solutions than CSA and PLSc. In particular, CSA tended to suffer from these problems when the sample size was small and/or the model was misspecified, which was consistent with the literature (e.g., Boomsma, 1982, 1985; Anderson and Gerbing, 1984). Conversely, GSCA largely resulted in biased parameter estimates in these simulation studies that were based on the assumption of factor-analytic models.

## An Empirical Application

The present example came from the American customer satisfaction index (ACSI; Fornell et al., 1996) database. The ACSI has been widely used to assess four different levels of customer satisfaction (national-, sector-, industry-, and company-level) in the United States over the past two decades. This example was company-level data collected in 2002 for 152 companies (*N* = 152). Figure 3 displays the ACSI model. We did not display the residual terms associated with all endogenous variables to make the figure concise. As depicted in Figure 3, the ACSI model contains fourteen indicators: z_{1} = customer expectations about overall quality, z_{2} = customer expectations about reliability, z_{3} = customer expectations about customization, z_{4} = overall quality, z_{5} = reliability, z_{6} = customization, z_{7} = price given quality, z_{8} = quality given price, z_{9} = overall customer satisfaction, z_{10} = confirmation of expectations, z_{11} = distance to ideal product or service, z_{12} = formal or informal complaint behavior, z_{13} = repurchase intention, and z_{14} = price tolerance. The measures and scales of the indicators are described in Fornell et al. (1996). This model also involves six latent variables that underlie the 14 indicators, as follows: CE = customer expectations, PQ = perceived quality, PV = perceived value, CS = customer satisfaction, CC = customer complaints, and CL = customer loyalty. The specified relationships in the ACSI model were well-derived from previous theories, and their detailed conceptual derivations can be found in Fornell et al. (1996).

We applied GSCA_{M}, CSA, and PLSc to fit the ACSI model to the data. We used the R packages lavaan (version 0.5-16) (Rosseel, 2012) to apply CSA and wrote MATLAB codes for GSCA_{M} and PLSc. As in the simulation study, we utilized maximum likelihood for CSA, and Mode A and the path weighting scheme for PLSc.

Note that in the ACSI model, only a single indicator (z_{12}) loads on customer complaints. As discussed earlier, PLSc was developed based on the basic design requiring at least two indicators per latent variable. When there is only one indicator for a latent variable, PLSc cannot estimate its loading and the path coefficients involving the latent variable because the correction constant for the indicator becomes zero (see Dijkstra and Henseler, 2015). Thus, we used the PLSPM estimate of the component weight for the indicator z_{12}, which was equal to one, by fixing the correction constant to one. This was also the case when estimating the path coefficients involving customer complaints (b_{7} and b_{9}), indicating that they might be suboptimal estimates of the path coefficients.

Tables 4, 5 provide the estimates of the standardized loadings and path coefficients of the ACSI model obtained from the methods. CSA yielded a number of improper solutions, including the standardized loading and path coefficient estimates greater than one in absolute value. In addition, the loading estimate for z_{1} (customer expectations about overall quality) was almost zero. This was inconsistent with that the indicator is expected to be highly and positively related to customer expectations (Fornell et al., 1996). PLSc also produced improper solutions, although it resulted in fewer than CSA. We have tried different schemes, but continued to have the same problem. Moreover, the positive signs of the path coefficient estimates involving customer complaints (b_{7} and b_{9}) were substantively contradictory, suggesting that more satisfied customers tended to complain more frequently (b_{7} = 0.46) and more frequent complainers were likely to be more loyal customers (b_{9} = 0.44). These counterintuitive signs were provided, albeit the signs of all the loading estimates remained positive as expected, indicating that the latent variables were not likely to be sign-reversed.

**Table 4**. The estimates of standardized loadings of the ACSI model obtained from CSA, PLSc, and GSCA_{M}.

**Table 5**. The estimates of standardized path coefficients of the ACSI model obtained from CSA, PLSc, and GSCA_{M}.

Conversely, GSCA_{M} resulted in neither improper solutions nor estimates that made little substantive sense. It provided that FIT = 0.85, indicating that the ACSI model accounted for about 85% of the variance of all the variables. Moreover, GSCA_{M} provided that FIT_{M} = 0.98 and FIT_{S} = 0.57. This indicates that the measurement model of the ACSI accounted for about 98% of the variance of the indicators, whereas the structural model explained about 57% of the variance of the latent variables. As also shown in Table 4, all the loading estimates were large and positive. The interpretations of the path coefficient estimates appeared to be generally consistent with those reported in the literature (e.g., Fornell et al., 1996; Anderson and Fornell, 2000). Specifically, customer expectations had a statistically significant impact on perceived quality (b_{1} = 0.93, 95% CI = 0.90 ~ 0.95), but had statistically non-significant effects on perceived value (b_{2} = −0.11, 95% CI = −0.48 ~ 0.22) and customer satisfaction (b_{3} = −0.05, 95% CI = −0.18 ~ 0.05). These non-significant effects were also discussed in previous studies (e.g., Johnson et al., 2001). Perceived quality had statistically significant effects on perceived value (b_{4} = 0.93, 95% CI = 0.54 ~ 1.33) and customer satisfaction (b_{5} = 0.80, 95% CI = 0.66 ~ 0.98). Perceived value had a statistically significant influence on customer satisfaction (b_{6} = 0.27, 95% CI = 0.21 ~ 0.34). Customer satisfaction had statistically significant effects on customer complaints (b_{7} = −0.46, 95% CI = −0.63 ~ −0.30) and customer loyalty (b_{8} = 0.50, 95% CI = 0.39 ~ 0.61). Customer complaints had a statistically significant effect on customer loyalty (b_{9} = −0.45, 95% CI = −0.56 ~ −0.37). We used 100 bootstrap samples for the estimation of the 95% confidence intervals of the GSCA_{M} estimates.

To summarize, in this application, CSA and PLSc yielded improper solutions that were problematic to interpret. The improper solutions may have occurred for reasons. For example, a few latent variables underlie only two indicators each in the ACSI model, the sample size was relatively small, the correlation between customer expectations and perceived quality, which was equivalent to the standardized path coefficient between them (b_{1}), was quite large (>|0.90|), or a combination of these issues (e.g., Chen et al., 2001). In addition, both CSA and PLSc provided estimates that were substantively counterintuitive. For the same data, conversely, GSCA_{M} did not result in improper solutions and its estimates were generally consistent with the hypothesized relationships in the literature.

## Concluding Remarks

We proposed an extension of GSCA, named GSCA_{M}, to explicitly accommodate errors in indicators. As with GSCA, GSCA_{M} can be viewed as a component-based approach to SEM in that it still approximates a latent variable by a component. Unlike GSCA, however, GSCA_{M} considers both common and unique parts of indicators as in factor-based SEM, and estimates a component of indicators with their unique parts excluded. In this way, GSCA_{M} deals with measurement errors in indicators, yielding parameter estimates comparable to those from factor-based SEM. In addition, it does not require a distributional assumption, such as multivariate normality of indicators, for parameter estimation because it estimates parameters via least squares. As a component-based approach, furthermore, it can avoid factor score indeterminacy (e.g., Guttman, 1955; Schönemann and Wang, 1972), enabling to provide unique latent variable scores.

In the simulation studies, GSCA_{M} performed equally well to CSA and PLSc in parameter recovery, when the model was correctly specified to satisfy the basic design assumption. Conversely, when the model was misspecified to contain additional cross loadings and path coefficients, only GSCA_{M} and CSA could be applied to fit the model; and GSCA_{M} tended to recover parameters equally to CSA. In the real data application, GSCA_{M} was the only method that involved no improper solutions.

Although we do not venture into generalizing the results of our analyses, GSCA_{M} may be a promising alternative to CSA, when researchers have difficulty to address such issues as non-convergence or convergence to improper solutions, or are interested in obtaining unique individual latent variable scores for subsequent analyses or modeling of these scores. This can contribute to widening the scope and applicability of GSCA. Nonetheless, it would be fruitful to apply the proposed method to a wide range of real-world problems to investigate its performance more thoroughly.

## Author Contributions

HH contributed to conducing all research activities including technical development, empirical analyses, and manuscript writing; YT contributed to technical development and manuscript writing; and KJ contributed to empirical analyses and manuscript writing.

## Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

## Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyg.2017.02137/full#supplementary-material

## References

Anderson, E. W., and Fornell, C. (2000). Foundations of the American customer satisfaction index. *Tot. Qual. Manag.* 11, S869–S882. doi: 10.1080/09544120050135425

Anderson, J. C., and Gerbing, D. W. (1984). The effect of sampling error on convergence, improper solutions, and goodness-of-fit indices for maximum likelihood confirmatory factor analysis. *Psychometrika* 49, 155–173. doi: 10.1007/BF02294170

Asparouhov, T., and Muthén, B. (2009). Exploratory structural equation modeling. *Struct. Equat. Model*. 16, 397–438. doi: 10.1080/10705510903008204

Bentler, P. M., and Huang, W. (2014). On components, latent variables, PLS and simple methods: reactions to Rigdon's rethinking of PLS. *Long Range Plann.* 47, 38–145. doi: 10.1016/j.lrp.2014.02.005

Bollen, K. A., Kirby, J. B., Curran, P. J., Paxton, P. M., and Chen, F. (2007). Latent variable models under misspecification: two-stage least squares (2SLS) and maximum likelihood (ML) estimators. *Sociol. Methods Res.* 36, 48–86. doi: 10.1177/0049124107301947

Boomsma, A. (1982). “The robustness of LISREL against small sample sizes in factor analysis models,” in, *Systems under Indirect Observation: Causality, Structure, Prediction, Part I*, eds K. G. Jöreskog and H. Wold (Amsterdam: North-Holland), 149–173.

Boomsma, A. (1985). Nonconvergence, improper solutions, and starting values in LISREL maximum likelihood estimation. *Psychometrika* 50, 229–242.

Campbell, D. T., and Fiske, D. W. (1959). Convergent and discriminant validation by the multitrait-multimethod matrix. *Psychol. Bull.* 56, 81–105.

Chen, F., Bollen, K. A., Paxton, P., Curran, P. J., and Kirby, J. B. (2001). Improper solutions in structural equation models: causes, consequences, and strategies. *Sociol. Methods Res.* 29, 468–508. doi: 10.1177/0049124101029004003

de Leeuw, J. (2004). “Least squares optimal scaling of partially observed linear systems,” in *Recent Developments on Structural Equation Models: Theory and Applications*, eds K. van Montfort, J. Oud, and A. Satorra (Dordrecht: Kluwer Academic Publishers), 121–134.

de Leeuw, J. (2008). *Factor Analysis as Matrix Decomposition*. Preprint Series: Department of Statistics, University of California, Los Angeles, CA.

Dijkstra, T. K. (2010). “Latent variables and indices: Herman Wold's basic design and partial least squares,” in *Handbook of Partial Least Squares: Concepts, Methods and Applications*, eds V. Esposito Vinzi, W. W. Chin, J. Henseler, and H. Wang (Berlin: Springer-Verlag), 23–46.

Dijkstra, T. K., and Henseler, J. (2015). Consistent and asymptotically normal PLS estimators for linear structural equations. *Comput. Stat. Data Anal*. 81, 10–23. doi: 10.1016/j.csda.2014.07.008

Duncan, T. D., Duncan, S. C., and Strycker, L. A. (2006). *An Introduction to Latent Variable Growth Curve Modeling: Concepts, Issues, and Applications*. Mahwah NJ: Lawrence Erlbaum Associates.

Esposito Vinzi, V., Trinchera, L., and Amato, S. (2010). “PLS path modeling: from foundations to recent developments and open issues for model assessment and improvement,” in *Handbook of Partial Least Squares. Concepts, Methods, and Applications*, eds V. Esposito Vinzi, W. W. Chin, J. Henseler, and H. Wang (Berlin: Springer-Verlag), 47–82.

Fornell, C., and Bookstein, F. L. (1982). Two structural equation models: LISREL and PLS applied to consumer exit-voice theory. *J. Mark. Res.* 19, 440–452.

Fornell, C., Johnson, M. D., Anderson, E. W., Cha, J., and Bryant, B. E. (1996). The American customer satisfaction index: nature, purpose, and findings. *J. Mark.* 60, 7–18.

Gleason, T. C. (1973). Improving the metric quality of questionnaire data. *Psychometrika* 38, 393–410.

Guttman, L. (1955). The determinacy of factor score matrices with implications for five other basic problems of common-factor theory. *Br. J. Stat. Psychol*. 8, 65–81.

Hwang, H., Malhotra, N. K., Kim, Y., Tomiuk, M. A., and Hong, S. (2010). A comparative study on parameter recovery of three approaches to structural equation modeling. *J. Mark. Res*. 47, 699–712. doi: 10.1509/jmkr.47.4.699

Hwang, H., and Takane, Y. (2004). Generalized structured component analysis. *Psychometrika* 69, 81–99. doi: 10.1007/BF02295841

Hwang, H., and Takane, Y. (2014). *Generalized Structured Component Analysis: A Component-Based Approach to Structural Equation Modeling*. Boca Raton, FL: Chapman & Hall/CRC Press.

Johnson, M. D., Gustafsson, A., Andreassen, T., Lervik, L., and Cha, J. (2001). The evolution and future of national customer satisfaction index models. *J. Econ. Psychol.* 22, 217–245. doi: 10.1016/S0167-4870(01)00030-7

Jöreskog, K. G. (1970). A general method for analysis of covariance structures. *Biometrika* 57, 409–426.

Jöreskog, K. G. (1973). “A generating method for estimating a linear structural equation system,” in *Structural Equation Models in the Social Sciences*, eds A. S. Goldberger and O. D. Duncan (New York, NY: Academic Press), 85–112.

Jöreskog, K. G., and Wold, H. (1982). “The ML and PLS techniques for modeling with latent variables: historical and comparative aspects,” in *Systems under Indirect Observation: Causality, Structure, Prediction, Part I*, eds H. Wold and K. G. Jöreskog (Amsterdam: North-Holland), 263–270.

Lei, P.-W., and Wu, Q. (2012). “Estimation in structural equation modeling,” in *Handbook of Structural Equation Modeling*, ed R. H. Hoyle (New York, NY: The Guilford Press), 164–180.

Lohmöller, J.-B. (1989). *Latent Variable Path Modeling with Partial Least Squares*. New York, NY: Springer-Verlag.

McArdle, J. J., and McDonald, R. P. (1984). Some algebraic properties of the reticular action model for moment structures. *Br. J. Math. Stat. Psychol.* 37, 234–251.

Rigdon, E. E. (2012). Rethinking partial least squares path modeling: in praise of simple methods. *Long Range Plan.* 45, 341–358. doi: 10.1016/j.lrp.2012.09.010

Rosseel, Y. (2012). lavaan: an R package for structural equation modeling. *J. Stat. Softw*. 48, 1–36. doi: 10.18637/jss.v048.i02

Sarstedt, M., Hair, J. F., Ringle, C. M., Thiele, K. O., and Gudergan, S. P. (2016). Estimation issues with PLS and CBSEM: where the bias lies! *J. Bus. Res.* 69, 3998–4010. doi: 10.1016/j.jbusres.2016.06.007

Schönemann, P. H., and Wang, M.-M. (1972). Some results on factor indeterminacy. *Psychometrika* 37, 61–91.

ten Berge, J. M. F. (1993). *Least Squares Optimization in Multivariate Analysis*. Leiden: DSWO Press, Leiden University.

Tenenhaus, M. (2008). Component-based structural equation modelling. *Total Qual. Manag. Bus. Excell*. 19, 871–886. doi: 10.1080/14783360802159543

Trendafilov, N. T., and Unkel, S. (2011). Exploratory factor analysis of data matrices with more variables than observations. *J. Comput. Graph. Stat.* 20, 874–891. doi: 10.1198/jcgs.2011.09211

Trendafilov, N. T., Unkel, S., and Krzanowski, W. (2013). Exploratory factor and principal component analyses: some new aspects. *Stat. Comput*. 23, 209–220. doi: 10.1007/s11222-011-9303-7

Unkel, S., and Trendafilov, N. T. (2013). Zig-zag exploratory factor analysis with more variables than observations. *Comput. Stat.* 28, 107–125. doi: 10.1007/s00180-011-0275-z

Velicer, W. F., and Jackson, D. N. (1990). Component analysis versus common factor analysis: some issues in selecting an appropriate procedure. *Multivariate Behav. Res.* 25, 1–28.

Wold, H. (1966). “Estimation of principal components and related methods by iterative least squares,” in *Multivariate Analysis*, ed P. R. Krishnaiah (New York, NY: Academic Press), 391–420.

Wold, H. (1973). “Nonlinear iterative partial least squares (NIPALS) modeling: some current developments,” in *Multivariate Analysis*, ed P. R. Krishnaiah (New York, NY: Academic Press), 383–487.

Wold, H. (1982). “Soft modeling: the basic design and some extensions,” in *Systems under Indirect Observations, Part II*, eds K. G. Jöreskog and H. Wold (Amsterdam: North-Holland), 1–54.

Keywords: generalized structured component analysis, uniqueness, measurement error, bias correction, structural equation modeling

Citation: Hwang H, Takane Y and Jung K (2017) Generalized Structured Component Analysis with Uniqueness Terms for Accommodating Measurement Error. *Front. Psychol*. 8:2137. doi: 10.3389/fpsyg.2017.02137

Received: 26 July 2017; Accepted: 23 November 2017;

Published: 06 December 2017.

Edited by:

Yanyan Sheng, Southern Illinois University Carbondale, United StatesReviewed by:

Edward E. Rigdon, Georgia State University, United StatesSunho Jung, Kyung Hee University, South Korea

Copyright © 2017 Hwang, Takane and Jung. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Heungsun Hwang, heungsun.hwang@mcgill.ca