Skip to main content

SYSTEMATIC REVIEW article

Front. Psychol., 22 October 2019
Sec. Quantitative Psychology and Measurement

A Scientometric Review of Rasch Measurement: The Rise and Progress of a Specialty

  • 1National Institute of Education, Nanyang Technological University, Singapore, Singapore
  • 2School of Social Sciences - Psychology, Nanyang Technological University, Singapore, Singapore

A recent review of the literature concluded that Rasch measurement is an influential approach in psychometric modeling. Despite the major contributions of Rasch measurement to the growth of scientific research across various fields, there is currently no research on the trends and evolution of Rasch measurement research. The present study used co-citation techniques and a multiple perspectives approach to investigate 5,365 publications on Rasch measurement between 01 January 1972 and 03 May 2019 and their 108,339 unique references downloaded from the Web of Science (WoS). Several methods of network development involving visualization and text-mining were used to analyze these data: author co-citation analysis (ACA), document co-citation analysis (DCA), journal author co-citation analysis (JCA), and keyword analysis. In addition, to investigate the inter-domain trends that link the Rasch measurement specialty to other specialties, we used a dual-map overlay to investigate specialty-to-specialty connections. Influential authors, publications, journals, and keywords were identified. Multiple research frontiers or sub-specialties were detected and the major ones were reviewed, including “visual function questionnaires”, “non-parametric item response theory”, “valid measures (validity)”, “latent class models”, and “many-facet Rasch model”. One of the outstanding patterns identified was the dominance and impact of publications written for general groups of practitioners and researchers. In personal communications, the authors of these publications stressed their mission as being “teachers” who aim to promote Rasch measurement as a conceptual model with real-world applications. Based on these findings, we propose that sociocultural and ethnographic factors have a huge capacity to influence fields of science and should be considered in future investigations of psychometrics and measurement. As the first scientometric review of the Rasch measurement specialty, this study will be of interest to researchers, graduate students, and professors seeking to identify research trends, topics, major publications, and influential scholars.

Introduction

A recent review of the literature concluded that Rasch measurement is an influential psychometric approach in psychology research (Edelsbrunner and Dablander, 2019). Rasch measurement refers to a family of unidimensional and multidimensional psychometric models inspired by the original formulation of a probabilistic model referred to as “the Rasch model” developed by a Danish mathematician called Georg Rasch in 1957 (Andersen, 1982; Andrich, 1997) to overcome the issues with psychometric testing at that time (Rasch, 1960; Wright and Stone, 1979). At its dawn, the Rasch model was reviewed positively by Coombs (1960), Loevinger (1965), and Sitgreaves (1960) and was embraced and promoted by scholars such as Benjamin Wright (Wright, 1996). In 1960, after learning about Rasch's work from Benjamin Wright, Jimmy Savage invited Rasch to Chicago for a series of lectures from March to June (Wright, 1996; Olsen, 2003). In an interview with David Andrich in May 1979 at Rasch's thatched cottage on the Danish island of Laesoe, Rasch stated:

I lectured on the contents of my book [Probabilistic Models.]. Jimmy Savage [Professor of Statistics] started to listen to some of it, but, of course, the mathematical details were so well-known to him. So in the long run, he got tired of it. I did not blame him. So did most of the audience. Only Ben[jamin] [Wright] stayed on. He was regular, took his notes, and I discussed data he brought with him. I also disclosed to him the generalization that Frisch [Nobel-Prize-winning Economist Ragnar] had inspired me to make and the points I was to present at Berkeley [Fourth Symposium on Mathematical Statistics and Probability]. (Andrich, 1997).

Soon after, Rasch (1963) presented the concept of objective measurement through Poisson processes at the International Congress of Psychology. Having a background in physics, Wright who launched courses on Rasch measurement at the University of Chicago in 1964, 1967 and 1968, demonstrated the application of the Rasch model to the Law School Admission Test supporting the argument that the Rasch model is indeed useful in psychometric testing. Another remarkable event was David Andrich's meeting with Georg Rasch in 1972. This marked the start of Rasch's scientific collaborations with Andrich, followed by Andrich's formulation of the rating scale model for polytomous data (Andrich, 1978) that was based on Andersen's (1977) results1.

In further two development, the logistic Rasch model that was originally developed to calibrate dichotomous data (x = 0 and 1) (Rasch, 1960; Engelhard, 2012; Engelhard and Wind, 2018), was extended to the partial credit model (Masters, 1982, 1988, 2010), many-facet Rasch measurement (MFRM2) (Linacre, 1989), linear logistic test model (Fischer, 1995), linear rating scale (Fischer and Parzer, 1991) and partial credit models (Fischer and Ponocny, 1994), linear and repeated measures models (Hoijtink, 1995), extended rating scale and expansions of the partial credit model for assessing change (Fischer and Parzer, 1991; Fischer and Ponocny, 1994), and the mixture distribution Rasch model (Rost, 1990; von Davier, 1996), to name a few. There were also a number of mixture distribution Rasch models, such as loglinear multivariate mixture Rasch models (Kelderman, 2007) and mixture hybrid models (von Davier and Yamamoto, 2007). Scholars modified the Rasch multidimensionality3 (Embretson, 1991; Ackerman, 1994), such as the multidimensional random coefficients multinomial logit (MRCML) model (Adams et al., 1997; Wang et al., 1997; Wu et al., 1998; for further developments of the model, please see e.g., Kelderman and Rijkes, 1994; Fischer and Molenaar, 1995; Briggs and Wilson, 2003).

Fischer (2010)—who had been a regular attendant at Rasch's office hours—wrote that after spending 2 months in Copenhagen with Rasch, he developed a computer software for conditional maximum likelihood estimation of the Rasch model in 1967, which was used extensively in German-speaking countries. In 1983, Benjamin Wright and Mike Linacre developed Microsale – the first Rasch software that was robust against missing data. Microscale was further developed into Bigscale in 1989, Bigsteps in 1991, and Winsteps in 1998 (Linacre, 1998, 2004). In addition to these computer programs, many other Rasch computer programs had been developed. For example, the first programs with wide usage were independently published around 1967 by Benjamin Wright and Gerhard Fischer, and the first Rasch software to be used for large-scale data analysis was BICAL4 (Wright et al., 1978). Some of the most commonly used Rasch measurement computer programs commercially available today include RUMM2030 (Andrich et al., 2009), ACER ConQuest (Wu et al., 2007), Winmira (von Davier, 2001), Facets and Winsteps (Linacre, 2019a,b; a list of Rasch software is available from: https://www.rasch.org/software.htm).

The method proposed by Georg Rasch for calibrating test items was preceded by Thorndike's (1919) and Thurstone's (1925) scaling methods which, like the Rasch models, were concerned with converting raw scores to scale-free measures of ability. According to Engelhard (1984, p. 26), the three approaches were similar in terms of “[t]he[ir] concept of an underlying latent trait [that] plays a central role in the quest for invariance with the three scaling methods.” While Thorndike's (1919) and Thurstone's (1925) scaling methods assumed a normal distribution for test data, Rasch's approach did not assume normality as it was set out to create scale free measures at the individual rather than group level of analysis (Engelhard, 1984). The normality assumption also underlies Birnbaum's (1967), Lord's (1952), and Lord and Novick's (1968) latent trait and item response theory (IRT) models, all of which strongly influenced educational and psychological measurement.

Finally, the enormous potential of Rasch models is recognized in various fields, including health and medical research (Tennant et al., 2004b; Pallant and Bailey, 2005; Tennant and Conaghan, 2007; e.g., Belvedere and de Morton, 2010; Sica da Rocha et al., 2013), assessment of educational and language skills (e.g., McNamara and Knoch, 2012; Eckes, 2015), rater training (Engelhard and Wind, 2018), and psychological measurement (e.g., Bond and Fox, 2015), to name a few.

The Present Study

There are reviews of the Rasch measurement specialty, such as Smith's (2019) and Wright's (1996) lists of key events, Olsen's (2003) extensive PhD thesis on the life and contributions of Georg Rasch, Bond's (2005) historical review of Rasch measurement, Panayides et al.'s (2010) historical account and defense of Rasch measurement in England, Belvedere and de Morton's (2010) review of studies validating mobility scales, Tesio et al.'s (2007) review of the application of Rasch measurement in rehabilitation research, Fischer's (2010) historical review of Rasch measurement in Europe, Wright's (1997) review of measurement in social sciences, and McNamara and Knoch's (2012) review of Rasch measurement in language assessment. In contrast to these positive accounts, Goldstein and Wood's (1989, p. 139) review of IRT (Rasch included) asserted that research in this field had shown a “disappointing lack of advance” in its 50 years of existence (see also Chien and Shao, 2019).

Several textbooks on Rasch measurement also reviewed the history and evolution of the speciality, highlighting the parts that were more relevant to the theme of the books (e.g., Andrich, 1989; Engelhard and Wind, 2018). In our view, to characterize the Rasch measurement specialty, it is important to determine the forces that have shaped its evolution over the years. These forces include impactful research trends, influential researchers, publications, and research outlets where the results of investigations have been published (Chen, 2016b). In addition, it is important to investigate the inter-domain trends through which the Rasch measurement specialty is linked to other fields of science.

To achieve this objective, we used a co-citation technique to identify the main players in the field. This method contrasts with previous investigations that were descriptive in nature (Tennant, 2011; Edelsbrunner and Dablander, 2019). In the present paper, we adopted advanced visualization (Chen et al., 2010) and text-mining methods. The text-mining methods include (i) author co-citation analysis (ACA) (Leydesdorff, 2005; Zhao and Strotmann, 2008), which considers two authors to be co-cited if they are cited together in a paper; (ii) document co-citation analysis (DCA) (e.g., Chen, 2004, 2006; Chen et al., 2008), where a co-citation instance occurs when two sources are cited together in one paper; (iii) journal co-citation analysis (JCA), which is used to identify journals cited together in one paper; and (iv) keyword analysis, where instances of two keywords appearing together are analyzed (Chen, 2017). These methods are similar to “big data” techniques suitable for computationally analyzing extremely large datasets to reveal active or developing research trends consisting of nodes with citation bursts defined as a publication's rapid increase in citation and influence (Chen, 2017). To investigate the inter-domain trends that link the Rasch measurement specialty to other specialties, we used a dual-map overlay (Chen and Leydesdorff, 2014) to investigate specialty-to-specialty connections (see Methodology for further details).

The co-citation technique, compared with methods such as narrative reviews (Chen et al., 2010), offers several main benefits such as: (i) the use of extensive bibliographic data adopted from Web of Science (WoS) and/or Scopus; (ii) reducing the inconvenience of analyzing huge datasets and providing computation of co-citations; (iii) leveraging computer programs for visualization and text-mining, such as CiteSpace (Chen et al., 2010; Chen, 2016a); and (iv) allowing researchers to produce a quantitative interpretation of the past and present state of specialties.

The research questions of the present study are as follows:

1. Where is Rasch measurement research situated on the map of the WoS and how is it linked with other research fields?

2. What are the impactful publications (bursts), major research trends, and keywords? What does the content of the major research trends reveal about the Rasch measurement specialty?

Methodology

Data Source

The data used for analyses comprised 5,365 publications on theory and practice in Rasch measurement between 01 January 19725 and 03 May 2019 with their 108,339 unique references downloaded from the WoS. These included 49,991 citing articles (that cited one or more publications in the corpus; articles without self-citations = 45,765). The reason for choosing WoS over other databases such as Scopus was its scope and coverage of published research. As our search of Scopus returned a smaller sample (<4,000 publications), we decided to adopt the database from WoS. However, a caveat is that the cited references in the database from WoS only included the first author6; Other contributors to the publications were thus not included in the analysis.

The search code used was “TOPIC: (“Rasch model”) OR TOPIC: (“Rasch measurement”) OR TOPIC: (“Rasch analysis”).” Figure 1 presents the descriptive statistics of the dataset downloaded from WoS. Rehabilitation (13.09%), education and educational research (11.97%), health care sciences services (9.84%), psychology / mathematics (8.80%), and health policy services (6.88%) constitute the top five disciplines with the highest application of the Rasch model. By contrast, linguistics (2.01%), orthopedics (2.05%), social sciences interdisciplinary (2.09%), multidisciplinary sciences (2.14%), and psychology (2.63%) form the bottom five adopters of the Rasch model. Each field can be further broken down into subfields. For example, in linguistics, which is the smallest field identified, U Koch (publications = 6), V Aryadoust (publications = 5), G Janssen (publications = 5), and J Tarace (publications = 5) achieved the highest number of publications among scholars in this field.

FIGURE 1
www.frontiersin.org

Figure 1. Frequency of publications using Rasch measurement in different fields (Source: WoS).

Research Question 1: The Dual-Map Overlay

To answer research question one, we used CiteSpace (Chen, 2016a) to generate a dual-map overlay that displayed a first base map of citing journals and a second base map of cited journals in the same user interface by using the influential journals of all fields retrieved from WoS to generate the overlay (Chen and Leydesdorff, 2014). Trajectories were then generated from the citing journals and cited journals to provide a better overview of the citations. Subsequently, the Blondel algorithm (Leydesdorff et al., 2013) was used to assign these journals to a cluster. This algorithm provided access to community networks of varying resolutions of community detection by finding high modularity partitions and unfolding a full hierarchical community layout (Blondel et al., 2008). The dual-map overlay allowed us to conduct several visual analyses as we were able to see the sources and targets of citations from various publications and the distributions of the citation arcs. This allowed us to investigate inter-specialty relationships and citation patterns from a group of publications.

Research Question 2: Bursts, Trends, and Their Influence

To answer the second research question, we adopted Chen et al.'s (2010, p. 1,389) “multiple-perspective co-citation analysis” technique that comprised the analysis of “structural, temporal, and semantic patterns as well as the use of both citing and cited items for interpreting the nature of co-citation clusters.” The components of this approach are discussed below.

Selection of Nodes

The two most recommended methods for node selection are Top N and Top N%. The Top N per slice procedure used in this study, selected the most cited items from each slice to form a network, according to the input value and node type determined by the user. We chose a value of 50 and multiple node types, so the top 50 most cited items were displayed and ranked accordingly. The Top N% per slice procedure displayed the percentage of most cited items according to a value determined by the user.

Network Development

The WoS dataset was used to construct co-citation networks for authors and publications (see below) for network development. Following Chen et al. (2010), ACA, DCA, JCA, and keyword analysis were performed to cluster co-citing authors (White and McCain, 1998; Chen, 1999; Leydesdorff, 2005; Zhao and Strotmann, 2008), co-citing publications, journals, and keywords (Small and Sweeney, 1985; Small and Greenlee, 1986; Chen, 2004, 2006; Chen et al., 2008), respectively.

Temporal and structural metrics were adopted to investigate network and cluster properties. Citation burstness and sigma (Σ) are temporal metrics. Knowing whether the citation count of a particular reference rose and when the rise occurred was important for citation analysis. Burst detection determined if the fluctuations for a specific frequency function, within a time period were significant. Sigma – the combination of betweenness centrality and burstness, was calculated as (centrality+1)burstness (Chen et al., 2010). This metric was used to identify and measure novel ideas presented in scientific publications (Chen et al., 2009). Sigma ranged from 0 to 1. Case studies by Chen et al. (2009) showed that the highest sigma values were usually associated with Nobel Prize and other award-winning researchers.

The average silhouette score (Rousseeuw, 1987), modularity Q (Newman, 2006), and betweenness centrality (Freeman, 1977) are structural metrics. The silhouette metric estimates the level of uncertainty when interpreting a cluster's nature. The silhouette value ranges from −1 to 1; a value of 1 suggests that a cluster is distinct from other clusters. The modularity Q ranges from 0 to 1 and measures the extent to which a network can be divided into modules. A high modularity score suggests a network with divisible structure, while a low modularity score suggest less distinct separation between clusters. The extent to which a node connects other nodes in a network is measured by the betweenness centrality metric. Scientific publications with high betweenness centrality values indicate potentially revolutionary material. These metrics are thus useful for finding influential scientific publications.

Visualization and Labeling of Clusters

We used the multidimensional clustering method for identification of clusters and their connections. We used two visualization methods to demonstrate the shape and form of the networks: the timeline view and the cluster view. The timeline view consisted of a range of vertical lines that represent time zones chronologically arranged from the left to right side (Chen, 2014). In this view, while the horizontal arrangement of nodes were restricted to the time zones they are located on, the nodes were allowed to have vertical links with nodes in other time zones (Chen, 2014). The cluster view, on the other hand, produced spatial network representations that were color-coded and automatically labeled in a landscape format.

We further used the log-likelihood ratio (LLR) method for automatic extraction of cluster labels. This method was found to provide the best results in terms of uniqueness and coverage (Dunning, 1993; Chen, 2014). Although the latent semantic indexing (LSI) and mutual information (MI) methods were available, they were not used in this study as their precision was lower compared with LLR (Chen, 2014).

Results

Dual-Map Overlay

Figure 2 shows the generated dual-map overlay: the citing journals are on the left, the cited journals are on the right, and the citation links tells us what journal the citing journal cited from. The trajectory of the citation links provides an understanding of inter-specialty relationships. A shift in trajectory from one region to another would indicate that a discipline was influenced by articles from another discipline. It was evident that medicine, sports, ophthalmology, neurology, and psychology were the dominant fields at the start of the trajectory. In contrast, health, nursing, sports, psychology, and economics dominated the end of the trajectory.

FIGURE 2
www.frontiersin.org

Figure 2. The dual-map overlay of the Rasch measurement specialty generated by CiteSpace (Chen, 2016a).

Author Co-citation Analysis (ACA)

The modularity Q score of the ACA network was 0.5411. As the modularity Q score represented how well the network was split into various independent clusters (Chen et al., 2010), a score of 0.5411 suggested that the networks and clusters were moderately well-structured. However, the boundaries that separated the clusters were not definitive. The average silhouette score was 0.3225, suggesting that the cluster had respectable heterogeneity. Table 1 presents the top 15, middle 5, and lowest 5 ranking publications as sample author bursts computed via ACA (for a complete list, please see Appendix 1). The author with the highest burst strength was WJ Boone (strength = 43.9175), whose burstness started in 2016 and continued to grow in 2019, followed by G Engelhard (strength = 42.3541, 2016–2019) and H Kelderman (strength = 32.5274, 1988–2005). Within the stipulated timeframe, the rapid changes in the number of citations received reflected the growing importance of the authors' papers and ideas.

TABLE 1
www.frontiersin.org

Table 1. Sample author bursts computed via author co-citation analysis (ACA).

As indicated on Table 1, SM Haley had the smallest burst (strength = 4.1974), whose burst lasted from 2010 to 2011. Notably, FM Lord had the longest lasting burst with a span of 27 years (strength = 31.2723). In contrast, SM Haley along with 12 other authors only had a burst span of one year (see Appendix 1 for a comprehensive list of chronologically ordered bursts).

Document Co-citation Analysis (DCA)

The timeline view and cluster view of DCA were generated to gain a clearer understanding of the occurrences and magnitudes of the bursts of different publications in the different clusters (Figures 3, 4). The clusters were numbered and ranked in terms of their size, with cluster #0 being the largest cluster. The size of the circle reflected the magnitude of the publication's influence: the larger circle, the higher the number of citations. The burstness of the author's publication was indicated by the red tree rings. From the DCA analysis, 57 clusters emerged. Figure 3 shows the top 14 largest clusters. Cluster #0 on visual questionnaires, was the largest cluster and showed activity from 2003 until the present. The presence of large nodes and nodes with the red tree rings in this cluster indicated that many publications from this cluster were highly influential or had high citation bursts. The size of the cluster was 174, and this accounted for 18.26% of all clusters. Cluster #1 on non-parametric item response theory, was the second largest cluster with a size of 114 (11.96%), followed by cluster #2, on visual function questionnaire, with a size of 103 (10.81%). Cluster #3 on valid measure (validity), cluster #4 on latent class model, and cluster #5 on many-facet Rasch model had sizes of 97 (10.17%), 80 (8.39%), and 75 (7.87%) respectively. The cluster view of the DCA network is presented in Figure 4. The cluster labels were turned off for clarity and only the authors' name of some influential publications were displayed.

FIGURE 3
www.frontiersin.org

Figure 3. Timeline view of the document co-citation analysis (DCA) network generated by CiteSpace (Chen, 2016a) (Modularity Q = 0.7319; average silhouette = 0.3199).

FIGURE 4
www.frontiersin.org

Figure 4. Cluster view of the document co-citation analysis (DCA) network generated by CiteSpace (Chen, 2016a) (modularity Q = 0.7319; average silhouette = 0.3199).

The top three largest clusters show activity for a span of roughly 20 years each. By contrast, the smaller clusters had shorter spans of 4–5 years each. The clusters had distinctive division of modules and respectable heterogeneity, with modularity Q score and average silhouette score at 0.7319 and 0.3199 respectively.

Table 2 presents the sample document bursts that were computed via DCA (see Appendix 2 for a comprehensive list of chronologically ordered bursts). Publications with high levels of strength signify major milestones in the development of Rasch measurement. Two notable publications by TG Bond (strength = 133.0131 and 88.1181, respectively) had the largest magnitudes of document bursts, with spans lasting 5 and 6 years, respectively. This suggested that works by TG Bond were not only highly influential but had greatly contributed to the development of Rasch measurement. A Tennant's publication (strength = 83.7317) also had high strength document burst with a span of 4 years. At the other end, the citation burst of the lowest magnitude was EB Andersen (strength = 4.1214) with a burst span of 2 years. Several authors, such as A Tennant, JM Linacre, K Pesudovs, and WJ van der Linden, shared the longest spanning document bursts at 7 years.

TABLE 2
www.frontiersin.org

Table 2. Sample document bursts computed via document co-citation analysis (DCA).

Table 3 shows the publications with high centrality and high sigma. As mentioned above, publications with high centrality scores were highly influential and items with high sigma scores indicate scientific novelty. A Tennant's publication (2007) was both highly influential and contained potentially scientific novel revelations (centrality = 0.15; sigma = 156925.6), followed by van der Linden and Hambleton (1997) (centrality = 0.15; sigma = 13.5) and two editions of Bond and Fox's (2001, 2007) monograph on Rasch measurement (centrality = 0.03 and sigma = 17.88; centrality = 0.09 and sigma = 88484.75, respectively) (see Appendix 1 for a comprehensive list of chronologically ordered bursts). High betweenness centrality indicated that the publications in Table 3 connected two or more clusters and, therefore, two or more themes (clusters) (Chen et al., 2009, 2010). As they connected different themes, they were also likely to be a synthesis of different ideas into a new one, and could be revolutionary in providing this connection. As previously stated, higher sigma values indicated the novelty of these publications (Chen et al., 2009; Chen, 2017).

TABLE 3
www.frontiersin.org

Table 3. Centrality and sigma estimated via document co-citation analysis (DCA).

Journal Co-citation Analysis (JCA)

Table 4 demonstrates sample journal bursts computed via JCA. A modularity Q score of 0.4626 suggested that the JCA network and clusters were moderately well-structured. However, the boundaries that delineated the clusters were not clear. The average silhouette score was 0.3522, suggesting that the cluster had respectable heterogeneity. The journal with the largest burst was PLoS ONE (strength = 83.752, 2015–2019), and the burstness was still growing. Arthritis and Rheumatism-Arthritis Care and Research (strength = 80.62, 2009–2015) had the second largest burst, followed by the Journal of Outcome Measurement (strength = 74.264, 1998–2007). The International Journal of Biometrics (strength = 9.6156), which had the longest burst span of 24 years from 1982 to 2006. Eight journals, including Applied Measurement in Education (strength = 7.073, 2009–2010), had a burst span of one year (see Appendix 3 for a comprehensive list of chronologically ordered bursts). Notably, the Journal of Outcome Measurement was a predecessor of Journal of Applied Measurement (JAM). Thus, its impact was combined with that of JAM (strength = 62.4672).

TABLE 4
www.frontiersin.org

Table 4. Sample publication bursts computed via journal co-citation analysis (JCA).

Keywords Co-citation Analysis

Keywords with high strength indicated influential ideas that originate from the clusters (Modularity Q = 0.406; average silhouette = 0.5166). Table 5 shows sample keyword bursts computed via keyword analysis. The keyword with the burst of the largest magnitude was “surveys and questionnaire(s)” (strength = 94.7513), with a burst span of 4 years, and it is still growing. Next were “psychology” (strength = 58.3953, 2014–2019) and “procedure” (strength = 48.8441, 2014–2019). The keyword “personality inventory” was not only the smallest burst, it also had a burst span of one year (strength = 9.9687). The longest burst spans of 20 years were “comparative study” (strength = 6.694, 1986–2006) and “clinical article” (strength = 8.201, 1986–2006).

TABLE 5
www.frontiersin.org

Table 5. Sample keyword bursts computed via co-occurring author keywords analysis and keyword plus.

Discussion

The present study adopted several co-citation techniques to investigate published research that applied or discussed Rasch measurement theory. The findings and research questions are discussed in this section.

First Research Question

To address the first research question on where Rasch measurement was located on the map of WoS and how it was linked to other research fields, a dual-map overlay was generated. As shown in the Results section, certain fields were especially prominent, such as medicine, neurology, and psychology, indicating that the Rasch models were commonly used in such fields. The left and the right base maps also had some common fields such as molecular biology and immunology on the left, and molecular biology and genetics on the right. Some of these fields were connected to each other as well. This was indicative that reciprocal citations within the field using Rasch models were common. To develop a full picture of the application of Rasch measurement in each field, additional intra-disciplinary studies will be needed to investigate how the model has contributed to research in each field.

Second Research Question

Influential Authors, Documents, and Journals

To address the second research question, a large database consisting of 5,365 publications on theory and practice in Rasch measurement between 01 January 1972 and 03 May 2019 and their 108,339 unique references were analyzed. We used a multiple perspectives approach consisting of network visualization methods, ACA, DCA, JCA, and keyword analysis to generate different networks and investigate the different dimensions of the Rasch measurement specialty. The top authors with the highest bursts included Boone WJ, Engelhard G, Kelderman H, Bock RD, Lamoreux EL, Lord FM, Hu LT, Christensen KB, Hagquist C, and Messick S (see the relevant tables and Appendix). Examining the areas of expertise of these authors revealed backgrounds not only in the Rasch models (e.g., Boone WJ and Engelhard G) but also in IRT models (e.g., Bock RD and Lord FM), and structural equation modeling (e.g., Hu LT), indicating the links between these fields and perhaps the influence that the Rasch measurement field has received from other fields (see section Limitations for a further discussion).

The most influential publication in Rasch measurement was the volume “Applying the Rasch Model: Fundamental Measurement in the Human Sciences” by Bond and Fox (2007), followed by its earlier edition (Bond and Fox, 2001). The third edition of the book ranks fourth after Tennant and Conaghan's (2007) article in Arthritis and Rheumatism-Arthritis Care and Research. From this perspective, Bond and Fox's book was exceptional as its three editions were among the top four influential publications in the Rasch measurement field.

Impactful journals include PLoS ONE and Arthritis & Rheumatism-Arthritis Care and Research, followed by Journal of Outcome Measurement (JOM), Psychometrika, and Journal of Applied Measurement (JAM). According to JAM Press (2018), “JOM was the predecessor of JAM and contains many articles related to Rasch measurement in education and the health sciences.” Lastly, impactful keywords included surveys, questionnaires, and psychology. Although there was no document from a journal or book by an author that had dominated the top of all lists, there were generally links where the most impactful authors had publications among the top 20 most impactful documents. This applied to comparisons of journals and authors as well as documents and journals, where the most influential authors published in the most influential journals. This indicated that there was a concentration of influence. Perhaps the Rasch measurement specialty was not extensively used beyond the fields in which those authors and journals were published. This hypothesis may be examined in future research.

Another general observation was the increase in number of fields that adopted the model over the years. Despite several decades after the publication of Rasch's (1960) book, the model was unknown to many fields. According to Fischer (2019, personal communication), “in 1970 there existed worldwide only four centers where research on the RM [Rasch model] was made on a continuous basis: Copenhagen (Georg Rasch and Erling Andersen), Chicago (Benjamin Wright and associates), Australia (David Andrich and associates), [and] Vienna (myself and associates).” These numbers had grown and the Rasch models are presently being used across different fields by different scholars in different parts of the world. In the following section, we provide a brief overview of several influential research clusters identified via DCA.

Major Research Clusters Identified by DCA

Different research patterns and themes were evident from the clusters identified. For example, Cluster#0 focused on measurement invariance in social, behavioral, and health sciences (Engelhard, 2013). Specifically, Rasch models assisted with overcoming several measurement problems, such as the limitations of rating scales (Hobart et al., 2007) and the construction of measures (Wilson, 2005; Bond and Fox, 2007, 2015; Boone et al., 2013). Rasch models were compared with IRT when applied in rating scales analysis (Hobart and Cano, 2009). In the same cluster, some bursts provided guidelines for researchers on how to utilize the advantages of the model (Tennant and Conaghan, 2007) and demonstrated its application in validating the hospital anxiety and depression scale (HADS) total score (HADS-14; Pallant and Tennant, 2007) and the nursing self-efficacy (NSE) scale (Hagquist et al., 2009).

Cluster#1 emphasized various streams of research on Rasch counterparts, including IRT models (Hambleton et al., 1991; van der Linden and Hambleton, 1997; Baker and Kim, 2004; de Boeck and Wilson, 2004), item and person fit computation in IRT (Glas and Verhelst, 1995; Meijer and Sijtsma, 2001), polytomous IRT models (Embretson and Reise, 2000), and the development of IRT models such as the generalized linear logistic test model (GLLTM) (Patz and Junker, 1999), an extension of the IRT model to overcome problems such as missing data and rated responses, and the multidimensional random coefficients multinomial logit model (Adams et al., 1997).

Cluster#2 had two primary themes: the unidimensionality requirement in Rasch measurement (e.g., Smith, 2002; Tennant et al., 2004a) and the application of Rasch measurement for validation of instruments that use Likert-type scales in medical fields (Massof and Fletcher, 2001; Massof, 2002; Pesudovs, 2006), such as patient-reported outcome measures (Pesudovs et al., 2007), the Quality of Life Impact of Refractive Correction (QIRC) questionnaire (Pesudovs et al., 2004), the Activities of Daily Vision Scale (ADVS; Pesudovs et al., 2003), and the Impact of Vision Impairment scale (Lamoureux et al., 2006). Finally, clusters#4 and #5 centered on the development of two forms of Rasch measurement: latent class Rasch measurement or the mixture Rasch model (e.g., Rost, 1991; see chapters in Fischer and Molenaar, 1995; von Davier, 1996) and many-facet Rasch measurement (e.g., Linacre, 1989; Eckes, 2015).

Based on our examination of the bursts' contents and personal communications with the identified scholars, we propose three groups of hypothetical factors that we view as facilitators of publication and author burstness, rendering three hypotheses. The first hypothesis is related to the publications' educational content and/or the perspectives that the authors on Rasch models have taken in their publications. Influential publications by Bond and Fox (2007, 2015), Engelhard (2013), Tennant and Conaghan (2007), and Boone et al. (2013), for example, are suitable for educators and medical practitioners who need to adopt rigorous measurement methods in their research, a contributing factor of their high citations and bursts. Engelhard (2019, personal communication) stated “I believe that my research is being cited because I write as a teacher […] Measurement is viewed as complex and statistical, while I view measurement as essentially a facet of clear thinking about the constructs in our theories […] I have tried to […] introduce the use of meaningful and invariant scales in numerous fields.” This resonated with Bond's (2019, personal communication) idea about the success of Bond and Fox's (2001, 2007, 2015) book, stressing that making an attempt to communicate the properties of the Rasch model to the ever-growing field of psychology, medicine, and social sciences is a key factor in attracting more scholars to this field. Similarly, Boone (2019, personal communication) highlighted his endeavors to find efficient ways “how to explain Rasch and how to encourage the use of Rasch among non-psychometricians.”

The second hypothesis about bursts concerns the timeliness of publications and the outreach and reputation of authors and publishers. According to Tennant (2019, personal communication), “the impact of [his] publication was a mix of publishing at a time when the use of the Rasch model was expanding rapidly […], the fact that it was written as a teaching paper suitable for clinicians, and the dissemination through Arthritis & Rheumatism-Arthritis Care & Research to the potentially large musculoskeletal community. It is also referenced in our Psychometric Laboratory teaching programme, which is undertaken in many European countries, to a wide range of professionals in health care.”

The third hypothesis exclusively relates to the available Rasch software. Two of the highest bursts were Linacre's Winsteps and Facets. There are several possible reasons for the status of these Rasch model packages, such as longevity (Winsteps and Facets started in 1998 and 1987 respectively), comprehensiveness of the outputs and computations, capacity for computation, robustness to missing data, and being instructive (having detailed manuals), and offering prompt support7 (Linacre, 2019, personal communication). We call for further research to shed light on the socio-cultural and ethnographic aspects of Rasch measurement research.

Limitations

The present study is not without its limitations. First, the co-citation method was only able to identify impactful publications and authors only after a sufficient amount of time from their emergence. Therefore, it was incapable of predicting the future of recently published works or publishing authors and, in our view, should not be used to evaluate recent publications. Second, the results were data-driven and accurate only when comprehensive databases were used. It might seem puzzling that Georg Rasch was not among the top 10 influential authors. One reason could be that Rasch did not publish most of his ideas in peer-reviewed journals or books. Accordingly, there was seemingly insufficient acknowledgment of Georg Rasch's role in pioneering the field.

Relatedly, only data from WoS were used in this study; data from other databases such as PubMed and PsyInfo were not used. According to Falagas et al. (2008), PubMed also provides up-to-date articles with early online articles available for access, specifically for authors who are investigating the uses of this database for medical purposes. In our view, WoS is better for investigating Rasch models given its wider coverage and scope compared with other available databases. Future research could compare PubMed and PsyInfo with WoS for investigating Rasch modeling or other research topics and decide on the more comprehensive database to use.

Although comprehensive databases such as the WoS are beneficial for generating the types of data shown in this paper, they may be casting a wide net. For example, one of the most influential authors identified was Hu LT and one of the most influential publications was on structural equation modeling (SEM) (Hu and Bentler, 1999; cited 31,582 times on WoS). The paper mainly compared various fit indexes for SEM analysis. A search for “Rasch” in the documents that cited this article produced 199 hits. These citing articles mainly used Rasch models alongside SEM and cited Hu and Bentler (1999) for use of the fit indices and the criterion provided in SEM analysis (e.g., Rowe et al., 2017; Finbråten et al., 2018; Mairesse et al., 2019). Although the work Hu and Bentler (1999) did not just focus on Rasch models, it had exerted a cross-disciplinary influence on Rasch-related research through its application in Rasch-SEM studies. This finding has two mutually exclusive implications for future research. First, based on this finding, we anticipate that measureable interdisciplinary connections may exist between Rasch measurement and IRT on the one hand and SEM research on the other hand. For example, increasingly more researchers adopt Rasch measurement to validate their data before submitting them to SEM analysis (see Bond and Fox, 2015, p. 240–241). Instead of being quantitatively distinct from each other, these research areas may be interwoven networks with high homogeneity. Alternatively, future researchers who aim for high precision can consider using more stringent keyword searches to reduce the likelihood of including cross-disciplinary publications in the dataset (e.g., “Rasch measurement” AND NOT “structural equation modeling”), although one must be cautious to achieve a good balance between stringent criteria and over-excluding. Having to decide whether to retain the publications that were used in SEM and Rasch measurement or remove them from the analysis is a challenge.

Additionally, the labeling of clusters involved identifying specialties and interpreting the nature of these specialties. Although rewarding and insightful, manual labeling, is tedious and time-consuming. Thus, automatic labeling using algorithms were used in this study, not only to increase the overall efficiency of labeling clusters but also reduce biases (Chen, 2017). However, the labels were limited to the vocabulary used in the data source. This is a key limitation of automatic labeling as the most suitable labels may not exist in the data pool. Using numerous data sources to increase the vocabulary pool may be beneficial in overcoming this limitation. It is important that researchers are in control of the labeling process even when automatic labeling is used (see Aryadoust and Ang, 2019; Aryadoust et al., 2020).

Lastly, only the names of the principal (first) authors were used in the co-citation analyses performed in this study. Databases of cited publications downloaded from WoS did not include the names of other contributing authors even though citing publications did not possess such restriction. If additional author names were made available by these databases, the co-citation analysis may yield different results.

Conclusion

In this study, we identified research clusters, authors, journals, and keywords that had significant impacts on Rasch measurement research. Using a dual-map overlay, we also revealed multiple inter-domain connections between journals and scientific fields. Informed by personal communications with some of the highly cited authors, we proposed three hypotheses that considered the ethnographic and sociocultural factors to provide a preliminary explanation for the findings. Further research is needed to investigate the relevance of these hypotheses. As the Rasch model required unidimensionality, data-model fit, and local independence, future research may investigate how influential publications on Rasch model are in explaining these concepts and how citing authors conceptualized and applied these concepts. Lastly, this paper took a holistic approach to co-citation analysis of the Rasch measurement field and did not separate different sub-fields. It may be useful to conduct Scientometric studies in specialized fields where Rasch measurement was used for item and person calibration and assessment validation. This may provide in-depth information regarding the status of Rasch measurement in various fields. We hope that the findings of the present study will lead to better understanding of the Rasch measurement frontier.

Author Contributions

VA conceived the study, created the dataset, conducted the co-citation analysis, and contributed to writing the paper. HT contributed to the literature review and writing the paper. LN contributed to the literature review and revising the paper.

Funding

This study was supported in part by the National Institute of Education of Nanyang Technological University (Grant number: RI 2/16 VSA) and by Pearson Education, UK (Grant: Using the Rasch Model to Investigate Differential Item Functioning in the PET Reading Test).

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

We would like to thank Gerhard Fischer and Mike Linacre for their comments on earlier drafts of this paper. We wish to thank Trevor Bond, George Engelhard, Alan Tennant, and William Boone for helping us to interpret the findings. The views and opinions expressed in this work are those of the authors and do not necessarily reflect the opinions of the fund providers or the esteemed colleagues mentioned above.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyg.2019.02197/full#supplementary-material

Footnotes

1. ^Andersen (1995) stated that, earlier, Rasch (1961) had suggested a model for polytomous data in Berkeley.

2. ^According to Fischer (1995, p. 132), multifactorial Rasch models were initially suggested by Rasch (1965) and later by Micko (1969, 1970) and “have been developed explicitly by Scheiblechner (1971) and Kempf (1972),” as well.

3. ^It appeared that Rasch was not against the concept of multidimensionality for person and item parameters. In his presentation on a polytomous model, Rasch (1961) wrote that his model could be extended to multidimensional item/person parameters (Andersen, 1995, p. 272).

4. ^See Fischer (1974) for an extended introduction to test theory with an appendix of several improved computer programs for Rasch measurement.

5. ^Although Rasch first published his model in 1961, our literature search started in 1972. Unfortunately, neither WoS nor Scopus has indexed any publications by Rasch between 1961 and 1972. Nevertheless, since WoS provided a more comprehensive database of Rasch measurement publications after 1972, the decision was made to use WoS in the present study.

6. ^It should be noted that citing articles are not subject to this limitation.

7. ^An example is a recent change to Winsteps that included increasing the limit on the rating scale from 255 to 32,000 for each item. This was at the request of a user whose data were percentages with two decimal places. This is equivalent to 10001 rating scale categories (Linacre, 2019, personal communication). In addition, Dr Linacre has made a time-sensitive full version of Winsteps available to educators and scholars who plan to teach Rasch model workshops.

References

Ackerman, T. (1994). Using multidimensional item response theory to understand what items and tests are measuring. Appl. Measurement Edu. 7, 255–278. doi: 10.1207/s15324818ame0704_1

CrossRef Full Text | Google Scholar

Adams, R. J., Wilson, M., and Wang, W. C. (1997). The multidimensional random coefficients multinomial logit model. Appl. Psychol. Measurement 21, 1–23. doi: 10.1177/0146621697211001

CrossRef Full Text | Google Scholar

Andersen, E. B. (1977). Sufficient statistics and latent trait models. Psychometrika 42, 69–81. doi: 10.1007/BF02293746

CrossRef Full Text | Google Scholar

Andersen, E. B. (1982). Georg Rasch (1901–1980). Psychometrika 47, 375–376. doi: 10.1007/BF02293703

CrossRef Full Text | Google Scholar

Andersen, E. B. (1995). “Polytomous Rasch models and their estimation,” in Rasch Models: Foundations, Recent Developments, and Applications, eds G. H. Fischer and I. W. Molenaar (New York, NY: Springer), 271–292. doi: 10.1007/978-1-4612-4230-7_15

CrossRef Full Text | Google Scholar

Andrich, D. (1978). A rating formulation for ordered response categories. Psychometrika 43, 561–573. doi: 10.1007/BF02293814

CrossRef Full Text | Google Scholar

Andrich, D. (1988). Quantitative Applications in the Social Sciences. Rasch Models for Measurement. Thousand Oaks, CA: Sage Publications, Inc.

Google Scholar

Andrich, D. (1989). Distinctions between assumptions and requirements in measurement in the social sciences. Math. Theor. Syst. 4, 7–16.

Google Scholar

Andrich, D. (1997). Georg Rasch in his own words. Rasch Measurement Transac. 11, 542–543.

Google Scholar

Andrich, D. (1998). Rasch Models for Measurement. Series: Quantitative Applications in the Social Sciences. London: Sage Publications.

Google Scholar

Andrich, D., Sheridan, B., and Luo, G. (2009). RUMM2030: Rasch Unidimensional Models for Measurement (Computer Program). Perth: RUMM Laboratory.

Google Scholar

Aryadoust, V., and Ang, B. H. (2019). Exploring the frontiers of eye tracking research in language studies: a novel scientometric review. Computer Assisted Language Learning 2019, 1–36. doi: 10.1080/09588221.2019.1647251

CrossRef Full Text | Google Scholar

Aryadoust, V., Ilang Kumaran, I., and Ferdinand, S. (2020). “A co-citation review of listening comprehension: a scientometric investigation of 70 years of research,” in The Handbook of Listening, eds D. L. Worthington and G. D. Bodie (West Sussex: Wiley-Blackwell), 1310.

Google Scholar

Baker, F. B., and Kim, S. H. (2004). Item Response Theory: Parameter Estimation Techniques. Oxford: CRC Press.

Google Scholar

Belvedere, S. L., and de Morton, N. A. (2010). Application of Rasch analysis in health care is increasing and is applied for variable reasons in mobility instruments. J. Clin. Epidemiol. 62, 1287–1297. doi: 10.1016/j.jclinepi.2010.02.012

CrossRef Full Text | Google Scholar

Birnbaum, Z. W. (1967). Statistical Theory for Logistic Mental Test Models with a Prior Distribution of Ability (ETS Research Bulletin RB-67-12). Princeton, NJ: Educational Testing Service.

Google Scholar

Blondel, V.D., Guillaume, J.L., Lambiotte, R., and Lefebvre, E. (2008). Fast unfolding of communities in large networks. J. Stat. Mechanics Theory Exp. 8:10008. doi: 10.1088/1742-5468/2008/10/P10008

CrossRef Full Text | Google Scholar

Bond, T. G. (2005). “Past, present and future: an idiosyncratic view of Rasch measurement,” in Applied Rasch measurement: A Book of Exemplars, eds S. Alagumalai, D. D. Curtis, and N. Hungi, eds J. P. Keeves (Dordrecht: Springer), 329–341. doi: 10.1007/1-4020-3076-2_18

CrossRef Full Text | Google Scholar

Bond, T. G., and Fox, C. M. (2001). Applying the Rasch Model: Fundamental Measurement in the Human Sciences. Oxford: Psychology Press.

Google Scholar

Bond, T. G., and Fox, C. M. (2007). Applying the Rasch Model: Fundamental Measurement in the Human Sciences, 2nd Edn. Mahwah, NJ: Lawrence Erlbaum Associates Publishers.

Google Scholar

Bond, T. G., and Fox, C. M. (2015). Applying the Rasch Model: Fundamental Measurement in the Human Sciences, 3rd Edn. Oxford: Psychology Press.

Google Scholar

Boone, W. J., Staver, J. R., and Yale, M. S. (2013). Rasch Analysis in the Human Sciences. Oxford: Springer Science and Business Media.

Google Scholar

Boone, W. J., Townsend, J. S., and Staver, J. (2011). Using Rasch theory to guide the practice of survey development and survey data analysis in science education and to inform science reform efforts: an exemplar utilizing STEBI self-efficacy data. Sci. Edu. 95, 258–280. doi: 10.1002/sce.20413

CrossRef Full Text | Google Scholar

Briggs, D. C., and Wilson, M. (2003). An introduction to multidimensional measurement using Rasch models. J. Appl. Measurement 4, 87–100.

PubMed Abstract | Google Scholar

Chen, C. (1999). Visualising semantic spaces and author co-citation networks in digital libraries. Info. Processing Manag. 35, 401–420. doi: 10.1016/S0306-4573(98)00068-5

CrossRef Full Text | Google Scholar

Chen, C. (2004). Searching for intellectual turning points: progressive knowledge domain visualization. Proc. Natl Acad. Sci. U.S.A. 101 (Suppl. 1), 5303–5310. doi: 10.1073/pnas.0307513100

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, C. (2006). Information Visualization: Beyond the Horizon. Oxford: Springer Science & Business Media.

Google Scholar

Chen, C. (2014) The CiteSpace Manual. Available online at: http://blog.sciencenet.cn/home.php?mod=attachment&filename=CiteSpaceManual.pdf&id=52563 (accessed July 25, 2019).

Google Scholar

Chen, C. (2016a). CiteSpace: A Practical Guide for Mapping Scientific Literature. Oxford: Nova Science Publishers, Incorporated.

Google Scholar

Chen, C. (2016b). Grand challenges in measuring and characterizing scholarly impact. Front. Res. Metrics Analytics 1:4. doi: 10.3389/frma.2016.00004

CrossRef Full Text | Google Scholar

Chen, C. (2017). Science mapping: a systematic review of the literature. J. Data Info. Sci. 2, 1–40. doi: 10.1515/jdis-2017-0006

CrossRef Full Text | Google Scholar

Chen, C., Chen, Y, Horowitz, M., Hou, H., Liu, Z., and Pellegrino, D. (2009). Towards an explanatory and computational theory of scientific discovery. J. Informetrics 3, 191–209. doi: 10.1016/j.joi.2009.03.004

CrossRef Full Text | Google Scholar

Chen, C., Ibekwe-SanJuan, F., and Hou, J. (2010). The structure and dynamics of cocitation clusters: a multiple-perspective cocitation analysis. J. Am. Soc. Info. Sci. Technol. 61, 1386–1409. doi: 10.1002/asi.21309

CrossRef Full Text | Google Scholar

Chen, C., and Leydesdorff, L. (2014). Patterns of connections and movements in dual-map overlays: a new method of publication portfolio analysis. J. Assoc. Info. Sci. Technol. 65, 334–351. doi: 10.1002/asi.22968

CrossRef Full Text | Google Scholar

Chen, C., Song, I. Y., Yuan, X., and Zhang, J. (2008). The thematic and citation landscape of data and knowledge engineering (1985–2007). Data Knowledge Eng. 67, 234–259. doi: 10.1016/j.datak.2008.05.004

CrossRef Full Text | Google Scholar

Chien, T.-W., and Shao, Y. (2019). The most-cited Rasch scholars on PubMed in 2018. Rasch Measurement Transac. 32, 1708–1712.

Google Scholar

Coombs, C. H. (1960). A theory of data. Psychol. Rev. 67, 143–159. doi: 10.1037/h0047773

PubMed Abstract | CrossRef Full Text | Google Scholar

de Boeck, P., and Wilson, M. (2004). Explanatory Item Response Models. New York, NY: Springer Science & Business Media.

Google Scholar

Dunning, T. (1993). Accurate methods for the statistics of surprise and coincidence. Comput. Linguistics 19, 61–74.

Google Scholar

Eckes, T. (2015). Introduction to Many-facet Rasch Measurement: Analyzing and Evaluating Rater-Mediated Assessments, 2nd Edn. Frankfurt am Main New York, NY: Peter Lang.

Google Scholar

Edelsbrunner, P. A., and Dablander, F. (2019). The psychometric modeling of scientific reasoning: a Review and Recommendations for Future Avenues. Edu. Psychol. Rev. 31, 1–34. doi: 10.1007/s10648-018-9455-5

CrossRef Full Text | Google Scholar

Embretson, S. E. (1991). A multidimensional latent trait model for measuring learning and change. Psychometrika 56, 495–515. doi: 10.1007/BF02294487

CrossRef Full Text | Google Scholar

Embretson, S. E., and Reise, S. P. (2000). Multivariate Applications Books Series. Item Response Theory for Psychologists. Mahwah, NJ: Lawrence Erlbaum Associates Publishers.

Google Scholar

Engelhard, G., and Wind, S. (2018). Invariant Measurement with Raters and Rating Scales: Rasch Models for Rater-mediated Assessments. New York, NY: Routledge.

Google Scholar

Engelhard, G. Jr. (1984). Thorndike, Thurstone, and Rasch: a comparison of their methods of scaling psychological and educational tests. Appl. Psychol. Measurement 8, 21–38. doi: 10.1177/014662168400800104

CrossRef Full Text | Google Scholar

Engelhard, G. Jr. (2012). Rasch measurement theory and factor analysis. Rasch Measurement Transac. 26, 1375.

Google Scholar

Engelhard, G. Jr. (2013). Invariant Measurement: Using Rasch Models in the Social, Behavioral, and Health Sciences. New York, NY: Routledge.

Google Scholar

Falagas, M. E., Pitsouni, E. I., Malietzis, G. A., and Pappas, G. (2008). Comparison of PubMed, Scopus, Web of Science, and Google Scholar: strengths and weaknesses. FASEB J. 22, 338–342. doi: 10.1096/fj.07-9492LSF

PubMed Abstract | CrossRef Full Text | Google Scholar

Finbråten, H. S., Wilde-Larsson, B., Nordström, G., Pettersen, K. S., Trollvik, A., and Guttersrud, Ø. (2018). Establishing the HLS-Q12 short version of the European Health Literacy Survey Questionnaire: latent trait analyses applying Rasch modelling and confirmatory factor analysis. BMC Health Serv. Res. 18:506. doi: 10.1186/s12913-018-3275-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Fischer, G.H. (1974). Einführung in Die Theorie Psychologischer Tests. Berne: Hans Huber.

Google Scholar

Fischer, G.H., and Molenaar, I. W. (1995). Rasch Models: Foundations, Recent Developments, and Applications. New York, NY: Springer-Verlag.

Google Scholar

Fischer, G.H., and Parzer, P. (1991). An extension of the rating scale model with an application to the measurement of treatment effects. Psychometrika 56, 637–651. doi: 10.1007/BF02294496

CrossRef Full Text | Google Scholar

Fischer, G. H. (1995). “The linear logistic test model,” in Rasch Models: Foundations, Recent Developments, and Applications, eds G. H. Fischer and I. W. Molenaar (New York, NY: Springer), 131–155. doi: 10.1007/978-1-4612-4230-7_8

CrossRef Full Text | Google Scholar

Fischer, G. H. (2010). The Rasch Model in Europe: a history. Rasch Measurement Transac. 24, 1294–1295.

Google Scholar

Fischer, G. H., and Ponocny, I. (1994). An extension of the partial credit model with an application to the measurement of change. Psychometrika 59, 177–192. doi: 10.1007/BF02295182

CrossRef Full Text | Google Scholar

Freeman, L. C. (1977). A set of measuring centrality based on betweenness. Sociometry 40, 35–41. doi: 10.2307/3033543

CrossRef Full Text | Google Scholar

Glas, C. A. W., and Verhelst, N. D. (1995). “Testing the Rasch model,” in Rasch Models: Foundations, Recent Developments, and Applications, eds G. H. Fischer and I. W. Molenaar (New York, NY: Springer-Verlag), 69–75.

Google Scholar

Goldstein, H., and Wood, R. (1989). Five decades of item response modelling. Br. J. Math. Stat. Psychol. 42, 139–167. doi: 10.1111/j.2044-8317.1989.tb00905.x

CrossRef Full Text | Google Scholar

Hagquist, C., Bruce, M., and Gustavsson, J. P. (2009). Using the Rasch model in nursing research: an introduction and illustrative example. Int. J. Nurs. Stud. 46, 380–393. doi: 10.1016/j.ijnurstu.2008.10.007

PubMed Abstract | CrossRef Full Text | Google Scholar

Hambleton, R. K., Swaminathan, H., and Rogers, H. J. (1991). Measurement Methods for the Social Sciences Series, Vol. 2. Fundamentals of Item Response Theory. Thousand Oaks, CA: Sage Publications, Inc.

Google Scholar

Hobart, J., and Cano, S. (2009). Improving the evaluation of therapeutic interventions in multiple sclerosis: the role of new psychometric methods. Health Technol. Assessment 13, 1–200. doi: 10.3310/hta13120

PubMed Abstract | CrossRef Full Text | Google Scholar

Hobart, J. C., Cano, S. J., Zajicek, J. P., and Thompson, A. J. (2007). Rating scales as outcome measures for clinical trials in neurology: problems, solutions, and recommendations. Lancet Neurol. 6, 1094–1105. doi: 10.1016/S1474-4422(07)70290-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Hoijtink, H. (1995). “Linear and repeated measures models for the person parameters,” in Rasch Models Foundations, Recent Developments, and Applications, eds G. H. Fischer and I. W. Molenaar (New York, NY: Springer-Verlag), 203–214. doi: 10.1007/978-1-4612-4230-7_11

CrossRef Full Text | Google Scholar

Hu, L. T., and Bentler, P. M. (1999). Cutoff criteria for fit indexes in covariance structure analysis: conventional criteria versus new alternatives. Struct. Equation Model. 6, 1–55. doi: 10.1080/10705519909540118

CrossRef Full Text | Google Scholar

JAM Press (2018). Journal of Applied Measurement. JAM Press. Retrieved from: http://jampress.org/JOM.htm (accessed July 25, 2019).

Google Scholar

Karabatsos, G. (2001). The Rasch model, additive conjoint measurement, and new models of probabilistic measurement theory. J. Appl. Measurement 2, 389–423.

PubMed Abstract | Google Scholar

Kelderman, H. (2007). “Loglinear multivariate and mixture Rasch models,” in Multivariate and Mixture Distribution Rasch Models: Extensions and Applications, eds M. von Davier and C. H. Carstensen (New York, NY: Springer-Verlag, 77–98.

Google Scholar

Kelderman, H., and Rijkes, C. P. M. (1994). Loglinear multidimensional IRT models for polytomously scored items. Psychometrika 59, 149–176. doi: 10.1007/BF02295181

CrossRef Full Text | Google Scholar

Kempf, W. (1972). Probabilistische Modelle experimentalpsychologischer Versuchssituationen. Psychol. Beitriige 14, 16–37.

Google Scholar

Klauer, K. C. (1995). “The assessment of person fit,” in Rasch Models: Foundations, Recent Developments, and Applications, eds G. H. Fischer and I. W. Molenaar (New York, NY: Springer), 97–110. doi: 10.1007/978-1-4612-4230-7_6

CrossRef Full Text | Google Scholar

Lamoureux, E. L., Pallant, J. F., Pesudovs, K., Hassell, J. B., and Keeffe, J. E. (2006). The impact of Vision Impairment Questionnaire: an evaluation of its measurement properties using Rasch analysis. Invest. Ophthalmol. Vis. Sci. 47, 4732–4741. doi: 10.1167/iovs.06-0220

PubMed Abstract | CrossRef Full Text | Google Scholar

Leydesdorff, L. (2005). Similarity measures, author cocitation analysis, and information theory. J. Am. Soc. Info. Sci. Technol. 56, 769–772. doi: 10.1002/asi.20130

CrossRef Full Text | Google Scholar

Leydesdorff, L., Rafols, I., and Chen, C. (2013). Interactive overlays of journals and the measurement of interdisciplinarity on the basis of aggregated journal–journal citations. J. Am. Soc. Info. Sci. Technol. 64, 2573–2586. doi: 10.1002/asi.22946

CrossRef Full Text | Google Scholar

Linacre, J. M. (1989). Many-faceted Rasch Measurement. Chicago: MESA Press.

Google Scholar

Linacre, J. M. (1998). Winsteps: History and Steps. Retrieved from: https://www.winsteps.com/winman/history.htm (accessed July 25, 2019).

Google Scholar

Linacre, J. M. (2002). Optimizing rating scale category effectiveness. J. Appl. Measurement 3, 85–106.

PubMed Abstract | Google Scholar

Linacre, J. M. (2004). From Microscale to Winsteps: 20 years of Rasch software development. Rasch Measurement Transac. 17:958.

Google Scholar

Linacre, J. M. (2009). A User's Guide to Winsteps-Ministep: Rasch-model Computer Programs. Program Manual 3.68.0. Oxford: Chicago, IL.

Google Scholar

Linacre, J. M. (2014). A User's Guide to FACETS Rasch-model Computer Programs. Retrieved from: https://www.winsteps.com/manuals.htm

Google Scholar

Linacre, J. M. (2019a). FACETS: Computer Program for Many Faceted Rasch Measurement (Version 3.82.1). Chicago, IL: Mesa Press.

Google Scholar

Linacre, J. M. (2019b). Winsteps® Rasch Measurement Computer Program User's Guide. Beaverton, OR. Available online at: Winsteps.com

Google Scholar

Lindsay, B., Clogg, C. C., and Grego, J. (1991). Semiparametric estimation in the Rasch model and related exponential response models, including a simple latent class model for item analysis. J. Am. Stat. Assoc. 86, 96–107. doi: 10.1080/01621459.1991.10475008

CrossRef Full Text | Google Scholar

Loevinger, J. (1965). Person and population as psychometric concepts. Psychol. Rev. 72, 143–155. doi: 10.1037/h0021704

PubMed Abstract | CrossRef Full Text | Google Scholar

Lord, F. M. (1952). A theory of test scores. Psychometr. Monograph 7:84. doi: 10.1002/j.1477-8696.1952.tb01448.x

CrossRef Full Text | Google Scholar

Lord, F. M., and Novick, M. R. (1968). Statistical Theories of Mental Test Scores. Reading, MA: Addison-Wesley.

Google Scholar

Mairesse, O., Damen, V., Newell, J., Kornreich, C., Verbanck, P., and Neu, D. (2019). The Brugmann fatigue scale: an analogue to the epworth sleepiness scale to measure behavioral rest propensity. Behav. Sleep Med. 17, 437–458. doi: 10.1080/15402002.2017.1395336

PubMed Abstract | CrossRef Full Text | Google Scholar

Mangione, C. M., Lee, P. P., Gutierrez, P. R., Spritzer, K., Berry, S., and Hays, R. D. (2001). Development of the 25-list-item national eye institute visual function questionnaire. Arch. Ophthalmol. 119, 1050–1058. doi: 10.1001/archopht.119.7.1050

CrossRef Full Text | Google Scholar

Massof, R. W. (2002). The measurement of vision disability. Optometry Vis. Sci. 79, 516–552. doi: 10.1097/00006324-200208000-00015

PubMed Abstract | CrossRef Full Text | Google Scholar

Massof, R. W., and Fletcher, D. C. (2001). Evaluation of the NEI visual functioning questionnaire as an interval measure of visual ability in low vision. Vis. Res. 41, 397–413. doi: 10.1016/S0042-6989(00)00249-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Masters, G. N. (1982). A Rasch model for partial credit scoring. Psychometrika 47, 149–174. doi: 10.1007/BF02296272

CrossRef Full Text | Google Scholar

Masters, G. N. (1988). “Partial credit model,” in Educational Research, Methodology and Measurement: An International Handbook, ed J. P. Keeves (Elmsford, NY: Pergamon Press, 292–297.

Google Scholar

Masters, G. N. (2010). “Partial credit model,” in Handbook of Polytomous Item Response Theory Models: Development and Applications, eds M. Nering and R. Ostini (New York, NY: Routledge Academic Press, 109–122.

Google Scholar

McHorney, C. A. (1997). Generic Health Measurement: past accomplishments and a measurement paradigm for the 21st century. Ann. Intern. Med. 127 (8 Pt 2), 743–750. doi: 10.7326/0003-4819-127-8_Part_2-199710151-00061

PubMed Abstract | CrossRef Full Text | Google Scholar

McNamara, T., and Knoch, U. (2012). The Rasch wars: the emergence of Rasch measurement in language testing. Language Testing 29, 555–576. doi: 10.1177/0265532211430367

CrossRef Full Text | Google Scholar

Meijer, R. R., and Sijtsma, K. (2001). Methodology review: evaluating person fit. Appl. Psychol. Measurement 25, 107–135. doi: 10.1177/01466210122031957

CrossRef Full Text | Google Scholar

Micko, H.-C. (1969). A psychological scale for reaction time measurement. Acta Psychol. 30:324. doi: 10.1016/0001-6918(69)90057-2

CrossRef Full Text | Google Scholar

Micko, H.-C. (1970). Eine Verallgemeinerung des Messmodells von Rasch miteiner Anwendung auf die Psychophysik der Reaktionen [A generalization of Rasch's measurement model with an application to the psychophysics of reactions]. Psychol. Beitriige 12, 4–22.

Google Scholar

Newman, M.E.J. (2006). Modularity and community structure in networks. PNAS 103, 8577–8582. doi: 10.1073/pnas.0601602103

PubMed Abstract | CrossRef Full Text | Google Scholar

Olsen, L. W. (2003). Essays on Georg Rasch and his contributions to statistics. (Unpublished PhD Thesis), Copenhagen, Institute of Economics, University of Copenhagen.

Google Scholar

Pallant, J. F., and Bailey, C. M. (2005). Assessment of the structure of the Hospital Anxiety and Depression Scale in musculoskeletal patients. Health Qual. Life Outcomes 3:82. doi: 10.1186/1477-7525-3-82

PubMed Abstract | CrossRef Full Text | Google Scholar

Pallant, J. F., and Tennant, A. (2007). An introduction to the Rasch measurement model: an example using the Hospital Anxiety and Depression Scale (HADS). Br. J. Clin. Psychol. 46, 1–18. doi: 10.1348/014466506X96931

PubMed Abstract | CrossRef Full Text | Google Scholar

Panayides, P., Robinson, C., and Tymms, P. (2010). The assessment revolution that has passed England by: Rasch measurement. Br. Edu. Res. J. 36, 611–626. doi: 10.1080/01411920903018182

CrossRef Full Text | Google Scholar

Patz, R. J., and Junker, B. W. (1999). Applications and extensions of MCMC in IRT: multiple item types, missing data, and rated responses. J. Edu. Behav. Stat. 24, 342–366. doi: 10.3102/10769986024004342

CrossRef Full Text | Google Scholar

Pesudovs, K. (2006). Patient-centred measurement in ophthalmology–a paradigm shift. BMC Ophthalmol. 6:25. doi: 10.1186/1471-2415-6-25

PubMed Abstract | CrossRef Full Text | Google Scholar

Pesudovs, K. (2010). Item banking: a generational change in patient-reported outcome measurement. Optometry Vis. Sci. 87, 285–293. doi: 10.1097/OPX.0b013e3181d408d7

PubMed Abstract | CrossRef Full Text | Google Scholar

Pesudovs, K., Burr, J. M., Harley, C., and Elliott, D. B. (2007). The development, assessment, and selection of questionnaires. Optometry Vis. Sci. 84, 663–674. doi: 10.1097/OPX.0b013e318141fe75

PubMed Abstract | CrossRef Full Text | Google Scholar

Pesudovs, K., Garamendi, E., and Elliott, D. B. (2004). The quality of life impact of refractive correction (QIRC) questionnaire: development and validation. Optometry Vis. Sci. 81, 769–777. doi: 10.1097/00006324-200410000-00009

PubMed Abstract | CrossRef Full Text | Google Scholar

Pesudovs, K., Garamendi, E., Keeves, J. P., and Elliott, D. B. (2003). The Activities of Daily Vision Scale for cataract surgery outcomes: re-evaluating validity with Rasch analysis. Invest. Ophthalmol. Vis. Sci. 44, 2892–2899. doi: 10.1167/iovs.02-1075

PubMed Abstract | CrossRef Full Text | Google Scholar

Raîche, G. (2005). Critical eigenvalue sizes in standardized residual principal components analysis. Rasch Measurement Transac. 19:1012. doi: 10.1016/B978-012471352-9/50004-3

CrossRef Full Text | Google Scholar

Rasch, G. (1960). Probabilistic Models for Some Intelligence and Attainment Tests. Copenhagen: Danish Institute for Educational Research.

Google Scholar

Rasch, G. (1961). “On general laws and the meaning of measurement in psychology,” in Proceedings of the IV. Berkeley Symposium on Mathematical Statistics and Probability, Vol. IV (Berkeley: University of California Press), 321–333.

Google Scholar

Rasch, G. (1963). “The poisson process as a model for a diversity of behavioral phenomena,” in International Congress of Psychology, Vol. 2 (Washington, DC).

Google Scholar

Rasch, G. (1965). Statistical Seminar. Copenhagen: Department of Statistics, University of Copenhagen.

Google Scholar

Rost, J. (1990). Rasch models in latent classes: an integration of two approaches to item analysis. Appl. Psychol. Measurement 14, 271–282. doi: 10.1177/014662169001400305

CrossRef Full Text | Google Scholar

Rost, J. (1991). A logistic mixture distribution model for polychotomous item responses. Br. J. Math. Stat. Psychol. 44, 75–92. doi: 10.1111/j.2044-8317.1991.tb00951.x

CrossRef Full Text | Google Scholar

Rousseeuw, P. J. (1987). Silhouettes: a graphical aid to the interpretation and validation of cluster analysis. J. Comput. Appl. Math. 20, 53–65. doi: 10.1016/0377-0427(87)90125-7

CrossRef Full Text | Google Scholar

Rowe, V. T., Winstein, C. J., Wolf, S. L., and Woodbury, M. L. (2017). Functional test of the Hemiparetic Upper Extremity: a Rasch analysis with theoretical implications. Arch. Phys. Med. Rehabil. 98, 1977–1983. doi: 10.1016/j.apmr.2017.03.021

PubMed Abstract | CrossRef Full Text | Google Scholar

Scheiblechner, H. (1971). A Simple Algorithm for CML-Parameter-Estimation in Rasch's Probabilistic Measurement Model with Two or More Categories of Answers. Vienna: Department of Psychology, University of Vienna.

Google Scholar

Sica da Rocha, N., Chachamovisch, E., de Almeida Fleck, M. P., and Tennant, A. (2013). An introduction to Rasch analysis for psychiatric practice and research. J. Psychiatr. Res. 47, 141–148. doi: 10.1016/j.jpsychires.2012.09.014

CrossRef Full Text | Google Scholar

Sitgreaves, R. (1960). Intraclass Correlation and the Analysis of Variance.

Google Scholar

Small, H., and Greenlee, E. (1986). Collagen research in the 1970s. Scientometrics 10, 95–117. doi: 10.1007/BF02016863

CrossRef Full Text | Google Scholar

Small, H., and Sweeney, E. (1985). Clustering thescience citation index® using co-citations. Scientometrics 7, 391–409. doi: 10.1007/BF02017157

CrossRef Full Text | Google Scholar

Smith, E. Jr. (2002). Detecting and evaluating the impact of multidimensionality using item fit statistics and principal component analysis of residuals. J. Appl. Measurement 3, 205–231.

PubMed Abstract | Google Scholar

Smith, E. V. Jr. (2000). Metric development and score reporting in Rasch measurement. J. Appl. Measurement 1:303.

PubMed Abstract | Google Scholar

Smith, R. M. (2019). The ties that bind with an invitation for contributions. Rasch Measurement Transac. 32, 1714–1716.

Google Scholar

Tennant, A. (2011). Appearance of “Rasch” in journal articles. Rasch Measurement Transac. 24:1311.

Google Scholar

Tennant, A., and Conaghan, P. G. (2007). The Rasch measurement model in rheumatology: what is it and why use it? When should it be applied, and what should one look for in a Rasch paper? Arthritis Rheumat. 57, 1358–1362. doi: 10.1002/art.23108

PubMed Abstract | CrossRef Full Text | Google Scholar

Tennant, A., Hillman, M., Fear, J., Pickering, A., and Chamberlain, M. A. (1996). Are we making the most of the Stanford Health Assessment Questionnaire? Rheumatology 35, 574–578. doi: 10.1093/rheumatology/35.6.574

PubMed Abstract | CrossRef Full Text | Google Scholar

Tennant, A., McKenna, S. P., and Hagell, P. (2004a). Application of Rasch analysis in the development and application of quality of life instruments. Value Health 7, S22–S26. doi: 10.1111/j.1524-4733.2004.7s106.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Tennant, A., and Pallant, J. F. (2006). Unidimensionality matters!(A tale of two Smiths?). Rasch Measurement Transac. 20, 1048–1051.

Google Scholar

Tennant, A., Penta, M., Tesio, L., Grimby, G., Thonnard, J. L., Slade, A., et al. (2004b). Assessing and adjusting for cross cultural validity of impairment and activity limitation scales through Differential Item Functioning within the framework of the Rasch model: the Pro-ESOR project. Medical Care 42 (Suppl. 1), 37–48. doi: 10.1097/01.mlr.0000103529.63132.77

PubMed Abstract | CrossRef Full Text | Google Scholar

Tesio, L., Simone, A., and Bernardinello, M. (2007). Rehabilitation and outcome measurement: where is Rasch analysis-going? Europa Medicophys. 43, 417–26.

PubMed Abstract | Google Scholar

Thorndike, E.L. (1919). An Introduction to the Theory of Mental and Social Measurements. New York, NY: Columbia University, Teachers College.

Google Scholar

Thurstone, L.L. (1925). A method of scaling psychological and educational tests. J. Edu. Psychol. 16, 433–451. doi: 10.1037/h0073357

CrossRef Full Text | Google Scholar

van der Linden, W. J., and Hambleton, R. K. (1997). Handbook of Modern Item Response Theory. New York, NY: Springer Science & Business Media.

Google Scholar

Velozo, C. A., Magalhaes, L. C., Pan, A. W., and Leiter, P. (1995). Functional scale discrimination at admission and discharge: Rasch analysis of the Level of Rehabilitation Scale-III. Arch. Phys. Med. Rehabil. 76, 705–712. doi: 10.1016/S0003-9993(95)80523-0

PubMed Abstract | CrossRef Full Text | Google Scholar

von Davier, M. (1996). “Mixtures of polytomous Rasch models and latent class models for ordinal variables,” in Softstat 95—Advances in Statistical Software 5, eds F. Faulbaum and W. Bandilla (Stuttgart: Lucius and Lucius).

Google Scholar

von Davier, M. (2001). WINMIRA [Computer Program]. Groningen: ASCAssessment Systems Corporation, USA and Science Plus Group.

Google Scholar

von Davier, M., and Yamamoto, K. (2007). “Mixture-distribution and HYBRID Rasch models,” in Multivariate and Mixture Distribution Rasch Models: Extensions and Applications, eds M. von Davier and C. H. Carstensen (New York, NY: Springer-Verlag), 99–118. doi: 10.1007/978-0-387-49839-3_6

CrossRef Full Text | Google Scholar

Wang, W. C., Wilson, M., and Adams, R. J. (1997). Rasch models for multidimensionality between items and within items. Objective measurement: Theory Pract. 4, 139–155.

Google Scholar

White, H. D., and McCain, K. W. (1998). Visualizing a discipline: an author co-citation analysis of information science, 1972–1995. J. Am. Soc. Info. Sci. 49, 327–355.

Google Scholar

Wilson, M. (2005). Constructing Measures: An Item Response Modeling Approach. Mahwah, NJ: Lawrence Erlbaum Associates.

Google Scholar

Wright, B. (1997). A History of Social Science Measurement. Retrieved from: https://www.rasch.org/memo62.htm (accessed July 25, 2019).

Google Scholar

Wright, B. D. (1996). Key events in Rasch measurement history in America, Britain and Australia (1960–1980). Rasch Measurement Transac. 10, 494–496.

Google Scholar

Wright, B. D., Mead, R. J., and Bell, S. R. (1978). BICAL: Calibrating Items with the Rasch Model. Retrieved from: https://www.rasch.org/ (accessed July 25, 2019).

Google Scholar

Wright, B. D., and Stone, G. (1979). Best Test Design. Chicago: MESA Press.

Google Scholar

Wu, M. L., Adams, R., and Wilson, M. (1998). ACER ConQuest (Version 1.0) [Computer Package]. Melbourne, VIC: Australian Council for Educational Research.

Google Scholar

Wu, M. L., Adams, R., Wilson, M., and Haldan, S. (2007). ACER ConQuest Version 2.0: Generalised Item Response Modelling Software. Melbourne, VIC: ACER.

Google Scholar

Zhao, D., and Strotmann, A. (2008). Information science during the first decade of the web: an enriched author cocitation analysis. J. Am. Soc. Info. Sci. Technol. 59, 916–937. doi: 10.1002/asi.20799

CrossRef Full Text | Google Scholar

Keywords: burst, co-citation analysis, Rasch measurement, review, Scientometrics

Citation: Aryadoust V, Tan HAH and Ng LY (2019) A Scientometric Review of Rasch Measurement: The Rise and Progress of a Specialty. Front. Psychol. 10:2197. doi: 10.3389/fpsyg.2019.02197

Received: 09 July 2019; Accepted: 12 September 2019;
Published: 22 October 2019.

Edited by:

Michela Balsamo, Università degli Studi G. d'Annunzio Chieti e Pescara, Italy

Reviewed by:

George Engelhard, University of Georgia, United States
John Siegert, Auckland University of Technology, New Zealand
Chung-Ying Lin, Hong Kong Polytechnic University, Hong Kong

Copyright © 2019 Aryadoust, Tan and Ng. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Vahid Aryadoust, vahid.aryadoust@nie.edu.sg

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.