REVIEW article

Front. Res. Metr. Anal., 23 August 2018

Sec. Research Policy and Strategic Management

Volume 3 - 2018 | https://doi.org/10.3389/frma.2018.00022

External Tests of Peer Review Validity Via Impact Measures

  • American Institute of Biological Sciences, McLean, VA, United States

Abstract

Peer review is used commonly across science as a tool to evaluate the merit and potential impact of research projects and make funding recommendations. However, potential impact is likely to be difficult to assess ex-ante; some attempts have been made to assess the predictive accuracy of these review decisions using impact measures of the results of the completed projects. Although many outputs, and thus potential measures of impact, exist for research projects, the overwhelming majority of evaluation of research output is focused on bibliometrics. We review the multiple types of potential impact measures with an interest in their application to validate review decisions. A review of the current literature on validating peer review decisions with research output impact measures is presented here; only 48 studies were identified, about half of which were US based and sample size per study varied greatly. 69% of the studies employed bibliometrics as a research output. While 52% of the studies employed alternative measures (like patents and technology licensing, post-project peer review, international collaboration, future funding success, securing tenure track positions, and career satisfaction), only 25% of all projects used more than one measure of research output. Overall, 91% of studies with unfunded controls and 71% of studies without such controls provided evidence for at least some level of predictive validity of review decisions. However, several studies reported observing sizable type I and II errors as well. Moreover, many of the observed effects were small and several studies suggest a coarse power to discriminate poor proposals from better ones, but not amongst the top tier proposals or applicants (although discriminatory ability depended on the impact metric). This is of particular concern in an era of low funding success, where many top tier proposals are unfunded. More research is needed, particularly in integrating multiple types of impact indicators in these validity tests, as well as considering the context of the research outputs relative to goals of the research program and concerns for reproducibility, translatability and publication bias. In parallel, more research is needed focusing on the internal validity of review decision making procedures and reviewer bias.

Grant peer review and impact assessment

Most would generally agree the purpose of biomedical research is to advance knowledge for societal benefit, with the hope of favorably impacting disease outcomes and improving global health. Indeed, the National Institutes of Health (NIH), the world's largest funder of biomedical research, characterizes their mission as to “seek fundamental knowledge about the nature and behavior of living systems and the application of that knowledge to enhance health, lengthen life, and reduce illness and disability” (NIH, 2017). To help select which research projects to fund to achieve this goal, NIH and other funders rely on a peer review process to assess the quality of the research approach and methodologies proposed, the feasibility of the investigators successfully conducting the project in the proposed environment, and the level of innovation and potential significance of the project (NIH, 2014). Of these criteria, it is likely the most difficult to accurately assess is the potential significance; particularly “if the aims of the project are achieved, how will scientific knowledge, technical capability, and/or clinical practice be improved” and “how will successful completion of the aims change the concepts, methods, technologies, treatments, services, or preventative interventions that drive this field?” (NIH, 2016).

In no small part, this is due to the role of serendipity in science, which has been identified as an important component in scientific discovery (Ban, 2006; Merton and Barber, 2011; Editorial, 2018), as well as a variety of unforeseen factors which may prevent the success of a research project. Thus, even in the best of cases, the potential impact of a research project may be difficult to gauge. However, there are also reports that the decision-making process can be hampered by subjectivity and the presence of biases (Marsh et al., 2008; Ginther et al., 2011; Lee et al., 2013; Boudreau et al., 2016; Kaatz et al., 2016). As one of the chief goals of peer review is to select projects for funding of the highest scientific quality that are likely to have the greatest impact, it stands to reason that objective measurements of the actual impact of fully funded and completed projects could be assessed ex-post funding and compared to peer review evaluations, so that we may determine the predictive validity of these decisions. Similarly, objective indicators of proposal quality (e.g., track record of the applicant) could be assessed ex-ante to funding to be compared to review decisions. These external tests of validity, which compare scientific inputs and outputs to review evaluations, likely offer an important assessment of the effectiveness of review decisions in choosing the best science, although admittedly do not necessarily validate other expectations of peer review, like impartiality (Wood and Wessely, 2003).

However, a central question in scientometrics is how best to evaluate research, as many metrics have considerable limitations or are influenced by a variety of factors that are not associated with research quality or scientific impact (Nieminen et al., 2006; Bornmann et al., 2008a; Hagen, 2008; Leydesdorff et al., 2016). For instance, citation levels are influenced by the number of co-authors, journal prestige and even by whether the results are positive or negative (Callaham et al., 2002; Dwan et al., 2008; Ioannidis, 2008). Moreover, for biomedical research, the societal impact of a study is not only measured in its contribution to the knowledge base (Bornmann, 2017), but also in actual improvements to human health; however, linking the influence of individual works to the development of new therapeutics is problematic, as they rely on large bodies of work through their evolution from bench to bedside (Keserci et al., 2017). Nevertheless, as the recent Leiden manifesto points out, performance measurements should “take into account the wider socio-economic and cultural contexts,” and that the “best practice uses multiple indicators to provide a more robust and pluralistic picture” (Hicks et al., 2015).

Thus, it seems a variety of impact measures should potentially be used to validate review decisions. However, at this time there has been no comprehensive review of studies in the literature, across a variety of impact measures, that have attempted to validate peer review decisions. We will review here many of these measures below, examining what has been done with respect to peer review of research funding applications, what measures still need to be explored, and what has been done to integrate these measures to achieve a more well-rounded assessment of research success and failures. It should be noted that this literature review is focused on work that is application based. That is, it includes studies that examine the ranking and funding fate of applications and applicants relative to either the quality of the input or the impact of the output from those applications and applicants after the funding decision across a variety of measures (Figure 1). Again, this includes only measures of external validity (external scientific quality measures for outputs and inputs) and not the internal validity of review procedures (e.g., bias, inter-rater reliability), which is beyond the scope of this review. It is based on the knowledgeable selection of relevant publications which includes both peer reviewed and non-peer reviewed articles, as some of this work has been conducted by funding agencies and published in non-traditional forums.

Figure 1

Publication productivity and citation impact

The most studied research outputs are bibliometric in nature, surrounding the number of published manuscripts, the impact of the journals they were published in, the raw and normalized citation levels of these manuscripts (normalized for time and research field), the h-indices of applicants and number of manuscripts in the top 10% of all other cited papers on the topic as well as citations and papers per dollar spent (Mavis and Katz, 2003; Van Noorden, 2010; Danthi et al., 2014; Li and Agha, 2015). As mentioned above, there are limitations to bibliometric indicators due to their complex nature and may not always reflect long term impact (Wang et al., 2013). Nevertheless, this is where much of the effort to study the validation of peer review has focused. Several types of similarly structured studies have resulted, which are summarized below.

Ex ante impact of applicants (funded vs. unfunded or review score)

In the last few years, several attempts have been made to examine the number of publications and their citation impact from funded and unfunded applicants. Several studies have tracked individual applicant ex ante performance before funding decisions to determine if reviewers can pick applicants with superior prior publication and citation performance. This is a powerful strategy as you can directly compare funded and unfunded applicants, and do not have to consider the effect of funding as a confounding factor on performance. Most studies show that overall funded applicants outperform unfunded (Bornmann and Daniel, 2006; van den Besselaar and Leydesdorff, 2009; Bornmann et al., 2010; van Leeuwen and Moed, 2012; Cabezas-Clavijo et al., 2013) and a few studies do not (Hornbostel et al., 2009; Neufeld et al., 2013; Saygitov, 2014), although typically the differences are small and dependent on the general quality level of applicants (if all applicants are very productive, smaller differences will be observed). A couple of studies examined the ex-ante productivity of applicants relative to review scores, and found significant correlations, as well as significant biases (Wenneras and Wold, 1997; Sandstrom and Hallsten, 2008). Also, some studies show when you compare the best of unfunded applicants with funded ex ante, they are comparable (van den Besselaar and Leydesdorff, 2009; Bornmann et al., 2010; Neufeld et al., 2013), suggesting significant type II error. Some of these studies have been summarized well by Boyack et al. (2018) as well as Van den Besselaar and Sandstrom (2015). Thus, these results may suggest that while peer review may be efficient at coarse discrimination between bad and good applicants, it may be limited in its ability for fine discrimination between good and excellent applicants. However, only looking at ex-ante results makes no comment on how applicants actually perform in the future, which is what reviewers are predicting via their score, therefore it is important to make ex-post observations as well.

Ex post impact of applicant and project (funded vs. unfunded)

Some studies examine the productivity of funded applicants ex post in comparison to unfunded, to see if reviewers chose applicants that in the end were productive. Multiple studies show that funded applicants are at least modestly more productive and more frequently cited after the award as compared to unfunded (Armstrong et al., 1997; Mavis and Katz, 2003; Mahoney et al., 2007; Bornmann et al., 2008b, 2010; Pion and Cordray, 2008; Reinhart, 2009; Campbell et al., 2010; Jacob and Lefgren, 2011a,b; Langfeldt et al., 2012; Robitaille et al., 2015; Van den Besselaar and Sandstrom, 2015; Gush et al., 2017), although some do not (Saygitov, 2014). Interpretation of these results is difficult because it is challenging to dissociate the productivity effect of funding from the validity of the review decision. However, while general research funding is related to scientific productivity and knowledge production (Lauer, 2015; Rosenbloom et al., 2015) and papers with funding acknowledgments are linked to higher citation counts (Gok et al., 2016), the effect of specific funding on an individual's productivity is not clear; some research looking at ex ante and ex post bibliographic levels for funded applicants show no effect of funding at all (Langfeldt et al., 2012; Robitaille et al., 2015), although it seems the length of time used to capture ex post bibliometric data is an important factor (Van den Besselaar and Sandstrom, 2015). Once again, many of these studies show significant type II errors (where unfunded applicants perform well) and sometimes only limited or no differences are found between funded and unfunded applicants with similar review scores or performance (Bornmann et al., 2008b, 2010; Pion and Cordray, 2008; Jacob and Lefgren, 2011a; Van den Besselaar and Sandstrom, 2015; Gush et al., 2017) although some similar comparisons do find differences (Robitaille et al., 2015).

These ex post studies are related to the above ex ante results in that some literature has indicated that one of the strongest predictors of future citation performance is prior citation performance (Kaltman et al., 2014; Hutchins et al., 2016). Thus again, if peer review selects for applicants with higher previous productivity, it stands to reason that their post-funding productivity will be higher than unfunded applicants as well. While this could be interpreted as further validation of the peer review process, the assumption is that some investigators are simply more inherently productive than others. However, this could also be interpreted as the Matthew effect, where the rich get richer; the subset with access to research funding have more opportunities to be productive, which leads to more funding, more security and prestige, and therefore better bibliometric output (Merton, 1968; Azoulay et al., 2013), although some studies find no evidence of this (Boyack et al., 2018). In addition, many grant proposals are judged around the assessment of a research idea and its methodological implementation, not just the investigator's track record. Thus, it is unclear looking at an individual's career productivity alone may be an appropriate measure of success to validate review decisions; analysis of ex post productivity of individual projects is also required.

Ex post impact of funded project vs. review score (no unfunded control)

Similar studies have been performed with projects, although admittedly these are harder to conduct as productivity and impact data from unfunded projects is impossible to access or difficult to interpret. Largely what has been done is to analyze the relative confidence in funding decisions (peer review scores) of funded projects and how these relate to citation impact. One issue has been how results are normalized and computed. For instance, several studies of NIH NHLBI data calculated output results on a per dollar spent basis, as some research (Berg, 2011; Fortin and Currie, 2013; Gallo et al., 2014) has predicted diminishing returns with larger investments; these studies found no correlation between review scores and output (Danthi et al., 2014; Doyle et al., 2015). However, a large NIH study of unnormalized bibliometric data found a moderate correlation (Li and Agha, 2015). In fact, several other studies using normalized and unnormalized citation impact measures also suggested a moderate correlation (Berg, 2011; Gallo et al., 2014). When NIH data were reanalyzed without using budget normalized citation impact, a moderate correlation was observed (Lauer et al., 2015). A few other studies have found no correlation between scores and citation impact, although one was a very small sample (Scheiner and Bouchie, 2013) and the other was from the second round of review, so the level of quality across these projects was already very high (Gush et al., 2017). In fact, similar results were found with NIH data (same data set as used by Li and Agha, 2015); if the poorer scoring applications were removed from the analysis to reflect current funding rates, correlations between output and review scores disappeared (Fang et al., 2016). Again, this suggests the coarse discrimination of peer review in separating good projects from poor ones, but not good from great.

One constant in all of these analyses is the high degree of variability in grant output and impact across projects. This variability reflects the complicated and potentially biased nature of bibliometrics (e.g., dependencies on field, number of authors, and on the research results themselves), but also the role of serendipity in science (not every discovery receives the same reception, and while breakthrough discoveries are rare, they often stand on the shoulders of previous less-cited research). Many attempts to normalize citation counts for confounding factors have been made (h-index, Hirsch, 2005, m-index, Bornmann et al., 2008c; RCR, Hutchins et al., 2016) but each method has strengths and weaknesses (Van Noorden, 2010). Complicating this is the observation that reviewers may treat higher risk projects differently than straightforward ones (Boudreau et al., 2016). Given this bibliometric complexity and the inherent riskiness of research projects, a strong correlation between peer review scores and citation patterns, where better scores predict high performance projects may be unattainable. In fact, some groups have asserted that “retrospective analyses of the correlation between percentile scores from peer review and bibliometric indices of the publications resulting from funded grant applications are not valid tests of the predictive validity of peer review” (Lindner and Nakamura, 2015), as citation values many times are higher for “exaggerated or invalid results” and that papers are often selected for citation based on their “rhetorical utility” and not “primarily based on their relevance or validity.”

Type I/ II error and peer review scores

While it may not be clear how to define relative success of productive projects, it is easily achievable to determine which projects published anything at all. To date there has not been an exploration of the relationship between peer review scores of projects and the likelihood of unproductive grants (funded projects yielding no publication output), despite suggestions that the failure is an important aspect to the scientific process of discovery (Firestein, 2015). To address this issue, we have re-analyzed previously published data focusing on the frequency of non-producing grants and its relationship to score. The data used in this analysis came from independent peer reviews of 227 R01 style awards (4 year, $1 million awards) funded from an anonymous biomedical research program (Gallo et al., 2014). We define type I error (ratio of unproductive grants/all grants of a given score) as projects that are funded but ultimately yield no publications after funding is completed and the grant is closed. Projects are rated on a scientific merit (SM) scale of 1–5 (1 being most meritorious). In Figure 2 below, we observe a moderate level of correlation between the proportion of funded projects with zero resultant publications and peer review score (R2 = 0.23; p = 0.07), with better scoring grants yielding lower error rates than poorer scoring grants (removal of the outlier at 1.2 yields an R2 = 0.58; p = 0.002). Across the entire scoring range, the overall type I rate was 33%, with unproductive grants having a median score of 1.9 ± 0.05, vs. 1.7 ± 0.03 for productive grants (non-zero). Others have defined type I errors as lower than median performance for funded projects (using metrics like the h-index) and have estimated these values at 26–37% (Bornmann et al., 2008b), which is similar to that observed here, albeit using a less generous cut-off. The fact that nearly a third of grants were unproductive and yet 50% of those unproductive grants scored a 1.9 or better perhaps speaks not only to the level of quality but also to the level of risk involved in research projects, and that flaws which impact the score of an application may also increase the risk of unproductive projects. Indeed, some studies have suggested more novel (but potentially higher risk) applications are penalized in review score (Boudreau et al., 2016).

Figure 2

The rate of false negatives, or type II error, could be defined as unfunded projects that were eventually completed and were highly productive. This is clearly a more difficult aspect to measure, as there are few follow-up data linking unfunded applications and their ideas to post-review publications. As such, few studies exist assessing type II error, although some attempts have been made tracking the h-indices of successful and unsuccessful applicants, estimating type II rates as 32–48% (Bornmann et al., 2008b). Type II errors are probably highly dependent on funding success rates. While it has been shown that reviewers agree more about what should not be funded than what should (Cole and Simon, 1981), it is likely that as scores approach the funding line, there will be higher levels of type II error, which may result in a graph similar to Figure 2, although there are no such studies in the literature currently.

Social media impact (altmetrics)

Most publishers now enable the use of altmetrics to capture the number of tweets and other social media posts about articles, as well as capture download rates and page views. These dynamic metrics capture in real time another sense of impact, “quantifying the interest and debate an article generates, from the moment it is published (Warren et al., 2017).” While critics have mentioned that altmetrics are not yet validated and represent popularity, not necessarily impact, proponents suggest social media discussions represent a new, broader channel of communication that could reach beyond discipline and even increase engagement outside the scientific community (Sugimoto et al., 2017). Altmetrics have the capability to capture and quantify types of outputs that are missed by traditional bibliometrics. For instance, white papers and non-peer reviewed publications do not necessarily yield citations in Web of Science, but yet may be of great importance and influence on science policy. In addition, blogs, conference presentations and other alternate publications may be the only route to announce negative results, which may be unpublishable in traditional journals but are still useful and important products of the research. Although one study suggests funded research is viewed online more often than unfunded research (Didegah et al., 2017), and another has examined the relation between views, Twitter counts and post-publication peer review results of manuscripts (Bornmann, 2017), there are currently no studies in the literature directly looking at funding decisions and altmetrics (Dinsmore et al., 2014).

Collaboration-fostering of research teams

There is an argument to be made that high degrees of collaborations between scientists (especially interdisciplinary collaborations) addressing a common research objective yield higher creativity and innovation, as well as higher translatability (Carayol and Thi, 2005). Also, higher collaboration may enhance reproducibility (Munafo et al., 2017). Thus, tracking the actual level of collaboration (both that contained in the original proposal as well as ex post published co-authorships) may be important, especially if this is one of the goals of the research funding program. In fact, it has been shown that receiving more funding may be a result of increased collaboration (Ebadi and Schiffauerova, 2015) and may result in larger future collaborations (Adams et al., 2005). While research into collaborative scientific activities is extensive, only a few studies have looked at this directly with regard to peer review decisions; both Melin and Danell (2006) and Langfeldt et al. (2012) found successful applicants have a higher degree of ex-post international co-authorship than unsuccessful applicants and both El-Sawi et al. (2009) and Ubfal and Maffioli (2011) have found increased levels of collaboration amongst funded groups. However, Robitaille et al. (2015) found funded applicants had lower levels of ex post interdisciplinarity and Bromham et al. (2016) also notes that projects with greater interdisciplinarity have lower funding success, even for projects with high degrees of collaboration. This may be due to the risk that interdisciplinarity brings, as some results have shown increased novelty (presumably high risk) is penalized by reviewers (Boudreau et al., 2016). This small amount of data suggests perhaps that peer review decisions can validly select projects that yield high degrees of collaboration but are not necessarily promotional of interdisciplinary research, although it also seems clear much more work needs to be done on this subject.

Post-funding review of outcomes

A few studies included in this review have looked at peer review evaluation of post-funding performance and quality (Claveria et al., 2000; Mutz et al., 2015) and compared it to the ex-ante evaluation of proposals; both of these findings observed significant predictive validity of the review decisions (although the work of Mutz relies strongly on some methodological assumptions and may not represent an independent observation). Post-funding evaluations of productivity and impact likely take into account contextual factors of the research that are not represented in bibliometric numbers. The obvious downside is that conducting post-funding review panels is likely cost prohibitive, preventing its regular use. Post-publication peer review (PPPR) sites like PubPeer and F1000 may also be used to get a sense of trustworthiness and robustness of individual publications via the comments and ratings (Knoepfler, 2015). However, while one could conceivably achieve a high number of reviewers per publication and therefore a high degree of confidence in the results, there is concern for potentially low and inconsistent levels of engagement and for some reviewers, the lack of anonymity will be an issue (Dolgin, 2018). Administrative review post-funding can also be done at the funding agency level to at least determine whether a variety of non-bibliometric outcomes were achieved, which can include whether the work was finished or left incomplete, whether the stated goals were achieved, whether the results or products were disseminated (including through non-traditional pathways) and tracking the level of reproducibility of the results. One recent example of this is by Decullier et al. (2014), who found that clinical projects chosen to be funded by an agency were much more likely to be initiated than unfunded projects. However, once a project was initiated, the authors observed that the likelihood of completion was unaffected by funding status, as was whether publications would result, the timeline to publications and the number of publications. Therefore, straight interpretation of publication output may mask type II error, as the productivity level of unfunded but initiated projects was similar to that of funded ones. Thus, these types of measures provide crucial context to the interpretation of the results.

Patents/technology development

Patents have been used as indicators of research impact, although some studies find that only about 10% of NIH grants over the last 3 decades directly yielded a patent as a product and only about 30% have work which is cited in a patent (Li et al., 2017). Other studies have shown that, to bring 5 patented therapeutics through testing and to the market required more than 100,000 papers, and nearly 20,000 NIH grants (Keserci et al., 2017; other funding sources not considered). In addition, some have argued that linkages between patents and the literature should not only rely on direct citation linkages, but on mapping analysis of whole bodies of work surrounding a concept to determine the influence of an individual (Gurney et al., 2014), further complicating analysis. Thus, attributing an individual grant to the creation and subsequent impact of a patent may be difficult, as not only do multiple research inputs cumulatively produce a patent, the success rate for producing an actual therapeutic in the market is very low (Stevens and Burley, 1997).

Nevertheless, some research has been conducted observing the predictive association of peer review scores of funded grant applications and patent production (Li and Agha, 2015); finding a decrease in score of one standard-deviation yielding 14% fewer patents. Galbraith et al. (2006) also compared peer review scores of individual funded projects to their ultimate success utilizing two metrics: (1) cooperative research and development agreements (CRADA) or licenses that were signed, SBIR or equity funding that was obtained or a product that was launched; and (2) the assessment of a senior project manager (not an author) of each technology as successful (evaluated one and a half to 3 years after the initial peer review evaluation). Using 69 early to mid-stage homeland defense technologies funded by the US DoD Center for Commercialization of Advanced Technologies (CCAT), the authors found that reviewer scores were weakly predictive of commercial success of funded projects. However, Melin and Danell (2006) found that, for a Swedish research funding program aiming to develop research with industrial applications with large, 6-year grants, funded applicants generated more patents and more spin-off companies than unfunded applicants, although the sample is small with large variation in patent output (which may in part be due to the wide breadth of scientific fields). Chai and Shih (2016) also found that firms funded by an academic-industry partnership received significantly more patents than unfunded applicant firms, although the effects depended on the size and age of the firm. These results suggest some level of review validity, although it is still unclear how and to what extent the funding can promote patent creation. It may be the direct effect is small; while some have observed small positive impacts on patent generation (Payne and Siow, 2003) or on patent originality and impact (Huang et al., 2006; Guerzoni et al., 2014), some have found no effect or even a negative effect (Sanyal, 2003; Beaudry and Kananian, 2013). Thus, patent productivity has some promise for use in tests of review validity, however future studies will likely require more subtle, nuanced approaches.

Data sharing

An important output of research is sharable data sets, which some have suggested have “vast potential for scientific progress” by facilitating reproducibility and allowing new questions to be asked with old data sets (Fecher et al., 2015). In fact, data sharing is associated in some cases with increased citations rates (Piwowar et al., 2007). Yet, several studies have indicated the majority of researchers do not share their data, in part because of the lack of incentives (Tenopir et al., 2011; Fecher et al., 2015; Van Tuyl and Whitmire, 2016). Multiple platforms are available to share data through journal publication sites (e.g., PloS One) or even sites hosting unpublished manuscripts and data (e.g., Figshare). Various metrics, such as download rates or even citations of data usage can be used to potentially capture impact. Yet, while one study examined data management plans for funded and unfunded National Science Foundation (NSF) proposals and found no significant differences in plans to share data (Mischo et al., 2014), currently no studies have explored ex post data sharing and its relationship to peer review decisions.

Career tracking

Some have focused efforts on assessing impact of early career funding through tracking of PI careers, using ex-post NIH funding as a metric. One study of the Howard Hughes Medical Institute's (HHMI) research training programs for medical students found that funding through their program was associated with significantly increased levels of NIH post-doctoral funding success post-HHMI award (21%) as compared to a control group of unfunded HHMI applicants (13%; Fang and Meyer, 2003). It should be noted that funded applicants still had higher success than unfunded applicants despite similar ex-ante qualifications. In addition, when ex-ante peer review results were taken into account, similar results were also seen with the Doris Duke Charitable Foundation (DDCF) Clinical Scientist Development Award (CSDA), where a greater proportion of CSDA funded applicants received at least one R01 grant (62%) vs. highly ranked but unfunded CSDA applicants (42%; Escobar-Alvarez and Myers, 2013). Moreover, NIH itself has observed differences between similarly scored funded and unfunded K grant applicants and their relative success in acquiring additional NIH funding (56% for K grant awardees vs. 43% for unfunded; Mason et al., 2013). Similar results were found between similarly scored funded and unfunded applicants by Tesauro et al. (2013). Mavis and Katz (2003) also observed higher post-award funding rates for successful applicants compared to unsuccessful ones, although there was no control for review score. Similarly, others have shown that, despite similar qualifications, funded applicants are more successful in gaining future funding and securing tenure track positions compared to unfunded applicants (Bol et al., 2018; Heggeness et al., 2018). However, many of these observations may be the result of the funding itself enabling future funding, as well as lowered levels of resubmissions by unfunded applicants. If possible, the effect of funding itself needs to be addressed in these tests, possibly by utilizing review scores to compare the amount of funded applicant's ex post funding success, although no such studies have been done.

Other metrics along the same vein have been used as well, including career satisfaction and faculty positions attained, both of which have been observed to be higher among funded applicants compared to similarly high ex-ante performing unfunded applicants (Hornbostel et al., 2009; Bloch et al., 2014; Van den Besselaar and Sandstrom, 2015). However, while Pion and Ionescu-Pioggia (2003) also found funded applicants of the Burroughs Welcome Career Award were more successful than unfunded in securing faculty positions and in acquiring future NIH funding (Pion and Cordray, 2008), these effects were diminished when adjusted for the ex-ante qualifications of the applicants. Career satisfaction is another variable to be tracked, although only two studies have examined this (Hornbostel et al., 2009; Langfeldt et al., 2012), tracking satisfaction via survey. While these groups found higher levels of satisfaction associated with funded applicants, there were no ex-ante controls for this measure and may be a result of the funding itself. Similarly, while (Langfeldt et al., 2012) has also monitored the number of successful graduate theses created stemming from funded applicants, again this work lacks the appropriate control to address peer review decisions. On the whole, while many of these results contrast bibliometric results above (given the high level of discrimination between competitive applicants), it is clear that future studies need to de-couple the effects of funding itself from the review decision before this measure can truly test review validity.

Integration of multiple impact metrics

Including a panel of indicators is likely to give a clearer picture of impact (Hicks et al., 2015), but they still need to be interpreted in the qualitative context of the science and the funding program (Chen, 2016), and the “right balance between comprehensiveness and feasibility must be struck” when determining how many and what type of indicators to include (Milat et al., 2015). In addition, just as reviewers weigh the relative importance of review criteria, how one weighs the importance of each indicator into the overall picture of impact is of crucial importance (Lee, 2015; Milat et al., 2015). Thus, integration of this information within a specific research context is crucial to getting an accurate picture of impact, but this is still represents a largely unexplored area, particularly with regard to validating peer review. One example of the use of multiple indicators in our survey was by Melin and Danell (2006), who found that subsequent to funding, while the number of publications was no different, funded applicants published in higher quality journals, as well as received more external funding for their group, produced more spin-off companies and produced more patents. Similarly, Hornbostel et al. (2009) found minor differences in bibliometric impact and output between funded and unfunded groups, yet both career satisfaction and number of faculty positions gained are higher among the funded group. Similar results are seen for Van den Besselaar and Sandstrom (2015).

Thus, the use of multiple indicators allows sensitivity to the multidimensional aspects of research impact. While it is likely the panel of most useful indicators will vary across research programs and funding goals, the methods for integrating these variables will vary as well. Some have argued that future holistic evaluation frameworks will need to involve qualitative and quantitative aspects of research quality and impact as well as peer and end-user evaluation to truly capture the public value of research (Donovan, 2007). In this vein, the Payback framework, which gauges “not just outputs but also outcomes derived from over a decade of investment” and takes into account the latency of impact and the attribution to multiple sources, has been suggested as best practice in research evaluation (Donovan, 2011). This framework integrates data from knowledge creation, benefits to future research, political benefits, health sector benefits and economic benefits (Bornmann, 2013). One downside to this very comprehensive approach is its labor-intensive nature and may not be relevant to assessment of individual projects. Others have focused on quantitating productive interactions between scientists and stakeholders, which is postulated to be a key generator of societal impact, although some have called for more studies to confirm this assumption (Molas-Gallart et al., 2000; Spaapen and Van Drooge, 2011; Bornmann, 2013; De Jong et al., 2014). One challenge to these types of integrations is the identification of criteria and measurable indicators for feasible assessment, and several frameworks have been suggested to address this (Sarli et al., 2010; Luke et al., 2018). Nevertheless, no standard method has been created that “can measure the benefit of research to society reliably and with validity” (Bornmann, 2017). Further, most evaluations of impact fail to take into account “inequality, random chance, anomalies, the right to make mistakes, unpredictability and a high significance of extreme events” which are hallmarks of the scientific process and likely distort any measurements of impact (Bornmann, 2017). Finally, the effect such impact assessment has on funding incentives is non-trivial, and likely influences ex-ante peer review decisions (Lindner and Nakamura, 2015; Bornmann, 2017); an important consideration when attempting to validate the peer review process.

Overview analysis of peer review validation studies

Table 1 lists the collection of papers we identified examining the validity of peer review decisions through research outputs, which were published over the last 21 years, with a median age of 6.5 years. In general, studies had to have access to funding decisions or peer review scores or both and their relationship to external research inputs/outputs to be included. There are 48 studies included, 44% (21) are US based, 46% are European (22), 4% are Canadian (2) and 4% from Australia/New Zealand (2) and 2% from South America (1). Sample size ranged from 20 to 130,000 with a median of 828 (standard error = 3,534). 69% (33) of the studies employed bibliometrics as a research output, although several studies employed alternative measures, like project initiation and completion, patents and technology licensing, post-project peer review, levels of international collaboration, future funding success, securing tenure track positions, and career satisfaction. Collectively, 52% (25) of the studies used non-bibliometric data but only 25% (12) of all projects used more than one measure of research output. Of the studies that rely on only one indicator (36), 64% (23) rely on bibliometric measures.

Table 1

PaperFunding sourceNScoring (S) or Funding decision (FD)Impact indicatorUnfunded controlImpact time period (years)Summary of results
Armstrong et al., 1997Heart and Stroke Foundation of Canada192FDNumber of publications and Citations ex-postY3–12 years ex-postFunded applicants cited more than unfunded
Berg, 2011NIH/NIGMS789SCitations and Publications ex-postN5 years ex-postModerate correlation between scores and citations/publications
Bloch et al., 2014Danish Agency for Science, Technology and Innovation3,027S/FDSecuring faculty positions ex postY3 years ex postFunded applicants acquire more funding and faculty positions compared to similarly scoring unfunded applicants
Bol et al., 2018Innovation Research Incentives Scheme (Netherlands)1255S/FDEx post grant fundingY7–12 years ex postFunded applicants acquire more funding over time compared to similarly scoring unfunded applicants
Bornmann and Daniel, 2006Boehringer Ingelheim Fonds Post-doctoral Fellowship397FDCitations ex-anteY1–9 years ex-anteFunded applicants have more citations than rejected applicants ex-ante
Bornmann et al., 2008bEuropean Molecular Biology Organization Post-doc and young investigator funds965FDCitations (ex-ante and ex-post)Y5 years ex ante and 8 years ex postFunded applicants higher citations than unfunded both ex-ante and ex-post (although there is significant type I/II error)
Bornmann et al., 2010European Molecular Biology Organization668S/FDNormalized citations, publications ex ante and ex postY3 years ex ante and 3 years ex postAwarded applicants more cited than rejected applicants, but best unfunded perform better as well as funded
Bromham et al., 2016Australian Research Council Discovery Programme18,476S/FDInterdisciplinarity and collaboration ex anteYIn situ measurement of applicationPoorer Scores are associated with higher interdisciplinarity, even when factoring in level of collaboration is taken into account
Cabezas-Clavijo et al., 2013Spanish National R&D Plan2,333S/FDCitations and number of Publications ex-anteY5 years ex anteAccepted proposals better ex-ante PI performance than rejected PIs, but low level of correlation between scores and ex-ante productivity
Campbell et al., 2010National Cancer Institute of Canada685S/FDCitation, Publications ex postY3 years ex postHigher citations for funded vs. unfunded
Chai and Shih, 2016Danish National Advanced Technology Foundation4,224FDPatent and publication productivity ex postY5 years ex postPublication and patent productivity higher for funded for younger firms and larger projects
Claveria et al., 2000Spanish Health Research Fund2,744SEx-post peer review of research outcomesN4–12 years ex-postEx-ante review scores significant predictor of ex-post review scores
Danthi et al., 2014NIH/NHLBI1,492SNormalized Citation ex-postN2 years ex-postNo Association between scores and citation impact/$
Decullier et al., 2014French Ministry of Health481FDClinical project initiation, completion, publication ex postY8–10 years ex postFunded projects were more likely to be initiated but once initiated, completion and publication were unaffected by funding status
Doyle et al., 2015NIH/NIMH1,755SNormalized Citation ex-postN6–15 years ex-postNo Association between scores and citation impact/$
El-Sawi et al., 2009Association of American Medical College (Medical Education Research)20FDSurvey on number of research products and level of collaboration ex postY3–8 years ex post (products and collaborations)Funded applicants had more research products and higher levels of collaboration than unfunded applicants
Escobar-Alvarez and Myers, 2013Doris Duke Charitable Trust1,441S/FDNIH funding success ex-postY1–13 years ex-postFunded applicants higher ex post NIH funding success than unfunded yet highly ranked applicants (ex-ante)
Fang and Meyer, 2003Howard Hughes Medical Institute867FDNIH funding success ex-postY5–13 years ex-postFunded applicants had higher NIH funding success than unfunded (similar ex-ante qualifications)
Fang et al., 2016NIH102,740SCitations and patents ex postN5 years ex postRe-analysis of Li and Agha, 2015, amongst the higher scoring, poor correlation between scores and productivity
Galbraith et al., 2006US Department of Defense69SSuccess of early stage technologies (CRADA, license or SBIR funding) ex-postN2–3 years ex postReviewer scores only weakly predictive of commercial success of funded projects
Gallo et al., 2014Anonymous (US)227SNormalized Citation ex-postN8–15 years ex-postModerate correlation between scores and citation impact
Gush et al., 2017New Zealand Marsden Fund1,263S/FDNormalized citations ex-postY5 years ex-postFunding success associated with research output increases, but no correlation with review scores (all high performers)
Heggeness et al., 2018NIH14,276S/FDEx post NIH grant fundingY7–19 years ex postFunded applicants acquire more funding over time compared to similarly scoring unfunded applicants
Hornbostel et al., 2009German Research Foundation695FDFaculty positions, Career Satisfaction, Citation and publication levels ex postY4 years ex ante and ex post (publications), 3 years ex post (Career)Career satisfaction and faculty positions gained are higher among funded, but only marginally better citation levels than unfunded
Jacob and Lefgren, 2011a[Postdoc]NIH Fellowships12,189S/FDNumber of Publications ex-post and active research careerY5 years ex post (publications and career)Funded have slightly elevated publication rate over highly ranked but unfunded ex-post, also more active research careers, some correlation with review score and productivity
Jacob and Lefgren, 2011b [Grant]NIH R01 Projects54,741S/FDNumber of Publications ex-postY5 years ex postFunded applicants have slight increase in productivity ex-post compared to highly ranked but unfunded, some correlation with review score and productivity
Langfeldt et al., 2012Research Council of Norway6,064FDNormalized citations ex-ante and ex-post, Survey, international co-authorshipY3–5 years ex post and 4–6 years ex ante (publications and co-authorship), 5–7 years (career)Successful applicants more cited than unsuccessful ex-ante and ex-post, successful have higher degree of international co-authorship than unsuccessful
Lauer et al., 2015NIH/NHLBI6,873STop 10% Normalized Citation ex-postN4–35 years ex-postModest relationship between scores and citation impact (no relationship with citation impact/$)
Li and Agha, 2015NIH130,000SCitations and Patents ex-postN5 years ex-postModerate correlation between scores and citations and patents
Mahoney et al., 2007American Academy of Family Physicians Foundation95FDNumber of publications ex-post and ex anteY5 years ex post and ex anteFunded higher publication rate compared to unfunded for both ex post/ex ante
Mason et al., 2013NIH2,893S/FDNIH funding success ex-postY3–31 years ex-postFunded applicants had higher NIH funding success than highly ranked but unfunded applicants
Mavis and Katz, 2003March of Dimes Birth Defects Foundation439FDNumber of Publications, Citations and additional funding success ex-postY10 years ex-postFunded applicants published more and received more citations than unfunded, as well as garnered more additional funding
Melin and Danell, 2006Swedish Foundation for Strategic Research40FDNumber of publications and impact factor (ex-ante and ex-post), patents, international collaboration, acquiring additional fundingY3 years ex ante and 3 years ex post (5 years ex post for patents and collaborations)Similar number of publications ex-ante and ex-post for funded/unfunded, but funded have higher average impact factor ex-post, more international collaborations ex-ante and ex-post and more success in future funding, as well as more patents
Mutz et al., 2015Austrian Science Fund1,689SEx-post peer review of research outcomesN (Data Imputation)5–15 years ex postModerate correlation between ex-ante peer review evaluations and ex-post reviews of performance (may not represent independent observation)
Neufeld et al., 2013European Research Council758FDEx-ante Normalized CitationsY6 years ex anteFunded and unfunded applicants have similar pre-application productivity and citation impact (all high performers)
Pion and Cordray, 2008Burroughs Wellcome Fund619S/FDSecuring faculty positions, R01 grant success, and publishing in top-ranked journals ex-postY4–7 years (career/grant), 1–5 years for publications ex-postFunded outperformed unfunded applicants in faculty position, R01 grant success, and publications, but differences diminished when controlled for review score
Pion and Ionescu-Pioggia, 2003Burroughs Wellcome Fund101FDSecuring tenure track positions ex-postN1–4 years ex-postFunded applicants secured more tenure track positions than unfunded
Reinhart (2009)Swiss National Science Foundation63FDCitation levels ex postY7 years ex postFunded applicants more cited than unfunded
Robitaille et al., 2015European Research Council5,100S/FDNumber of publications, normalized citations ex-ante and ex post, interdisciplinarityY27–31 years ex ante, 2–6 years ex post (publications and collaborations)Funded had higher citations than unfunded overall and highly ranked but unfunded. Funding did not affect interdisciplinarity of output.
Sandstrom and Hallsten, 2008Swedish Research Council280SPublications and Citations ex anteN6 years ex anteReview scores correlated with publication productivity
Saygitov, 2014Russian Foundation for Basic Research190FDCitations and publications ex ante and ex postY5 years ex ante and 5 years ex postFunded and unfunded applicants both have similar productivity
Scheiner and Bouchie, 2013NSF48SCitation levels ex-postN11 years ex-postNo correlation between scores and citation levels
Tesauro et al., 2013NIH/NCI184S/FDEx post grant fundingY2–10 years ex postFunded more likely to acquire ex post funding than similarly scoring unfunded applicants
Ubfal and Maffioli, 2011Fund for the Scientific and Technological Research (Argentina)496FDCollaboration ex postY12–13 years ex postFunded applicants had more collaboration ex post compared to unfunded
van den Besselaar and Leydesdorff, 2009Netherlands Research Council (Social Sciences)1,178S/FDNumber of Publications and Citations ex-anteY4–6 years ex anteFunded more cited than unfunded; however top tier unfunded cited more than funded
Van den Besselaar and Sandstrom, 2015Netherlands Social Science Council260S/FDCitations/Top 10% publications ex-post, securing faculty positionsY8–10 years ex-postFunded applicants higher performance than overall unfunded, but no difference with best performing unfunded. Funded applicants secured position more than unfunded
van Leeuwen and Moed, 2012Netherlands Organization for Scientific Research3,660FDNormalized Citations ex anteY4–8 years ex anteFunded applicants higher citations than unfunded
Wenneras and Wold, 1997Swedish Medical Research Council114SPublications and Citations ex anteNUnclear (probably 5 year or less ex ante)Applicant competence scores correlated with publication impact

Summary of literature.

Twenty-nine percent (14) are conducted without an unfunded control, and all but one of this group examines review scores and output of funded projects. Of this subset, 71% (10) provided evidence for some level of predictive validity of review decisions. Of the 29% (4) that did not, two studies used citation level per dollar spent (Danthi et al., 2014; Doyle et al., 2015) which can mask correlations, one only looked at a limited range of peer review scores, ignoring poorer scoring projects (Fang et al., 2016) and one study had a very small sample size of 40 (Scheiner and Bouchie, 2013). 71% (34) of studies listed have unfunded controls and of those, 91% (31) showed some level of predictive validity of review decisions. It has been previously suggested that another important variable in testing validity is the time window when impact is measured, especially for bibliometric impact (Van den Besselaar and Sandstrom, 2015). We find for bibliometric studies that, while most have a range, the median maximum time at which impact is measured is 5.0 ± 1.0 years after the review decision, and that 17% (3) showed no predictive validity for 5 years or less vs. 20% for more than 5 years.

It should be noted that many of the differences in impact observed were small, especially with regard to bibliometric measures. Also, several studies indicated that, when the poorer scoring unfunded applicants or poorer scoring projects were excluded from analysis, the validity disappears, although this depended on the metric used (Fang and Meyer, 2003; Hornbostel et al., 2009; Escobar-Alvarez and Myers, 2013). Also, several have noted the large degree of variability in bibliometric measures, especially with regard to projects, which obfuscate strong correlations or firm conclusions. In addition, interpretation of results was sometimes made difficult due to the potential effect of the funding itself. Nevertheless, overall these results suggest at least a coarse discriminatory power, able to separate poor proposals from better ones, but not necessarily good from great. While these results should give us pause in the current era of low funding success rates, they also suggest that more needs to be done to include a variety of external impact measures for validation studies, as well as in parallel, focusing on the internal validity of review decision making procedures.

Conclusions

It is clear that despite the importance of the peer review process in determining billions of research dollars funded in the US, there are still only a handful of studies conducted with this focus (most of which were published in the last 7 years) and less than half are US based. More research needs to be done to understand the scientific validity of this process, which means improved access to pre-funding peer review data. Academics should work with funding agencies (both federal and private funders) to negotiate agreements to gain access to this data. Funding agencies should invest in these studies.

Second, it is clear that there are many ways to identify success, and the scientometrics community has warned that multiple indicators and a well-rounded approach should be used to assess the value of research (Hicks et al., 2015). Yet, the majority of these studies here use only one type of indicator, and of those, bibliometric measures are the most used. Many issues surround the use of bibliometric measures as an accurate indicator of impact, as they can depend on many other factors unrelated to research quality (Sarli et al., 2010). More work into indicators that take into account social impact and non-bibliometric methods are also needed (Bornmann, 2013). For instance, as some have pointed out that traditional citation analysis may underestimate the true impact of clinical research (Van Eck et al., 2013); prioritizing citation counts from clinical trials or clinical guidelines may be one way to highlight translational impact (Thelwall and Maflahi, 2016). Similarly, while methodological innovations are usually well cited, getting some sense of rate of usage in a field (e.g., through the use of a survey) may give a more appropriate estimation of impact beyond what is published (Brueton et al., 2014). And as the importance of reproducibility in science cannot be overstated (Ioannidis, 2005), assessments of reproducibility (e.g., the r-factor) are currently in development (Chawla, 2018). As impact indicators are generated and validated, they should be used in review validation studies.

Third, these future studies should use a combination of metrics in order to produce a more comprehensive analysis, context and validity. Only 25% of these studies used more than one impact indicator. However, some that did found peer review decisions to be predictive of success by one measure, but much less predictive by another (Melin and Danell, 2006; Hornbostel et al., 2009). Studies show huge variability in bibliometric indicators, so they need to be supplemented to give robustness to the test for validity (Danthi et al., 2014; Gallo et al., 2014). Also, different research programs have different goals which may include both bibliometric and non-bibliometric outcomes, both should be observed to give context. Similarly, program specific context should be considered. For example, research programs can evolve over time in terms of quality of applications received and funding success rates (Gallo et al., 2014). Also, one must also consider how scientific excellence is defined and measured and how the incentivization through metrics can influence research output and the review itself (Lindner and Nakamura, 2015; Bornmann, 2017; Moore et al., 2017; Ferretti et al., 2018). Subjective definitions of excellence may not always equate to high innovation or impact, and thus the context of how the review was conducted and how reviewers were instructed to interpret excellence should be considered (Luukkonen, 2012). Once a panel of indicators is decided upon, the results should be integrated and interpreted in the context of the area of science, the goals of the research program, and the implementation of the peer review. In addition, the overall societal impact needs to be considered, as well as the inherent volatility of the scientific discovery process.

Fourth, the structure of the tests of validity vary considerably across studies, some of which lack crucial controls. For instance, examining ex-post applicant performance without comparing ex-ante performance may fail to remove the effect of funding itself. Also, for studies looking at ex ante performance as a predictor of future performance, they should take into account the Matthew effect in their interpretation, as some results show that funding less-awarded groups may actually have higher impact than more distinguished groups (Langfeldt et al., 2015; Mongeon et al., 2016), and thus reviewers choosing high ex ante performers may not always pay off. For studies examining scores vs. applicant or project output, they are usually missing crucial information about the unfunded group, which limits the ability to test validity (Lindner and Nakamura, 2015). In addition, many studies have indicated low inter-rater reliability amongst panelists (Cole and Simon, 1981) and some studies indicate that review scores and rankings are much more dependent on the individual reviewer than on the proposal (Jayasinghe et al., 2003; Pier et al., 2018). Thus, there is a need to look at the internal validity of the review process with examinations of potential reviewer bias, review structures and baselines of decision making (Magua et al., 2017). These types of internal tests of review process validity are not included in this manuscript, but are crucial for assessing other expectations of peer review (Wood and Wessely, 2003), like fairness (Lee et al., 2013), efficiency (Carpenter et al., 2015) and rationality (Gallo et al., 2016).

Finally, from the results summarized in this review, it seems that peer review likely does have some coarse discrimination in determining the level and quality of output from research funding, suggesting the system does have some level of validity, although admittedly the span of funding agencies and mechanisms included in this review complicates generalization somewhat. While it may be able to separate good and flawed proposals, discrimination amongst the top tier proposals or applicants may be more difficult, which is what the system is currently charged to do given recent funding levels (Fang et al., 2016). Nevertheless, this seems to depend on the metric used, as some studies found a high degree of discrimination when tracking career success of funded and top tier unfunded applicants (Fang and Meyer, 2003; Hornbostel et al., 2009; Escobar-Alvarez and Myers, 2013), although the effects of funding itself have to be teased out (Bol et al., 2018). Also, some level of validity was found with studies involving patents, post-funding review of outcomes and levels of collaboration as well, suggesting validity across multiple outputs. Nevertheless, as the decisions become more subjective, the likelihood for bias increases, and thus much effort must be focused on ensuring the fidelity and equity of the review process. It is likely unavoidable that some meritorious research will not be funded, putting more pressure on research funding administrators to incorporate into the final funding decisions considerations of portfolio diversification, programmatic concerns, promotion of collaborations and risk considerations (Galis et al., 2012; Janssens et al., 2017; Peifer, 2017; Wahls, 2018). These considerations, as well as the creation of new funding mechanisms (e.g., funds for early career investigators; Kaiser, 2017) should complement research into peer review processes. Given that some aspects of scientific discovery may be “fundamentally unpredictable,” the development of science policies that “cultivate and maintain a healthy ecosystem of scientists rather than focus on predicting individual discoveries” may be the ideal to strive for (Clauset et al., 2017).

Statements

Author contributions

SAG and SRG contributed to the conception of the review. SAG performed the statistical analysis and did the initial gathering of the literature. SAG wrote the first draft of the manuscript. SAG and SRG wrote sections of the manuscript. All authors contributed to manuscript revision, read and approved the submitted version.

Acknowledgments

Thanks to the American Institute of Biological Sciences (AIBS) Scientific Peer Advisory and Review Services (SPARS) staff.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  • 1

    AdamsJ. D.BlackG. C.ClemmonsJ. R.StephanP. E. (2005). Scientific teams and institutional collaborations: evidence from US universities, 1981–1999. Res. Policy34, 259285. 10.1016/j.respol.2005.01.014

  • 2

    ArmstrongP. W.CaversonM. M.AdamsL.TaylorM.OlleyP. M. (1997). Evaluation of the heart and stroke foundation of Canada research scholarship program: research productivity and impact. Can. J. Cardiol.13, 507516.

  • 3

    AzoulayP.StuartT.WangY. (2013). Matthew: effect or fable?Manage. Sci.60, 92109. 10.1287/mnsc.2013.1755

  • 4

    BanT. A. (2006). The role of serendipity in drug discovery. Dial. Clin. Neurosci.8. 335344.

  • 5

    BeaudryC.KananianR. (2013). Follow the (industry) money–The Impact of science networks and industry-to-university contracts on academic patenting in nanotechnology and biotechnology. Indus. Innov. 20, 241260. 10.1080/13662716.2013.791125

  • 6

    BergJ. M. (2011). Productivity Metrics and Peer Review Scores, Continued. NIGMS Feedback Loop (blog). Available online at: https://loop.nigms.nih.gov/2011/06/productivity-metrics-and-peer-review-scores/ (Accessed July 2018).

  • 7

    BlochC.GraversenE. K.PedersenH. S. (2014). Competitive research grants and their impact on career performance. Minerva52, 7796. 10.1007/s11024-014-9247-0

  • 8

    BolT.de VaanM.van de RijtA. (2018). The Matthew effect in science funding. Proc. Natl. Acad. Sci. U.S.A.115, 48874890. 10.1073/pnas.1719557115

  • 9

    BornmannL. (2013). What is societal impact of research and how can it be assessed? A literature survey. J. Assoc. Inform. Sci. Technol.64, 217233. 10.1002/asi.22803

  • 10

    BornmannL. (2017). Measuring impact in research evaluations: a thorough discussion of methods for, effects of and problems with impact measurements. High. Educ.73, 775787. 10.1007/s10734-016-9995-x

  • 11

    BornmannL.DanielH. D. (2006). Selecting scientific excellence through committee peer review-A citation analysis of publications previously published to approval or rejection of post-doctoral research fellowship applicants. Scientometrics68, 427440. 10.1007/s11192-006-0121-1

  • 12

    BornmannL.LeydesdorffL.Van den BesselaarP. (2010). A meta-evaluation of scientific research proposals: different ways of comparing rejected to awarded applications. J. Informetr.4, 211220. 10.1016/j.joi.2009.10.004

  • 13

    BornmannL.MutzR.DanielH. D. (2008c). Are there better indices for evaluation purposes than the h index? A comparison of nine different variants of the h index using data from biomedicine. J. Assoc. Inform. Sci. Technol.59, 830837. 10.1002/asi.20806

  • 14

    BornmannL.MutzR.NeuhausC.DanielH. D. (2008a). Citation counts for research evaluation: standards of good practice for analyzing bibliometric data and presenting and interpreting results. Ethics Sci. Environ. Polit.8, 93102. 10.3354/esep00084

  • 15

    BornmannL.WallonG.LedinA. (2008b). Does the committee peer review select the best applicants for funding? An investigation of the selection process for two european molecular biology organization programmes. PLoS ONE3:e3480. 10.1371/journal.pone.0003480

  • 16

    BoudreauK. J.GuinanE. C.LakhaniK. R.RiedlC. (2016). Looking across and looking beyond the knowledge frontier: intellectual distance, novelty, and resource allocation in science. Manage. Sci.62, 27652783. 10.1287/mnsc.2015.2285

  • 17

    BoyackK. W.SmithC.KlavansR. (2018). Toward predicting research proposal success. Scientometrics114, 449461. 10.1007/s11192-017-2609-2

  • 18

    BromhamL.DinnageR.HuaX. (2016). Interdisciplinary research has consistently lower funding success. Nature534:684. 10.1038/nature18315

  • 19

    BruetonV. C.ValeC. L.Choodari-OskooeiB.JinksR.TierneyJ. F. (2014). Measuring the impact of methodological research: a framework and methods to identify evidence of impact. Trials15:464. 10.1186/1745-6215-15-464

  • 20

    Cabezas-ClavijoA.Robinson-GarcíaN.EscabiasM.Jiménez-ContrerasE. (2013). Reviewers' ratings and bibliometric indicators: hand in hand when assessing over research proposals?PLoS ONE8:e68258. 10.1371/journal.pone.0068258

  • 21

    CallahamM.WearsR. L.WeberE. (2002). Journal prestige, publication bias, and other characteristics associated with citation of published studies in peer-reviewed journals. JAMA287, 28472850. 10.1001/jama.287.21.2847

  • 22

    CampbellD.Picard-AitkenM.CoteG.CarusoJ.ValentimR.EdmondsS.et al. (2010). Bibliometrics as a performance measurement tool for research evaluation: the case of research funded by the National Cancer Institute of Canada. Am. J. Eval.31, 6683. 10.1177/1098214009354774

  • 23

    CarayolN.ThiT. U. N. (2005). Why do academic scientists engage in interdisciplinary research?Res. Eval.14, 7079. 10.3152/147154405781776355

  • 24

    CarpenterA. S.SullivanJ. H.DeshmukhA.GlissonS. R.GalloS. A. (2015). A retrospective analysis of the effect of discussion in teleconference and face-to-face scientific peer-review panels. BMJ Open5:e009138. 10.1136/bmjopen-2015-009138

  • 25

    ChaiS.ShihW. (2016). Bridging science and technology through academic–industry partnerships. Res. Policy45, 148158. 10.1016/j.respol.2015.07.007

  • 26

    ChawlaD. S. (2018). Online Tool Calculates Reproducibility Scores of PubMed papers. Science. Available online at: http://www.sciencemag.org/news/2018/01/online-tool-calculates-reproducibility-scores-pubmed-papers last accessed 4/07/18 (Accessed 22 January, 2018)

  • 27

    ChenC. (2016). Grand challenges in measuring and characterizing scholarly impact. Front. Res. Metr. Anal.1:4. 10.3389/frma.2016.00004

  • 28

    ClausetA.LarremoreD. B.SinatraR. (2017). Data-driven predictions in the science of science. Science355, 477480. 10.1126/science.aal4217

  • 29

    ClaveriaL. E.GuallarE.CamiJ.CondeJ.PastorR.RicoyJ. R.et al. (2000). Does peer review predict the performance of research projects in health sciences?Scientometrics47, 1123. 10.1023/A:1005609624130

  • 30

    ColeS.SimonG. A. (1981). Chance and consensus in peer review. Science214, 881886. 10.1126/science.7302566

  • 31

    DanthiN.WuC. O.ShiP.LauerM. (2014). Percentile ranking and citation impact of a large cohort of National Heart, Lung, and Blood Institute–funded cardiovascular R01 grants. Circ. Res.114, 600606. 10.1161/CIRCRESAHA.114.302656

  • 32

    De JongS.BarkerK.CoxD.SveinsdottirT.Van den BesselaarP. (2014). Understanding societal impact through productive interactions: ICT research as a case. Res. Eval.23, 89102. 10.1093/reseval/rvu001

  • 33

    DecullierE.HuotL.ChapuisF. R. (2014). Fate of protocols submitted to a French national funding scheme: a cohort study. PLoS ONE9:e99561. 10.1371/journal.pone.0099561

  • 34

    DidegahF.BowmanT. D.HolmbergK. (2017). On the Differences Between Citations and Altmetrics: An Investigation of Factors Driving altmetrics vs. Citations for Finnish articles. arXiv preprint arXiv:1710.08594.

  • 35

    DinsmoreA.AllenL.DolbyK. (2014). Alternative perspectives on impact: the potential of ALMs and altmetrics to inform funders about research impact. PLoS Biol.12:e1002003. 10.1371/journal.pbio.1002003

  • 36

    Dolgin (2018) PubMed Commons Closes its Doors to Comments. Nature Feb 02 2018. Available online at: https://www.nature.com/articles/d41586-018-01591-4 (Accessed June 4, 2018).

  • 37

    DonovanC. (2007). The qualitative future of research evaluation. Sci. Public Policy34, 58559710.3152/030234207X256538

  • 38

    DonovanC. (2011). State of the art in assessing research impact: introduction to a special issue. Res. Eval.20, 175179. 10.3152/095820211X13118583635918

  • 39

    DoyleJ. M.QuinnK.BodensteinY. A.WuC. O.DanthiN.LauerM. S. (2015). Association of percentile ranking with citation impact and productivity in a large cohort of de novo NIMH-funded R01 grants. Mol. Psychiatry20:1030. 10.1038/mp.2015.71

  • 40

    DwanK.AltmanD. G.ArnaizJ. A.BloomJ.ChanA. W.CroninE.et al. (2008). Systematic review of the empirical evidence of study publication bias and outcome reporting bias. PLoS ONE3:e3081. 10.1371/journal.pone.0003081

  • 41

    EbadiA.SchiffauerovaA. (2015). How to receive more funding for your research? Get connected to the right people!. PLoS ONE10:e0133061. 10.1371/journal.pone.0133061

  • 42

    Editorial (2018) The serendipity test. Nature554:5. 10.1038/d41586-018-01405-7

  • 43

    El-SawiN. I.SharpG. F.GruppenL. D. (2009). A small grants program improves medical education research productivity. Acad. Med.84, S105S108. 10.1097/ACM.0b013e3181b3707d

  • 44

    Escobar-AlvarezS. N.MyersE. R. (2013). The Doris Duke clinical scientist development award: implications for early-career physician scientists. Acad. Med.88, 17401746. 10.1097/ACM.0b013e3182a7a38e

  • 45

    FangD.MeyerR. E. (2003). Effect of two Howard Hughes Medical Institute research training programs for medical students on the likelihood of pursuing research careers. Acad. Med.78, 12711280. 10.1097/00001888-200312000-00017

  • 46

    FangF. C.BowenA.CasadevallA. (2016). NIH peer review percentile scores are poorly predictive of grant productivity. Elife5:e13323. 10.7554/eLife.13323

  • 47

    FecherB.FriesikeS.HebingM. (2015). What drives academic data sharing?PLoS ONE10:e0118053. 10.1371/journal.pone.0118053

  • 48

    FerrettiF.PereiraÂ. G.VértesyD.HardemanS. (2018). Research excellence indicators: time to reimagine the ‘making of'? Sci. Public Policy111. 10.1093/scipol/scy007

  • 49

    FiresteinS. (2015). Funding Failure, in Failure: Why Science is so Successful (New York, NY: Oxford University Press), 177204.

  • 50

    FortinJ. M.CurrieD. J. (2013). Big science vs. little science: how scientific impact scales with funding. PLoS ONE8:e65263. 10.1371/journal.pone.0065263

  • 51

    GalbraithC. S.EhrlichS. B.DeNobleA. F. (2006). Predicting technology success: identifying key predictors and assessing expert evaluation for advanced technologies. J. Technol. Transf.31, 673684. 10.1007/s10961-006-0022-8

  • 52

    GalisZ. S.HootsW. K.KileyJ. P.LauerM. S. (2012). On the value of portfolio diversity in heart, lung, and blood research. Am. J. Respir. Crit. Care Med.186:575. 10.1164/rccm.201208-1437ED

  • 53

    GalloS. A.CarpenterA. S.IrwinD.McPartlandC. D.TravisJ.ReyndersS.et al. (2014). The validation of peer review through research impact measures and the implications for funding strategies. PLoS ONE9:e106474. 10.1371/journal.pone.0106474

  • 54

    GalloS. A.SullivanJ. H.GlissonS. R. (2016). The influence of peer reviewer expertise on the evaluation of research funding applications. PLoS ONE11:e0165147. 10.1371/journal.pone.0165147

  • 55

    GintherD. K.SchafferW. T.SchnellJ.MasimoreB.LiuF.HaakL. L.et al. (2011). Race, ethnicity, and NIH research awards. Science333, 10151019. 10.1126/science.1196783

  • 56

    GokA.RigbyJ.ShapiraP. (2016). The impact of research funding on scientific outputs: evidence from six smaller European countries. J. Assoc. Inform. Sci. Technol.67, 715730. 10.1002/asi.23406

  • 57

    GuerzoniM.AldridgeT. T.AudretschD. B.DesaiS. (2014). A new industry creation and originality: insight from the funding sources of university patents. Res. Policy43, 16971706. 10.1016/j.respol.2014.07.009

  • 58

    GurneyT.HorlingsE.Van den BesselaarP.SumikuraK.SchoenA.LaurensP.et al. (2014). Analysing knowledge capture mechanisms: methods and a stylised bioventure case. J. Informetr.8, 259272. 10.1016/j.joi.2013.12.007

  • 59

    GushJ.JaffeA.LarsenV.LawsA. (2017). The effect of public funding on research output: the New Zealand Marsden Fund. NZ Econ. Papers122. 10.1080/00779954.2017.1325921

  • 60

    HagenN. T. (2008). Harmonic allocation of authorship credit: source-level correction of bibliometric bias assures accurate publication and citation analysis. PLoS ONE3:e4021. 10.1371/journal.pone.0004021

  • 61

    HeggenessM. L.GintherD. K.LarenasM. I.Carter-JohnsonF. D. (2018). The Impact of Postdoctoral Fellowships on a Future Independent Career in Federally Funded Biomedical Research (No. w24508).National Bureau of Economic Research.

  • 62

    HicksD.WoutersP.WaltmanL.De RijckeS.RafolsI. (2015). The Leiden Manifesto for research metrics. Nature520:429. 10.1038/520429a

  • 63

    HirschJ. E. (2005). An index to quantify an individual's scientific research output. Proc. Natl. Acad. Sci. U.S.A.102:16569. 10.1073/pnas.0507655102

  • 64

    HornbostelS.BohmerS.KlingspornB.NeufeldJ.von InsM. (2009). Funding of young scientist and scientific excellence. Scientometrics79, 171190. 10.1007/s11192-009-0411-5

  • 65

    HuangZ.ChenH.LiX.RocoM. C. (2006). Connecting NSF funding to patent innovation in nanotechnology (2001–2004). J. Nanopart. Res.8, 859879. 10.1007/s11051-006-9147-9

  • 66

    HutchinsB. I.YuanX.AndersonJ. M.SantangeloG. M. (2016). Relative Citation Ratio (RCR): a new metric that uses citation rates to measure influence at the article level. PLoS Biol.14:e1002541. 10.1371/journal.pbio.1002541

  • 67

    IoannidisJ. P. (2005). Why most published research findings are false. PLoS Med.2:e124. 10.1371/journal.pmed.0020124

  • 68

    IoannidisJ. P. (2008). Measuring co-authorship and networking-adjusted scientific impact. PLoS ONE3:e2778. 10.1371/journal.pone.0002778

  • 69

    JacobB. A.LefgrenL. (2011a). The impact of research grant funding on scientific productivity. J. Public Econ.95, 11681177. 10.1016/j.jpubeco.2011.05.005

  • 70

    JacobB. A.LefgrenL. (2011b). The impact of NIH postdoctoral training grants on scientific productivity. Res. Policy40, 864874. 10.1016/j.respol.2011.04.003

  • 71

    JanssensA. C. J.MillerG. W.NarayanK. V. (2017). The data and analysis underlying NIH's decision to cap research support lacked rigor and transparency: a commentary. PeerJ Preprints.5:e3106v1. 10.7287/peerj.preprints.3106v1

  • 72

    JayasingheU. W.MarshH. W.BondN. (2003). A multilevel cross-classified modelling approach to peer review of grant proposals: the effects of assessor and researcher attributes on assessor ratings. J. R. Stat. Soc. A.166, 279300. 10.1111/1467-985X.00278

  • 73

    KaatzA.LeeY. G.PotvienA.MaguaW.FilutA.BhattacharyaA.et al. (2016). Analysis of NIH R01 Application Critiques, Impact and Criteria Scores: Does the Sex of the Principal Investigator Make a Difference?Acad. Med. J. Assoc. Am. Med. Coll.91:1080. 10.1097/ACM.0000000000001272

  • 74

    KaiserJ. (2017). Updated: NIH Abandons Controversial Plan to Cap Grants to Big Labs, Creates New Fund for Younger Scientists Science June 8 2017. Available online at: http://www.sciencemag.org/news/2017/06/updated-nih-abandons-controversial-plan-cap-grants-big-labs-creates-new-fund-younger (Accessed July 4, 2018).

  • 75

    KaltmanJ. R.EvansF. J.DanthiN. S.WuC. O.DiMicheleD. M.LauerM. S. (2014). Prior publication productivity, grant percentile ranking, and topic-normalized citation impact of NHLBI cardiovascular R01 grants. Circ. Res.115, 617624. 10.1161/CIRCRESAHA.115.304766

  • 76

    KeserciS.LivingstonE.WanL.PicoA. R.ChackoG. (2017). Research synergy and drug development: bright stars in neighboring constellations. Heliyon3:e0044210.1016/j.heliyon.2017.e00442

  • 77

    KnoepflerP. (2015). Reviewing post-publication peer review. Trends Genet.31, 221223. 10.1016/j.tig.2015.03.006

  • 78

    LangfeldtL.BennerM.SivertsenG.KristiansenE. H.AksnesD. W.BorlaugS. B.et al. (2015). Excellence and growth dynamics: a comparative study of the Matthew effect. Sci. Public Policy42, 661675. 10.1093/scipol/scu083

  • 79

    LangfeldtL.RambergI.SivertsenG.BlochC.OlsenD. S. (2012). Evaluation of the Norwegian Scheme for Independent Research Projects (FRIPRO) Available online at: http://www.technopolis-group.com/wp-content/uploads/2014/04/1545_RCN_Background_Report_No07_Users_Experience.pdf (Accessed July, 2018).

  • 80

    LauerM. (2015). Perspectives on Peer Review at the NIH. Available online at: https://nexus.od.nih.gov/all/2015/11/12/perspectives-on-peer-review-at-the-nih/ (Accessed May 4, 2018).

  • 81

    LauerM. S.DanthiN. S.KaltmanJ.WuC. (2015). Predicting productivity returns on investment: thirty years of peer review, grant funding, and publication of highly cited papers at the National Heart, Lung, and Blood Institute. Circ. Res.117, 239243. 10.1161/CIRCRESAHA.115.306830

  • 82

    LeeC. J. (2015). Commensuration bias in peer review. Philos. Sci.82, 12721283. 10.1086/683652

  • 83

    LeeC. J.SugimotoC. R.ZhangG.CroninB. (2013). Bias in peer review. J. Assoc. Inform. Sci. Technol.64, 217. 10.1002/asi.22784

  • 84

    LeydesdorffL.BornmannL.CominsJ. A.MilojevicS. (2016). Citations: indicators of quality? The impact fallacy. Front. Res. Metr. Analyt.1:1. 10.3389/frma.2016.00001

  • 85

    LiD.AghaL. (2015). Big names or big ideas: do peer-review panels select the best science proposals?Science348, 434438. 10.1126/science.aaa0185

  • 86

    LiD.AzoulayP.SampatB. N. (2017). The applied value of public investments in biomedical research. Science356, 7881. 10.1126/science.aal0010

  • 87

    LindnerM. D.NakamuraR. K. (2015). Examining the predictive validity of NIH peer review scores. PLoS ONE10:e0126938. 10.1371/journal.pone.0126938

  • 88

    LukeD. A.SarliC. C.SuiterA. M.CarothersB. J.CombsT. B.AllenJ. L.et al. (2018). The translational science benefits model: a new framework for assessing the health and societal benefits of clinical and translational sciences. Clin. Transl. Sci.11, 7784. 10.1111/cts.12495

  • 89

    LuukkonenT. (2012). Conservatism and risk-taking in peer review: emerging ERC practices. Res. Eval.21, 4860. 10.1093/reseval/rvs001

  • 90

    MaguaW.ZhuX.BhattacharyaA.FilutA.PotvienA.LeatherberryR.et al. (2017). Are female applicants disadvantaged in National Institutes of Health peer review? Combining algorithmic text mining and qualitative methods to detect evaluative differences in R01 reviewers' critiques. J. Wom. Health26, 560570. 10.1089/jwh.2016.6021

  • 91

    MahoneyM. C.VermaP.MorantzS. (2007). Research productivity among recipients of AAFP foundation grants. Anna. Fam. Med.5, 143145. 10.1370/afm.628

  • 92

    MarshH. W.JayasingheU. W.BondN. W. (2008). Improving the peer-review process for grant applications: reliability, validity, bias, and generalizability. Am. Psychol.63:160. 10.1037/0003-066X.63.3.160

  • 93

    MasonJ. L.LeiM.Faupel-BadgerJ. M.GinsburgE. P.SegerY. R.DiJosephL.et al. (2013). Outcome evaluation of the National Cancer Institute career development awards program. J. Cancer Educ.28, 917. 10.1007/s13187-012-0444-y

  • 94

    MavisB.KatzM. (2003). Evaluation of a program supporting scholarly productivity for new investigators. Acad. Med.78, 757765. 10.1097/00001888-200307000-00020

  • 95

    MelinG.DanellR. (2006). The top eight percent: development of approved and rejected applicants for a prestigious grant in Sweden. Sci. Public Policy33, 702712. 10.3152/147154306781778579

  • 96

    MertonR. K. (1968). The Matthew effect in science: the reward and communication systems of science are considered. Science159, 5663. 10.1126/science.159.3810.56

  • 97

    MertonR. K.BarberE. (2011). The Travels and Adventures of Serendipity: A Study in Sociological Semantics and the Sociology of Science. Princeton, NJ: Princeton University Press.

  • 98

    MilatA. J.BaumanA. E.RedmanS. (2015). A narrative review of research impact assessment models and methods. Health Res. Policy Syst.13:18. 10.1186/s12961-015-0003-1

  • 99

    MischoW. H.SchlembachM. C.O'DonnellM. N. (2014). An analysis of data management plans in University of Illinois National Science Foundation grant proposals. J. eSci. Librarianship3, 3. 10.7191/jeslib.2014.1060

  • 100

    Molas-GallartJ.TangP.MorrowS. (2000). Assessing the non-academic impact of grant-funded socio-economic research: results from a pilot study. Res. Eval.9, 171182. 10.3152/147154400781777269

  • 101

    MongeonP.BrodeurC.BeaudryC.LariviereV. (2016). Concentration of research funding leads to decreasing marginal returns. Res. Eval.25, 396404. 10.1093/reseval/rvw007

  • 102

    MooreS.NeylonC.EveM. P.O'DonnellD. P.PattinsonD. (2017). “Excellence R Us:” university research and the fetishisation of excellence. Palgr. Commun.3:16105. 10.1057/palcomms.2016.105

  • 103

    MunafoM. R.NosekB. A.BishopD. V.ButtonK. S.ChambersC. D.du SertN. P.et al. (2017). A manifesto for reproducible science. Nat. Hum. Behav.1:21. 10.1038/s41562-016-0021

  • 104

    MutzR.BornmannL.DanielH. D. (2015). Testing for the fairness and predictive validity of research funding decisions: a multilevel multiple imputation for missing data approach using ex-ante and ex-post peer evaluation data from the Austrian science fund. J. Assoc. Inform. Sci. Technol.66, 23212339. 10.1002/asi.23315

  • 105

    NeufeldJ.HuberN.WegnerA. (2013). Peer review-based selection decisions in individual research funding, applicants' publication strategies and performance: the case of the ERC Starting Grants. Res. Eval.22, 237247. 10.1093/reseval/rvt014

  • 106

    NieminenP.CarpenterJ.RuckerG.SchumacherM. (2006). The relationship between quality of research and citation frequency. BMC Med. Res. Methodol.6:42. 10.1186/1471-2288-6-42

  • 107

    NIH (2014). Review Criteria at a Glance. Available online at: https://grants.nih.gov/grants/peer/Review_Criteria_at_a_Glance_MasterOA.pdf (Accessed June 4, 2018).

  • 108

    NIH (2016). Overall Impact Versus Significance. Available online at: https://grants.nih.gov/grants/peer/guidelines_general/impact_significance.pdf (Accessed June 4, 2018).

  • 109

    NIH (2017). Mission and Goals. Available online at: https://www.nih.gov/about-nih/what-we-do/mission-goals (Accessed June 4, 2018).

  • 110

    PayneA. A.SiowA. (2003). Does federal research funding increase university research output?Adv. Econ. Anal. Policy3. 10.2202/1538-0637.1018

  • 111

    PeiferM. (2017). The argument for diversifying the NIH grant portfolio. Mol. Biol. Cell28, 29352940. 10.1091/mbc.e17-07-0462

  • 112

    PierE. L.BrauerM.FilutA.KaatzA.RaclawJ.NathanM. J.et al. (2018). Low agreement among reviewers evaluating the same NIH grant applications. Proc. Natl. Acad. Sci. U.S.A.115, 29522957. 10.1073/pnas.1714379115

  • 113

    PionG.Ionescu-PioggiaM. (2003). Bridging postdoctoral training and a faculty position: initial outcomes of the Burroughs Wellcome Fund Career Awards in the Biomedical Sciences. Acad. Med.78, 177186. 10.1097/00001888-200302000-00012

  • 114

    PionG. M.CordrayD. S. (2008). The burroughs wellcome career award in the biomedical sciences: challenges to and prospects for estimating the causal effects of career development programs. Eval. Health Prof.31, 335369. 10.1177/0163278708324434

  • 115

    PiwowarH. A.DayR. S.FridsmaD. B. (2007). Sharing detailed research data is associated with increased citation rate. PLoS ONE2:e308. 10.1371/journal.pone.0000308

  • 116

    ReinhartM. (2009). Peer review of grant applications in biology and medicine. Reliability, fairness, and validity. Scientometrics81, 789809. 10.1007/s11192-008-2220-7

  • 117

    RobitailleJ. P.MacalusoB.PollittA.GunashekarS.LarivièreV. (2015). Comparative Scientometric Assessment of the Results of ERC-Funded Projects. Bibliometric Assessment Report (D5). Available online at: https://erc.europa.eu/sites/default/files/document/file/ERC_Alternative_Metrics_report.pdf (Accessed July, 2018).

  • 118

    RosenbloomJ. L.GintherD. K.JuhlT.HeppertJ. A. (2015). The effects of research & development funding on scientific productivity: academic chemistry, 1990-2009. PLoS ONE10:e0138176. 10.1371/journal.pone.0138176

  • 119

    SandstromU.HallstenM. (2008). Persistent nepotism in peer-review. Scientometrics74, 175189. 10.1007/s11192-008-0211-3

  • 120

    SanyalP. (2003). Understanding patents: the role of R&D funding sources and the patent office. Econ. Innov. N. Technol.12, 507529. 10.1080/714933760

  • 121

    SarliC. C.DubinskyE. K.HolmesK. L. (2010). Beyond citation analysis: a model for assessment of research impact. J. Med. Libr. Assoc.98:17. 10.3163/1536-5050.98.1.008

  • 122

    SaygitovR. T. (2014). The Impact of Funding through the RF President's Grants for Young Scientists (the field–Medicine) on Research Productivity: a Quasi-Experimental Study and a Brief Systematic Review. PLoS ONE9:e86969. 10.1371/journal.pone.0086969

  • 123

    ScheinerS. M.BouchieL. M. (2013). The predictive power of NSF reviewers and panels. Front. Ecol. Envir.11, 406407. 10.1890/13.WB.017

  • 124

    SpaapenJ.Van DroogeL. (2011). Introducing ‘productive interactions' in social impact assessment. Res. Eval.20, 211218. 10.3152/095820211X12941371876742

  • 125

    StevensG. A.BurleyJ. (1997). 3,000 raw ideas = 1 commercial success!Res. Technol. Manag.40, 1627.

  • 126

    SugimotoC. R.WorkS.LariviereV.HausteinS. (2017). Scholarly use of social media and altmetrics: a review of the literature. J. Assoc. Inform. Sci. Technol.68, 20372062. 10.1002/asi.23833

  • 127

    TenopirC.AllardS.DouglassK.AydinogluA. U.WuL.ReadE.et al. (2011). Data sharing by scientists: practices and perceptions. PLoS ONE6:e21101. 10.1371/journal.pone.0021101

  • 128

    TesauroG. M.SegerY. R.DiJosephL.SchnellJ. D.KleinW. M. (2013). Assessing the value of a Small Grants Program for behavioral research in cancer control. Transl. Behav. Med.4, 7985. 10.1007/s13142-013-0236-x

  • 129

    ThelwallM.MaflahiN. (2016). Guideline references and academic citations as evidence of the clinical value of health research. J. Assoc. Inform. Sci. Technol.67, 960966. 10.1002/asi.23432

  • 130

    UbfalD.MaffioliA. (2011). The impact of funding on research collaboration: evidence from a developing country. Res. Policy40, 12691279. 10.1016/j.respol.2011.05.023

  • 131

    van den BesselaarP.LeydesdorffL. (2009). Past performance, peer review and project selection: a case study in the social and behavioral sciences. Res. Eval.18, 273288. 10.3152/095820209X475360

  • 132

    Van den BesselaarP.SandstromU. (2015). Early career grants, performance, and careers: a study on predictive validity of grant decisions. J. Informetr.9, 826838. 10.1016/j.joi.2015.07.011

  • 133

    Van EckN. J.WaltmanL.van RaanA. F.KlautzR. J.PeulW. C. (2013). Citation analysis may severely underestimate the impact of clinical research as compared to basic research. PLoS ONE8:e62395.

  • 134

    van LeeuwenT. N.MoedH. F. (2012). Funding decisions, peer review, and scientific excellence in physical sciences, chemistry, and geosciences. Res. Eval.21, 189198. 10.1093/reseval/rvs009

  • 135

    Van NoordenR. (2010). Metrics: a profusion of measures. Nature465, 864866. 10.1038/465864a

  • 136

    Van TuylS.WhitmireA. L. (2016). Water, water, everywhere: defining and assessing data sharing in academia. PLoS ONE11:e0147942. 10.1371/journal.pone.0147942

  • 137

    WahlsW. P. (2018). Point of View: the NIH must reduce disparities in funding to maximize its return on investments from taxpayers. Elife7:e34965. 10.7554/eLife.34965

  • 138

    WangD.SongC.BarabásiA. L. (2013). Quantifying long-term scientific impact. Science342, 127132. 10.1126/science.1237825

  • 139

    WarrenH. R.RaisonN.DasguptaP. (2017). The rise of altmetrics. Jama317, 131132. 10.1001/jama.2016.18346

  • 140

    WennerasC.WoldA. (1997). Nepotism and sexism in peer-review. Nature387:341. 10.1038/387341a0

  • 141

    WoodF.WesselyS. (2003). Peer review of grant applications: a systematic review, in Peer Review in Health Sciences, eds GodleeJefferson (London: BMJ Publications), 1431.

Summary

Keywords

research funding, peer review, impact metrics, validity, grant

Citation

Gallo SA and Glisson SR (2018) External Tests of Peer Review Validity Via Impact Measures. Front. Res. Metr. Anal. 3:22. doi: 10.3389/frma.2018.00022

Received

11 April 2018

Accepted

20 July 2018

Published

23 August 2018

Volume

3 - 2018

Edited by

George Chacko, NET eSolutions Corporation (NETE), United States

Reviewed by

Peter Van Den Besselaar, VU University Amsterdam, Netherlands; Kevin Boyack, SciTech Strategies, Inc., United States

Updates

Copyright

*Correspondence: Stephen A. Gallo

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics