This study analyzes funding acknowledgments in scientific papers to investigate relationships between research sponsorship and publication impacts. We identify acknowledgments to research sponsors for nanotechnology papers published in the Web of Science during a one-year sample period. We examine the citations accrued by these papers and the journal impact factors of their publication titles. The results show that publications from grant sponsored research exhibit higher impacts in terms of both journal ranking and citation counts than research that is not grant sponsored. We discuss the method and models used, and the insights provided by this approach as well as it limitations.
Citation: Wang J, Shapira P (2015) Is There a Relationship between Research Sponsorship and Publication Impact? An Analysis of Funding Acknowledgments in Nanotechnology Papers. PLoS ONE 10(2): e0117727. https://doi.org/10.1371/journal.pone.0117727
Academic Editor: K. Brad Wray, State University of New York, Oswego, UNITED STATES
Received: September 4, 2014; Accepted: January 2, 2015; Published: February 19, 2015
Copyright: © 2015 Wang, Shapira. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited
Funding: This research was supported in part by the Center for Nanotechnology in Society at Arizona State University (CNS-ASU) under National Science Foundation Grant No. 0531194. (http://www.nsf.gov) (PS JW). Additional support was provided by the Manchester Institute of Innovation Research, at the Manchester Business School, University of Manchester. (https://research.mbs.ac.uk/innovation/) (PS). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
With demands by scientists, universities, and other stakeholders to expand research and development (R&D) funding, there are also growing pressures to assess and justify the impacts of R&D expenditures. Governments typically sponsor most basic R&D and thus underwrite a significant share of the knowledge production and resulting scientific publications that are generated by national R&D investments . For government agencies in many countries, public accountability is requiring more attention to the effectiveness and efficiency of public research funding. In the US, for example, the Government Performance and Result Act of 1993  aims to increase the efficiency, effectiveness and accountability of federal spending, leading to the use of performance measures with identified R&D goals and outcomes. The Science of Science and Innovation Policy (SciSIP) Program of the US National Science Foundation particularly supports research on approaches to measure returns from R&D investments . Similarly, the Japan Science and Technology Agency, through its Research Institute of Science and Technology for Society, has established a Program in Science of Science, Technology and Innovation Policy to assess the economic and social impacts of R&D investments . While much attention is of course focused on industrial and commercial outcomes, there is widespread interest in scientific results, including both the quantity and quality of publication outputs from R&D, and this interest extends not only to the public sector but also to foundations and corporate research sponsors.
Many studies have attempted to quantify the economic return and non-economic return of investment in R&D. The former includes those measuring the impact of private R&D investment [5–9] and impact of public R&D investment [10–14] on economic growth. The latter includes studies measuring the effect of R&D funding on the amount of scientific output such as publications and patents [15–18] and the impact and quality of scientific output [19–24].
This paper seeks to extend work in the second category, on the relationships between R&D sponsorship and publication outputs. We examine funding acknowledgments reported in scientific papers to investigate the relationships between funding and research impacts. Utilizing the context information in the acknowledgment fields in publications, the approach allows for large scale analysis of funding acknowledgment variables [25–27]. This approach uses publicly-accessible data sources and, unlike many prior approaches, is not limited to any particular funding agency or research institution, which facilitates comparisons across various boundaries (institutional, sponsor, national, and discipline). While this study makes use of publications in the interdisciplinary domain of nanotechnology, the approach can readily be extended to other fields.
A two-stage regression model is used to test the effects of grant funding on research impacts. The first stage regression examines whether papers supported by grant funding sources are more likely to get published in high impact factor journals, which can be viewed as recognition by peer reviewers. The second stage regression examines whether papers with grant funding are more likely to receive attention and generate citations after being published. This can be viewed as indicating use by and diffusion to the broader scientific community. Two bibliometric indicators—journal impact factor and citation counts—are used to measure research outputs in these two stages. Both indicators are not without critiques. It is argued that these measures are proxies for quality, although it is also argued that such claims should be treated carefully due to field differences, biases in peer review, citations clubs, and the possibility of negative as well as positive types of citations . Nevertheless, these two measures are frequently used as indicators for assessing the publication outputs of scientists and their research groups, by promotion and tenure committees and in studies of the performance of research institutions and nations. The use of both indicators in combination can offset some of the problems associated if they are used individually. A dataset of over 89000 nanotechnology publications is used to empirically test the model effects. The results show that publications associated with grant sponsored research do exhibit higher impacts in terms of both journal ranking and citation counts, controlling for field differences. At the same time, impacts also vary by funding sources and patterns. Our empirical results and their implications are discussed in detail later in the paper, following discussion of the existing literature and of our study hypotheses.
This section of the paper reviews a range of literature that reports on prior work related to research sponsorship and publication impacts. In this process, we draw out four hypotheses to test in our subsequent large-scale empirical work. We first consider the subject hypotheses in the context of the literature, and then discuss how we operationalize the notion of publication impact.
In general, the extant literature suggests that there are connections between research funding and research outputs, although there is a nuanced debate about the nature and direction of effects. A predominant perspective is that research funding has positive impacts on research outputs. Researchers who secure extra funding may well be more determined and ambitious in their research goals and able to garner more resources (e.g. for personnel, equipment, materials, and travel) for implementing their research. In some instances, efforts to secure additional research funding may reflect departmental or institutional pressures to develop research credibility or enhance visibility. Grant proposals typically go through review processes in order to get funded—for examples, see relevant sections on peer and merit review processes promulgated by leading funding agencies such as the UK Engineering and Physical Sciences Research Council (EPSRC) , the US National Institutes of Health (NIH)  and the US National Science Foundation (NSF) . Funding agencies may have their own research agendas, and target proposals that fit these priorities, and they may also have open submission rounds where researchers propose new research topics. In either case, peer review processes aim to filter out weaker research proposals and reward more promising research ideas. On the other hand, peer review can be flawed, biased or conservative, and the administrative time and costs associated with proposing and managing grant sponsored research may distract from scientific advancement . One US study of faculty workloads found that more than two-fifths of time allocated to federally funded research was spent on pre- and post-award administrative activities rather than active research, with variations across institutions and disciplines . Research groups with better administrative coordination or more institutional support may be less distracted by grant award transaction costs. However, such issues may be less apparent when research is block-supported, i.e. where scientists receive support for their research work on a regular basis, usually from internally-allocated resources, without the need to compete for external sources of grant funds. Block-supported research may allow researchers the flexibility and security of exploring riskier new ideas—or, conversely, such research might become staid without the stimulus or requirement of justifying new research ideas to external reviewers.
The empirical evidence to date on the effects of different modes of sponsoring research is mixed. For example, a study of NIH-supported researchers shows that funded publications appear to have higher citations than average (including non-grant sponsored research and presumably block-supported) publications . Similar findings were reported in a study of researchers in an Australian university  and a study of researchers funded by the National Cancer Institute of Canada , where grant funded publications appear to have higher citations than non-grant sponsored research papers. By contrast, Harter and Hooten  found that articles published in the Journal of the American Society for Information Science (JASIS) exhibit no difference in citation counts with regard to their funding status. This result was confirmed in their subsequent study which expanded the scope of JASIS articles under study to additional years . Similarly, Cronin and Shaw  in their study of four information science journals found no significant difference in citations between grant sponsored and non-grant sponsored research articles.
The mixed results found in previous studies can be attributed partly to different subjects under study and partly to differences in the analytical methods used. More specifically, subjects in these studies were limited in sample size and varied from publications by researchers funded by a particular agency  , to publications by researchers in a particular institution , and to articles from selected journals [19–20] . The inherent variances in the characteristics of these subjects may lead to observed differences in the relationship between funding status and citation. Additionally, these studies used different analytical approaches. Harter and Hooten employed a correlation test between funding status and citation counts  . Cronin and Shaw , Trochim et al.  and Campbell et al.  compared the difference of means between citation of funded and non-funded publications. A major problem with both correlation tests and comparison of means is that neither control for other factors affecting citations so as to isolate the impact of funding. Sandstrom  undertook a regression analysis with variables including several grant aspects as independent variables and citation as a dependent variable. However, these regression results need to be interpreted with caution due to the unclear causal effect direction between funding and citation in the dataset as noted by the author.
In short, while these prior studies offer insights, there remains considerable scope (and need) to advance work on the relationships between research sponsorship and publication impact and to introduce appropriate controls. The recent availability of large-scale information on funding acknowledgments in scientific research papers allows us to initiate a further and more extensive examination of this question. In this paper, we investigate four straightforward yet still critical hypotheses. To start, we examine the funding status of publications and explore if grant sponsored research receives more attention and has higher impact. By “grant sponsored research”, we mean research that explicitly acknowledges a research sponsor or grant award as providing support to the research work reported in a paper. In contrast, “non-grant sponsored research” refers to papers without funding acknowledgment. The research reported in such papers may indeed not be explicitly funded although in most cases it is likely to be block-supported by internal institutional resources. As outside research grants are often competitive and typically require external peer-review, low-quality research proposals should be screened out. It is thus plausible to conjecture that grant sponsored research, taken as a whole, will be of better quality compared with research that is not sponsored or is block-supported without going through competitive external review processes. This leads us to the initial hypothesis (H1) that grant sponsored research will be associated with higher publication impacts.
Grant sponsored research publications may have multiple funding sources. In some cases, it is the result of joint solicitation from funding agencies. For example, the National Institute of General Medical Sciences (NIGMS) and the National Science Foundation’s Division of Mathematical Sciences (NSF/DMS) jointly call for proposals titled “Research at the Interface of the Biological and Mathematical Sciences.” In other cases, the publication is based on a research project funded by different agencies at different stages. However, the majority of papers which acknowledge co-funding are the product of intellectual collaboration between authors from different organizations/countries with separate funding support. These authors report the particular funding sources brought together to support the collaboration. Indeed, the relationship between co-authorship and co-funding is found to be positive. As such, a question arises whether single or multiple funding sources will achieve higher research impacts. Arguably, research with multiple funding sources has gone through more stages of review, and thus might have higher quality. Alternatively, more funding sources might require more attention to administration and coordination, which could distract from the research effort. To test, we put forward a second hypothesis (H2): research that receives sponsorship from a greater number of funding sources will have higher publication impacts.
International collaboration has been an increasing pattern in the scientific world. Not only are scientists from different countries working together, funding agencies are also seeking cooperation with foreign partners in supporting the advance of science. The US NSF recognizes the importance of international science and engineering partnership and suggests using it as a tool to address global challenges . The European Commission Framework Programme intentionally requires that the proposal should be comprised of teams from multiple countries. International research groups are believed to have stronger scientific power because of their access to additional resources . Narin et al.  also suggested a self-selection effect as scientists doing well in research are more likely to travel and co-author papers internationally. These observations support the proposition that research which involves international collaboration will tend to have higher impact. Similarly, publications with international co-funding support will also exhibit higher impact. This suggests a third hypothesis (H3): research that receives grant support from multiple countries will have higher publication impacts.
We further posit that leading countries in key scientific fields have particular research capabilities and the potential to identify and support emerging topics that can then more broadly influence future scientific directions. Research areas funded by leading countries not only direct the research interests of scientists but also influence funding priorities of other countries. Hence, we anticipate that research funded by leading countries will receive more attention in the field. This leads to our fourth hypothesis (H4): research receiving grant funding from leading countries in a scientific domain will tend to have higher publication impacts.
Two indicators are used to measure the publication impact of research: citation counts and journal impact factors. Citation counts are often used as an indicator for the impact and diffusion of research. Citations map the intellectual and knowledge linkages between the source article and reference article. High citation counts have been positively correlated with other recognized impact indicators such as peer assessment and honorific awards . The frequency of a paper being cited shows that other researchers have recognized it. We fully recognize that the use of citation counts as a measure of research quality is subject to well-known limitations, such as critical citations, self-citations, and field variance .
The journal impact factor reflects the average citation performance of a journal. It was proposed by Garfield  and is used in the Journal Citation Report (JCR)  published by Thomson Reuters to indicate the relative ranking of journals. The impact factor is calculated as the ratio of citations received by the journal in a particular year to the number of publications in that journal in the previous two years. No impact factor is reported for journals that are newly indexed in JCR or that do not publish regularly. A journal with a higher impact factor suggests that the journal is of more prominent in its field as a source of reference knowledge and attracts papers which are more likely, on average, to be well-cited (recognizing that not all papers in journals with high impact factors receive citations). There is critical debate about using the impact factor to measure journal quality . For example, the impact factor is discipline dependent where some disciplines have much more citations than other disciplines due to field size or differential propensity to cite. Additionally, journals with more review articles tend to receive more citations.
Data and Model
The study uses the funding acknowledgment analysis approach proposed by Wang and Shapira  to detect the linkage between funding and research output. The acknowledgment section of a publication provides information on the financial and intellectual support received by the authors, where the names of funding organizations and often the grant numbers are specified. The acknowledgment section of the Thomson Reuters Web of Science (WoS) publication index allows for analysis of funding and publications at a large scale.
The analysis uses data of nanotechnology publications over the one year period August 2008 – July 2009. August 2008 represents the first month that funding acknowledgments were available (at the time of our analysis) in WoS publication records. We extracted publications that recorded publication year and month for our study time period (inclusive) from a global nanotechnology publication database that was developed by the Nanotechnology Research and Innovation Systems Assessment Group at Georgia Institute of Technology (Georgia Tech), Atlanta, USA. Nanotechnology is a broad domain of research at the nanoscale (1–100 nanometers) involving multiple fields including engineering, physics, chemistry, microscopy, materials, electronics, and biology. Details of the definition of nanotechnology and the search strategy are contained in Porter et al. . After excluding publications with missing journal impact factor information, the dataset used in this study contains 89,605 nanotechnology publications, with 67% reporting funding acknowledgment information (Table 1). Where funding acknowledgment is not included, this is primarily because the research was funded internally or unsponsored but using available resources. There is also the possibility of unintentional or intentional neglect to acknowledge funding . We judge that this does not pose a serious threat to our analysis as it is frequently (and increasingly) mandatory for researchers funded by public and other research sponsors to disclose funding sources in resulting publications. Empirical examination has found a strong correlation between grant data and publication funding acknowledgment . The share of publications reporting funding acknowledgment is also comparable with other studies  . A quarter of the publications have a single funding source and 42% have multiple funding sources. Over half of the publications receive funding from only one country (Table 2). Funding from two or more countries supports around 11% of publications. As funding arrangements can be complicated and varied in how they are reported in the acknowledgment text, with differing acronyms and misspelling of funding agencies, a large amount of work was required to text mine, clean, match and validate the data .
In our analysis, two dependent variables are used: impact factor (IMPACT_FACTOR) and citation (CITATION). Both impact factor and citation indicate impact but in different ways. The impact factor shows the broad relative citation ranking of a journal and in our case it measures acceptance by peer reviewers, while citation measures recognition by the research field. It is possible that individual articles in highly-ranked (“good”) journals do not receive many citations while some articles in less-well ranked journals can be cited extensively. The impact factor is also used as a control variable in the citation model to test whether articles in good journals receive more citations. The journal impact factor is derived from the WoS Journal Citation Reports (JCR) . JCR impact factors were sought for the journal titles of all publications in our nanotechnology data set. Impact factors (2009) were found for 3686 or 93.3% of the 3952 titles, covering 89605 or 97.8% of the papers. For the titles in our nanotechnology publications database, the median journal impact factor was 2.7, with a mean of 3.2. Data on forward citations is also retrieved from WoS and counted as of December 2010. This should be regarded as an early measure of citations, representing those garnered between 17 to 28 months following journal publication. Around 19% of articles received no citations from publication to this date. Overall, the average citation count is 4.7 and the median is 3. Since citation is a count variable, a negative binomial regression is used to model the relationship between funding and citation. The negative binomial regression is good for modeling over-dispersed count outcome variables .
Variables on author numbers, affiliations, and countries, as well as fields of research are included as control variables to isolate the impact of funding. Collaborative work, academic affiliation, and country locations are believed to be positively associated with research quality. Collaborative work is measured by the number of authors on a publication (AUTHORS). Author affiliation is indicated by four dummy variables, coded as 1 if one of the author affiliations is in academia, industry, government research centers, or hospitals respectively. The count on author affiliations (AFFILIATIONS) for an article is included to show if the research involves cross-institution collaboration. Four dummy variables on country location are coded for authors from the US (US_AUTH), China (CHINA_AUTH), Germany (GERMANY_AUTH) and Japan (JAPAN_AUTH), the top four countries in nanotechnology publications. The number of author country locations (AUTH_COUNTRIES) is counted to indicate whether the article is a product of international collaboration. In addition, research impact measures based on citations are discipline dependent, where certain fields generate more citations than others. A set of 16 field dummy variables is used to control for field disciplinary impacts. When modeling against citations, the effect of the age of an article (ARTICLE_AGE) is also controlled since earlier publications have more time to accrue citations than later publications. Our benchmark is December 2010 (the month of data download), and we model the number of months since publication backwards from this time period. A star scientist variable (STAR_SCIENTIST) is introduced to control for the influence of the reputation associated with notable scientists. Scientists who are highly cited in the past may tend to receive more citations in the future, in part due to their standing and because their work may be more likely to be accepted into high impact journals, as well as because of the quality of their work. Star scientist is a dummy variable coded as 1 if the article has at least one author with at least 1500 citations (from all WoS papers) in 2000–2007, and 0 if otherwise. Table 3 and Table 4 present the description and summary statistics of all the variables.
To test the hypotheses proposed above, we use two sets of models. In the first set of models, we compare grant sponsored with non-grant sponsored research papers and use a dummy variable (FUNDED) showing the status of funding as the key independent variable. In the second set of models, we look at grant sponsored publications only and examine whether the diversity of funding sources has any impact on research quality. To give consideration to the argument that too much collaboration or too many funding organizations might impose added burdens related to administration and communication and thus detract from research quality, we add squared terms to account for quadratic relationships . Count variables for collaboration and funding are tested together with their quadratic counterparts, including the number of authors, the number of author affiliations, the number of author countries, the number of funding sources and the number of funding countries.
In addition, we are interested in exploring whether funding support influences research impact differently for high impact papers or high journal placement papers. A more comprehensive picture of the covariant effects can be obtained by using quantile regression. Quantile regression allows us to see the relationship between independent variables and specific quantiles of the dependent variable . This allows us to examine if funding influence varies among papers in different quantiles of research impact. The four hypotheses are again tested in seven quantiles of journal impact factor and citations. The same set of control variables are used in each of these quantile models.
In the first set of models, we compare research impact measured by journal impact factor (Model 1) and citation counts (Model 2) with funding status. The dummy variable FUNDED is used to indicate whether the paper declares its funding source in the acknowledgment text. The coefficients of this variable are positive and significant in both models. The result shows that publications with funding acknowledgment are more likely to get published in high impact journals and receive more citations thereafter (Table 5). The coefficients in these two models are rather consistent. However, research collaboration exhibits interesting and mixed impacts. While the number of authors contributes to research impact, the number of author affiliations and the number of author countries show U-shaped and inverted U-shaped relationships respectively. According to the coefficients, research impact declines along with the increase of the number of affiliations from 1 to 10, and then increases when the number of affiliations is beyond 10. With the increase of the number of author countries, research impact tends to diminish except in the beginning where authorship from two countries seems to be better than single country authorship. As for other control variables, authorship involving star scientists increases research impact. The type of author affiliation is not a major factor, although papers with corporate authors are less likely to get into high impact journals while those with hospital affiliations are more likely to publish in such journals. It seems that clinical research attracts more citations. Papers published by authors from leading scientific countries garner higher impacts except for authors from China where there are negative coefficients in both models. As expected, the journal impact factor and the age of article both positively influence citations.
In the next set of models, we test the relationship between funding diversity and research impact (Table 6). Three types of variables are used to measure funding diversity: the number of funding sources as in FUNDERS (Models 3 and 6), the number of funding countries as in FUND_COUNTRIES (Models 4 and 7), and funding provided by four leading countries the US, China, Germany and Japan as in US_FUND, CHINA_FUND, GERMANY_FUND and JAPAN_FUND (Models 5 and 8). The European Union (EU) as a separate entity also provides various research funding opportunities, including through EU Framework Programmes and the European Network of Excellence. Scientists from European countries are not only eligible for funding support from their own countries, but also from European Union. Therefore we include European Union funding (EU_FUND) as a separate funding source to delineate its impact. Funding sources are not mutually exclusive. Overall, the funding diversity variables show a positive impact on research quality with some slight variations. The more funding sources acknowledged in a paper, the more likely the paper will be found to be published in a highly ranked journal. A quadratic relationship is found between the number of funding sources and received citation counts. Citations increase with the number of funding sources, peaks at around 14 sources and then decrease, implying that an optimal level of funding diversity exists for citations. The number of funding countries is found to be positively linked to both impact factor and citations, although nonlinear in terms of the former. The journal impact factor increases at an accelerating pace with the increase of funding countries. Regarding the country origin of funding, the EU, the US, Germany and Japan all appear to have positive impacts on the rank of published journals, but positive effects are retained only for the EU, the US and Germany when it comes to the citation measure. The impact of control variables is found to be consistent with those in Model 1 and Model 2 with the coefficients of control variables very similar to those reported in Table 5.
Using the same set of control variables, the relationship between funding and research impact is tested in quantile regressions. Table 7 and Table 8 report the results for funding status, where the coefficients of control variables are omitted. The impact of funding is much stronger in the upper quantiles of the distribution. For example, the difference in journal impact factor between grant sponsored and non-grant sponsored research publications is 0.24 at the 5th percentile of the conditional distribution and 0.482 at the 75th percentile. Similarly, the disparity between citations of grant sponsored and non-grant sponsored research paper is close to 0 at the 5th quantile and is 0.2 at the 90th quantile. Generally speaking, high impact papers are much more likely to be associated with acknowledged funding compared with low impact papers. However, we also note that the difference is not as big in the very high end of the distribution of journal impact factor (90th and 95th percentiles).
In terms of funding diversity, both the number of funding sources and the number of funding countries show stronger impact in the right tail of the distribution (Tables 9 and 10). The marginal impact of the number of funding organizations is 0.079 at the 5th percentile and 0.198 at the 95th percentile for impact factor, and 0.007 at the 5th percentile and 0.277 at the 95th percentile for citations. Meanwhile, the marginal impact of the number of funding countries increases constantly across percentiles and is tripled at the 95th percentile compared with the 5th percentile for impact factor, and even 30 times at the 95th percentile compared with the 5th percentile for citations. The importance of funding diversity is more evident in high impact articles.
Funding from selected leading countries shows its role in improving research impact, which also increases across quantiles. The impact of funding from the EU, the US, Germany and Japan is higher on the upper quantiles of the distribution of impact factor. For citation counts, only EU funding and US funding are mostly significant in the whole range of the distribution. Similarly, the disparity between publications received funding from these two sources and those without is much higher in the right tail of the distribution.
Conclusion and Discussion
The study explored how funding affects research impacts. It is among the first set of studies that text mine funding acknowledgment in WoS records to systematically examine the relationship between funding and impact at a large scale. Our analysis of the funding acknowledgment section of nanotechnology publications in 2008–2009 finds that outputs from grant sponsored research exhibit higher impacts than outputs from non-grant sponsored research. Grant sponsored articles are not only more likely to get published in highly ranked journals, but also to generate more research interest in the field as measured by forward citations. The diversity of funding sources has a more variable influence on research impact. The number of funding sources acknowledged in publications is positive on placement in high quality journals, but tends to be concave on received citations, increasing before reaching the optimal number of funding sources, and then decreasing. This suggests that the quality assurance of multiple review processes is effective to some extent but also that too many funding entities may increase transactional burdens and distract from the research itself. The number of funding countries contributes to both journal placement and citation counts. Research supported by funding from multiple countries does imply international value and potential, which may make it more acceptable to peer reviewers and the research community. Research grants provided by selected leading countries/blocs such as the EU, the US and Germany are even more influential. Publications indicating financial support from these countries/blocs appear more often in good journals and receive more attention through citations. It is plausible that research funded by these countries is of higher quality, although other factors may also be at work. For example, it is possible that research priorities funded by leading countries shapes research directions in other countries and thus generates recognition and follower citations.
In addition, according to the quantile regression results, grant funding appears to have stronger influence in the production of high impact research publications. This suggests that the grant review process does select more promising projects, since it is at the higher ends of impact where the disparity between grant sponsored and non-grant sponsored research is most substantial. To some extent, this finding validates the use of grant funding mechanisms to allocate R&D funds, although we emphasize that this result should be interpreted cautiously. There is some circularity in the process: more motivated and innovative researchers are more likely to get grants and in turn produce research that achieves higher impact. This is a positive cyclic process. Those who are not able to break into this cycle may face difficulties in an increasingly competitive research world.
There are caveats that are relevant to this study. On the one hand, not all grant funding is acknowledged. Some might be undisclosed intentionally or simply neglected. On the other hand, not all acknowledged funding is acquired through competitive peer-reviewed process. Some may be earmarked or supported by institutional funds. Therefore, acknowledged funding cannot be equated with competitive grant funding in every case. Nevertheless, given the high correlation between grant funding and publication acknowledgment  , we judge that this limitation does not overly bias the results. As discussed in the paper, we acknowledge that there are limitations to the use of journal impact factors and citation counts to measure research quality.
This study seeks to untangle the relationships between grant funding and research impact. It provides empirical evidence of the effectiveness of grant funding schemes and the scientific publication impacts of different combinations of research investments. The study suggests that research collaboration is beneficial not only by bringing together different skills and mindsets but also by coupling financial resources to produce research outputs that generate higher publication impacts. As for policy implications, the study suggests that joint solicitation for research proposals by different funding agencies is a productive way to promote effective research partnerships. Research funding agencies may also find it productive to encourage more international collaborative research activities by their home scientists and to actively collaborate with counterparts in other countries.
We acknowledge the assistance of colleagues with the Nanotechnology Research and Innovation Systems Assessment group at Georgia Institute of Technology, particularly Jan Youtie and Alan Porter, in making the underlying dataset available to us. We also acknowledge comments provided by participants at the symposium panel on “Leveraging Resources, Organization, and Collaboration for Breakthrough Science” at the Annual Meeting of the American Association for the Advancement of Science (AAAS), Chicago, IL, February 17, 2014, and by anonymous reviewers of this journal.
Conceived and designed the experiments: JW PS. Performed the experiments: JW. Analyzed the data: JW PS. Wrote the paper: PS JW.
- 1. OECD (2011) Main Science and Technology Indicators (MSTI) Vol. 2011/1. OECD Publishing. DOI: https://doi.org/10.1787/msti-v2011-1-en-fr. pmid:23243387
- 2. GPRA (1993) Government Performance and Results Act of 1993. Public Law (103–62).
- 3. NSF (2011) Science of Science and Innovation Policy (SciSIP). National Science Foundation. Available: http://www.nsf.gov/funding/pgm_summ.jsp?pims_id=501084&org=sbe. Accessed 2011 September 13. https://doi.org/10.1080/17437199.2011.587961 pmid:25473706
- 4. RISTEX (2011) R&D Program: Science of Science, Technology and Innovation Policy. Research Institute of Science and Technology, Tokyo, Japan. Available: http://www.ristex.jp/EN/examin/stipolicy/index.html. Accessed 2011 September 13.
- 5. Solow RM (1957) Technical change and the aggregate production function. The Review of Economics and Statistics 39(3): 312–320.
- 6. Terleckyj NE (1974) Effects of R&D on the Productivity Growth of Industries: An exploratory study. Washington, DC: National Planning Association. pmid:25057719
- 7. Griliches Z (1985) Productivity, R&D, and basic research at the firm level in the 1970s. NBER Working Paper No. W1547. Boston, MA: National Bureau of Economic Research.
- 8. Mansfield E (1980) Basic research and productivity increase in manufacturing. American Economic Review 70(5): 863–873.
- 9. Mansfield E (1981) How economists see R&D. Harvard Business Review 59(6): 98–106.
- 10. Terleckyj NE (1985) Measuring economic effects of federal R&D expenditures: Recent history with special emphasis on federal R&D performed in industry. Papers Commissioned for a Workshop on the Federal Role in Research and Development. The National Academies Press. pp.151–172.
- 11. Lichtenberg FR (1985) Assessing the impact of federal industrial R&D expenditures on private R&D activity. Papers commissioned for a workshop on the federal role in research and development. The National Academies Press. pp. 115–150.
- 12. Pavitt K (1991) What makes basic research economically useful? Research Policy 20: 109–119.
- 13. Martin B, Salter A, Hicks D, Pavitt K, Senker J, Sharp M, von Tunzelmann N (1996) The relationship between publicly funded basic research and economic performance: A SPRU review. Report prepared for HM treasury. University of Sussex, Brighton.
- 14. Paasi M (1998) Efficiency of innovation systems in the transition countries. Economic Systems 22(3): 217–234.
- 15. McAllister PR, Narin F, Corrigan JG (1983) Programmatic evaluation and comparison based on standardized citation scores. IEEE Transactions on Engineering Management EM-30 4.
- 16. Adams JD, Griliches Z (1998) Research productivity in a system of universities. Annales d’Economie et de Statisque 49/50.
- 17. Payne AA, Siow A (2003) Does federal research funding increase university research output? Advances in Economic Analysis & Policy 3(1) Article 1. pmid:25605453
- 18. Huang Z, Chen H, Li X, Roco MC (2006) Connecting NSF funding to patent innovation in nanotechnology (2001–2004). Journal of Nanoparticle Research 8: 859–879.
- 19. Cronin B, Shaw D (1999) Citation, funding acknowledgment and author nationality relationships in four information science journals. Journal of Documentation 55(4): 402–408.
- 20. Harter SP, Hooten PA (1992) Information science and scientists: JASIS, 1972–1990. Journal of the American Society for Information Science 43(9): 583–593.
- 21. van Leeuwen TN, van der Wurff LJ, van Raan AFJ (2001) The use of combined bibliometric methods in research funding policy. Research Evaluation 10(3): 195–201.
- 22. Trochim WM, Marcus SE, Masse LC, Mose RP, Weld PC (2008) The evaluation of large research initiatives: a participatory integrative mixed-methods approach. American Journal of Evaluation 29: 8–28.
- 23. Sandstrom U (2009) Research quality and diversity of funding: A model for relating research money to output of research. Scientometrics 79(2): 341–349.
- 24. Campbell D, Picard-Aitken M, Cote G, Caruso J, Valentim R, Edmonds S, Williams GT, Macaluso B, Robitaille JP, Bastien N, Laframboise MC, Lebeau LM, Mirabel P, Lariviere V, Archambault E (2010) Bibliometrics as a performance measurement tool for research evaluation: The case of research funded by the National Cancer Institute of Canada. American Journal of Evaluation 31(1): 66–83.
- 25. Shapira P, Wang J (2010) Follow the money: What was the impact of the nanotechnology funding boom of the past ten years? Nature 468: 627–628. pmid:21124430
- 26. Wang J, Shapira P (2011) Funding acknowledgment analysis—an enhanced tool to investigate research sponsorship impacts: The case of nanotechnology. Scientometrics 87(3): 563–586.
- 27. Costas R, van Leeuwen T (2012) Approaching the “reward triangle”: General analysis of the presence of funding acknowledgments and “peer interactive communication” in scientific publications. Journal of the American Society for Information Science and Technology 63(8): 1647–1661.
- 28. King J (1987) A review of bibliometrics and other science indicators and their role in research evaluation. Journal of Information Science 13: 261–276.
- 29. EPSRC (2013) Peer Review Principles. Engineering and Physical Sciences Research Council. Available: http://www.epsrc.ac.uk/funding/guidance/basics/Pages/prprinciples.aspx. Accessed 2013 March 30.
- 30. NIH (2011) Peer review process. National Institutes of Health, Office of Extra Mural Research, US Department of Health and Human Services. Available: http://grants.nih.gov/grants/peer_review_process.htm. Accessed 2011 December 7.
- 31. NSF (2013) Proposal and award policies and procedure guide. National Science Foundation. NSF 14–1. Available: http://www.nsf.gov/pubs/policydocs/pappguide/nsf14001/nsf14_1.pdf. Accessed 2014 December 10. https://doi.org/10.14802/jmd.14038 pmid:25616262
- 32. Langfeldt L (2001) The decision-making constraints and processes of grant peer review, and their effects on the review outcome. Social Studies of Science 31(6): 820–841.
- 33. Decker RS, Wimsatt L, Trice AG, Konstan JA (2011) A profile of federal-grant administrative burden among federal demonstration partnership faculty. A Report of the Faculty Standing Committee of the Federal Demonstration Partnership. Available: http://www.iscintelligence.com/archivos_subidos/usfacultyburden_5.pdf. Accessed 2014 December 10.
- 34. Harter SP, Hooten PA (1990) Factors affecting funding and citation rates in information science publications. Library and Information Science Research 10: 263–280.
- 35. National Science Foundation (2014). Joint DMS/NIGMS Initiative to Support Research at the Interface of the Biological and Mathematical Sciences. Program Solicitation 13–570. Available: http://www.nsf.gov/pubs/2013/nsf13570/nsf13570.htm. Accessed 2014 December 10.
- 36. NSB (2008) International Science and Engineering Partnerships: A priority for U.S. Foreign Policy and Our Nation’s Innovation Enterprise. National Science Board.
- 37. Aksens DW (2003) Characteristics of highly cited papers. Research Evaluation 12(3): 159–170.
- 38. Narin F, Stevens K, Whitlow ES (1991) Scientific Co-operation in Europe and the Citation of Multinationally Authored Papers. Scientometrics 21(3): 313–323.
- 39. MacRoberts MH, MacRoberts B (1989) Problems of citation analysis: A critical review. Journal of the American Society for Information Science 40(5): 342–349.
- 40. Garfield E (1972) Citation analysis as a tool in journal evaluation. Science 178 (4060): 471–479. pmid:5079701
- 41. Thomson Reuters (2012) Journal Citation Reports—Journal Impact Factor. Available: http://admin-apps.webofknowledge.com/JCR/help/h_impfact.htm. Accessed 2014 December 10.
- 42. Pendlebury DA, Adams A (2012) Comments on a critique of the Thomson Reuters journal impact factor. Scientometrics 92 (2): 395–401.
- 43. Porter AL, Youtie J, Shapira P, Schoeneck DJ (2008) Refining search terms for nanotechnology. Journal of Nanoparticle Research 10(5): 715–728.
- 44. Lewison G, Carding P. (2003) Evaluating UK research in speech and language therapy. International Journal of Language and Communication Disorders 38(1): 48–65.
- 45. Butler L (2001) Revisiting bibliometric issues using new empirical data. Research Evaluation 10(1): 59–65.
- 46. Long JS, Freese J (2006) Regression models for categorical dependent variables using Stata. Stata Press. pmid:25590126
- 47. Wooldridge JM (2009) Introductory econometrics: a modern approach. Mason, OH: South-Western Cengage Learning. pmid:25506961
- 48. Koenker R, Hallock KF (2001) Quantile regression. Journal of Economic Perspectives 15(4): 143–156.