Open Access

Does Monetary Support Increase the Number of Scientific Papers? An Interrupted Time Series Analysis

   | Mar 13, 2018

Cite

Introduction

The number of scholarly papers and citations thereto are indirect indicators of the level of scientific development of countries. The number of scholarly papers with Turkish affiliations listed in citation indexes has increased tremendously over the years and Turkey ranks 18th in the world in terms of number of publications. Over 36,000 papers were published in 2015 alone, although their scientific impact in terms of the number of citations they gather is well below the average of the world, the European Union (EU) and the OECD countries.

In 1993, the Turkish Scientific and Technological Research Council (TÜBİTAK) initiated a monetary support program (UBYT) to incentivize researchers and increase the number, impact, and quality of international publications authored by Turkish researchers. Considerable percentages of papers with Turkish affiliations were supported in the early years of this program, even though the rate of support has gradually decreased (to c. 30%) over the years due to the steep increase in the number of published papers with Turkish affiliations. As part of the program, some 157,000 publications (93% of which were papers/articles) were supported between 1997 and 2015. The amount of support paid for each paper has been determined on the basis of the impact factor of the journal in which it was published.

The total amount of support was about 124 million Turkish Liras (in 2015 current prices; equal to c. 35 million USD). The number of papers supported, the total number of publications, and the amount of support increased four-, 10- and 13-fold, respectively, during this period.

The support program has been in place for almost a quarter century. Yet, its impact has not been evaluated in the past. We have been asked by TÜBİTAK to evaluate the effectiveness of the program and given the payment records of 157,000 supported publications. They included, among others, journal information (name, year, its class based on Journal Citation Reports’ subject categories), type of contribution (e.g., article, review) and the amount of support.

Based on the payment records provided, the characteristics (i.e., impact factors) of journals in which supported papers with Turkish affiliations appeared have been analyzed, the functioning of the support algorithm has been studied, and the effectiveness of the overall support program has been evaluated. Findings indicate that the authors of mediocre papers published in journals with relatively low impact factors have mostly been supported due to the use of skewed distributions of journal impact factors in determining the amount of support. The existing support algorithm, on the other hand, does not seem to function as conceived.

This paper presents only the findings of the interrupted time series analysis with a view to find out if the support program has had any impact on the increase of the number of papers with Turkish affiliations. It is organized as follows: The Literature Review section briefly discusses the findings of relevant studies including those that provide some background on the Turkish case. The Data Sources and Method section describes the data used and provides information on interrupted time series analysis. The detailed findings are presented thereafter (Findings and Discussion) along with the limitations of the study. The paper ends with Conclusions.

Literature Review

Performance-based research funding systems (PRFSs) came into being in the 1980s. Based on rewarding the outputs, the rationale of PRFSs is to provide more support to institutions (or individuals) with higher performances so that the ones with lower performances will strive to improve theirs in order to get more support (Herbst, 2007). Yet, it is not clear if PRFSs based on outputs and competition increase the scientific productivity and the impact of outputs. In a relatively recent study comparing PRFSs and outputs of eight countries, countries with less competitive PRFSs such as Denmark turned out to be as effective as the ones with more competitive PRFSs such as the UK and Australia (Auranen & Nieminen, 2010). Some researchers drew attention to the potential “side effects” of PRFSs based on competition, as they tend to “homogenize” research outputs, discourage experiments using new approaches, and reward researchers playing “safe” even though their contributions may not have any societal impact (Geuna & Martin, 2003). The idea of increasing productivity on the basis of outputs and competition seems more complicated than what decision-makers have initially thought (Auranen & Nieminen). For instance, there appears to be some evidence (albeit with relatively small effect sizes) that China’s “cash-for-publication” policy tends to increase researchers’ productivity (Heywood, Wei, & Ye, 2011). Yet, such cash incentives for publications that are in effect in China, South Korea, and Turkey seem to increase the number of submissions but are negatively correlated with acceptance rates (Franzoni, Scellator, & Stephan, 2011).

Based on the countries of the first authors of papers submitted to the journal Science between 2000 and 2009, some 6,228, 1,345, and 84 papers came from China, South Korea, and Turkey, respectively. Yet, only 93 papers from China (1.5%), 18 papers from South Korea (1.3%), and 3 papers from Turkey (3.6%) have been accepted for publication during this period (Franzoni, Scellato, & Stephan, 2011); number of papers submitted to Science along with the accepted number of papers from respective countries come from the Excel tables included in the Supporting Online Material of this article. We calculated the acceptance rates based on the figures provided.

There are mainly two types of PRFSs in use: (1) the ones based on peer review or informed peer review supported with bibliometric measures; and (2) the ones based solely on bibliometric measures such as journal impact factors. The UK’s Research Excellence Framework (REF) is the largest research assessment system in the world (De Boer et al., 2015). Based on peer review, REF has been used since 1986 to distribute funds to research institutes and universities on the basis of their performances. Despite their shortcomings, PRFSs based on bibliometric measures only are on the rise, as they are, in comparison to peer review, easier and less costly to apply as a “proxy” to assess performance. Therefore, they tend to get preferred by increasingly more countries lately.

PRFSs and publication support systems based on bibliometric measures generally use the number of papers published in refereed journals and their impact in terms of citations as the main criteria to determine the research institutes and researchers to be supported. Impact factors (IF) and article influence scores (AIS) of journals are the two most commonly used metrics.

Journal IF was originally proposed by the late Eugene Garfield (1972) to help librarians in their selection of journals for subscription. It is an indicator of the quality of a journal in general and measures the citation impact of an “average” paper published therein. It does not say anything about the quality of an individual paper in that journal and how many citations, if any, it would gather in a certain period of time after its publication (e.g., two years).

Citation distributions used to calculate the IFs of journals are quite skewed, indicating that few papers published in a given journal get cited much more frequently while the majority get unnoticed or rarely cited (Marx & Bornmann, 2013). This is the case even for the most prestigious journals with the highest IFs such as Nature (IF = 38) and Science (IF = 35). As high as 75% of articles published in these journals get cited fewer times than their journal IFs indicate (Larivière et al., 2016, Table 2). Journal IFs vary by scientific discipline, too, as the number of researchers in each field, publication types (i.e., journal articles as opposed to books), and scholarly communication patterns tend to differ. In general, some 9%-10% of all the articles listed in Web of Science collect 44% of the total number of citations (Albarrán et al., 2011). More importantly, there exists no positive relationship between the number of citations that an article gets and the IF of the journal in which it is published (Zhang, Rousseau, & Sivertsen, 2017), and a large body of literature detailing the shortcomings of the use of journal IFs as a performance measure is readily available (e.g., Casadevall & Fang, 2012; Glänzel & Moed, 2002; Marx & Bornmann, 2013; Seglen, 1997; van Raan, 2005; Wouters et al., 2015). Yet, rather than checking the number of citations to the papers of individual researchers, PRFSs based on bibliometric measures continue to use journal IFs to assess the performance of individuals. Journal IFs are quite misleading in predicting the number of citations that any given article might get. What follows are a few examples of PRFS using journal IFs as a research assessment tool.

Number of publications with Turkish affiliations (1976–2015).

YearPapersOtherTotalYearPapersOtherTotal
N%N%NN%N%N
197621680532026919963,35984623163,982
197722972912832019973,84483796174,640
1978272721082838019984,460821,001185,461
1979256711062936219995,201831,078176,279
1980343741232646620005,462841,059166,521
1981299731102740920016,684841,271167,955
1982315701323044720028,985861,4341410,419
19833547214128495200310,662841,9781612,640
19844207712923549200413,199842,4881615,687
19854477614524592200514,194832,8771717,071
19865067715123657200615,070794,0992119,169
19875887717423762200717,853804,4142022,267
19886727522725899200819,327824,3791823,706
198982980209201,038200921,655824,6271826,282
199091278261221,173201022,833834,7601727,593
19911,13480290201,424201123,588825,3251828,913
19921,35177406231,757201225,254825,6071830,861
19931,51976482242,001201326,526797,2002133,726
19941,75473643272,397201427,242797,3152134,557
19952,23372885283,118201528,662797,5302136,192
Total / Avg.318,7098174,72719393,436

Number of papers supported by TÜBİTAK (1997–2015).

Year# of papers supported by TÜBİTAK# of papers with Turkish affiliations (WoS)Percentage supported (%)
19972,2473,84458
19982,6574,46060
19993,0885,20159
20003,2985,46260
20014,2166,68463
20025,8888,98566
20037,51710,66271
20049,51113,19972
20057,03614,19450
20068,12215,07054
200710,55117,85359
200810,41119,32754
200911,55421,65553
201011,59222,83351
20119,57423,58841
201210,64125,25442
201310,20326,52638
201410,25727,24238
20158,01428,66228
Total146,377300,70149

PRFSs are reviewed by several researchers (e.g., De Boer et al., 2015; European Commission, 2010; Geuna & Martin, 2003; Hicks, 2012). Most EU countries, Norway, USA, Australia, New Zealand, and China have some PRFSs in place. We provide a few examples of PRFSs that either solely use journal IF or use it in combination with peer review (excluding the ones based only on peer review such as REF in the UK).

Italy uses a PRFS where an expert panel decides whether to use citation analysis or peer review (or both) for each publication. Universities are ranked on the basis of a quality score consisting of citations and other journal metrics, which determine the amount of support each university gets. Some 30% of the research funds are distributed according to the outcome of this evaluation (Abramo, D’Angelo, & Di Costa, 2011; Abramo & D’Angelo, 2011, 2016).

Similarly, Spain uses a mixed system, although researchers are encouraged to publish in journals that are listed in the top quarters of JCR’s subject categories. Researchers who publish in such journals receive monetary support that ranges somewhere between 3% and 15% of their monthly salaries (Osuna, Cruz-Castro, & Sanz-Menéndez, 2011).

A number of countries such as Czech Republic, China, Finland, and Australia use journal IF exclusively to support research institutes and individual researchers. Finland, for instance, linked journal IF directly with research support by legislation (Adam, 2002). Similarly, Australia and the Czech Republic make direct linkage between research evaluation and funding by counting scholarly outputs and assigning a score to each on the basis of bibliometric measures. These scores are then used to determine the amount of monetary support and papers that appear in refereed journals or journals with relatively higher IFs get the highest scores (Butler, 2004; Butler, 2003; Good et al., 2015, Table 3). Norway also has a similar system based on weighting journals on the basis of various criteria and created three different journal lists (Schneider, 2009). China, on the other hand, uses journal IF most comprehensively in that academic recruitments and promotions, university rankings (and the amount of research support they get), support of Chinese journals that are listed in Chinese Citation Indexes all rely on journal IFs. The procedure seems to have been automated, as a researcher publishing in a journal with a certain IF knows how much support s/he would get. For instance, the author of a paper published in a journal with IF higher than 15 receives 300,000 Yuan (c. 43,000 USD) (Shao & Shen, 2012)! However, the use and appropriateness of such formulaic approaches has been questioned lately with a suggestion that China needs “to rethink its metrics- and citation-based research reward policies” (Teixeira da Silva, 2017).

Test statistic (Ljung Box).

Model Statistics
ModelNumber of PredictorsModel Fit statisticsLjung-Box Q(18)Number of Outliers
Stationary R-squaredStatisticsDFSig.
Makale sayısı-Model_13.60723.53117.1330

Turkey is no exception: journal IFs are considered as an indicator of quality and have been used as an important criterion in academic promotions since early 1990s. In addition to individual universities, TÜBİTAK has initiated a nationwide monetary support system based exclusively on journal IFs. Journals classified under Q1, Q2, etc., in JCR’s subject categories have been used to determine the monetary compensation (Tonta, 2015). More recently (2016), Turkish Higher Education Council (HEC) started a new support scheme based mostly on journal IFs and the faculty whose scores are above a certain threshold in terms of number of academic activities (mostly publications) during the previous year get an additional 10% to 15% on top of their regular monthly salaries throughout the year (Akademik, 2015).

It should be noted that performance-based research funding and publication support systems based on quantitative measures tend to have some adverse effects. Researchers seem to adjust to the requirements very easily and change their publication patterns and behaviors. Such systems are prone to “gaming,” too, and researchers become more “opportunistic” (e.g., publication “inflation”) and less ethical (e.g., “fake” citations) in time. Unintended consequences of PRFSs in several countries (e.g., Australia, Czech Republic, and Spain) were reported in the literature (Butler, 2003; Butler, 2004; Good et al., 2015; Osuna, Cruz-Castro, & Sanz- Menéndez, 2011; Tonta, 2014). For example, more papers tend to get published in journals with relatively lower IFs. A similar trend has also been observed in Turkey (Kamalski et al., 2017;Önder et al., 2008; Yurtsever et al., 2001, 2002). As the Goodhart’s Law states, “When a measure becomes a target, it ceases to be a good . measure.”

It should also be noted that correlation between competitive PRFSs and the research productivity is not clear-cut (Auranen & Nieminen, 2010). Excessive competition seems to reduce the time and energy otherwise to be expended for research. In this paper, we test the conjecture if TÜBİTAK’s publication support system has had an impact on the increase of number of publications listed in citation indexes with Turkish affiliations.

Data Sources and Method

We performed a search on Web of Science (WoS) (December 19, 2016) to identify all the publications with Turkish affiliations listed in Science Citation Index (SCI), Social Sciences Citation Index (SSCI) and Arts & Humanities Citation Index (A&HCI) between 1976 and 2015. More than 390,000 records were retrieved, 81% of which were full papers (articles) while the rest were other types of publications (e.g., reviews, notes, and letters to the editor).

TÜBİTAK provided the payment data for about 157,000 supported publications (93% of which were papers). These records were first cleaned, then coded as either “full papers” (articles) or “other” types of publications, classified under various criteria (e.g., year, class of journal, amount of support paid), ranked and combined, if necessary.

We used MS Excel and SPSS 23 for the detailed analysis of data and prepared both WoS and TÜBİTAK records for interrupted time series analysis outlined below (Interrupted, 2013). (See Appendix A for time series data prepared for interrupted time series analysis.)

The interrupted time series (ITS) analysis technique (also known as quasiexperimental time series analysis or intervention analysis) is used in this paper to measure the impact of TÜBİTAK’s support program. ITS analysis measures if an “event” occurring at any given stage has an immediate or delayed effect on the time series data. For instance, an unexpected political development in a given country may increase the exchange rates, or a terrorist attack may reduce the number of tourists. These “events” (called “interventions”) may be planned or not planned. As ITS analysis is a quasi-experimental method, it is possible (by means of using a control group) to verify if the change has occurred because of the intervention.

ITS analysis is based on the following statistical model:

Yt=βpre+βpost+et,$$\begin{array}{} Y_t=\beta_{pre}+\beta_{post}+e_t, \end{array} $$

where Yt represents the t’th observation in the time series, ßpre and ßpost represent the levels of series before and after the intervention, respectively, and et is the error related with Yt. The null hypothesis

H0=βpreβpost=0,$$\begin{array}{} H_{\it 0}=\beta_{pre}-\beta_{post}={\it 0}, \end{array} $$

states that there is no statistically significant difference between the levels of series before and after the intervention (i.e., it has no impact on dependent variable (McDowall et al., 1980). It is assumed that the parameters in time series models stay the same before and after the intervention and that no other events that affect the parameters take place. ITS analysis can be applied to both static and dynamic (“ergodic”) time series. The ARIMA model is used for non-static series whose arithmetic means, variances, and co-variances change as time passes. This model is expressed as ARIMA (p, d, q) where p, d, and q represent the autoregressive operator (AR), the integrated operator (I), and the moving average operator (MA), respectively. If time series data is not stationary (d), it will first be made stationary to make its mean and variance constant over the years studied.

We have WoS data of publications with Turkish affiliations (1976–2015) and data of supported publications by TÜBİTAK (1997–2015). The program (“intervention”) started in 1993 and enough data points exist both before (1976–1992) and after (1993–2015) the intervention so as to be able to apply ITS analysis to time series data (Cochrane, 2002).

It is not always easy to determine when the performance-based funding system in a given institution is exactly introduced and how long it takes for the system to start to have some impact on the publication output of that institution (van den Besselaar, Heyman, & Sandstrom, 2017; Butler, 2017; Hicks, 2017). We took the date of the decision of TÜBİTAK’s Scientific Board to initiate the support program (June 12, 1993) as the starting date. As relatively fewer researchers benefited from the support program in the early years, we thought that the effect of the program might be observed with some delay (lag) and therefore measured its delayed effect one (1994), four (1997) and 10 years (2003) after of its start.

We have no data on papers (full articles) whose authors have not been supported. However, a relatively small group of authors of other types of contributions can function as a control group, as only 3% of the total amount of support on average was set aside for such contributions even though 19% of publications were of such nature. The authors of other types of contributions were paid half of what the authors of the full papers were, and a mere 1% of the support budget was allocated to them in 2013, for example.

This percentage should ideally be 0 (zero) in order for it to function as a true control group. Yet, we think that it can be used as a control group with some caution and the generalization should be interpreted accordingly.

In other words, we can find out if TÜBİTAK’s support program has had any impact on the increase in the number of papers by comparing it with that of other types of contributions. If the number of other types of contributions that were not well supported did not increase but the number of papers supported increased, we can deduce that the source of the impact was the support program. Conversely, if, despite lack of support, the number of other types of contributions increased along with the number of papers receiving full monetary support, then the increase in the latter cannot be attributed to the program, suggesting that some factor(s) other than the support program may have played a role in this increase.

Findings and Discussion

The descriptive data about the number of papers and the total number of publications originating from Turkey are presented in Table 1 and Figure 1. The rate of increase is quite steep, especially starting from 2000s. This rate of increase made Turkey in those years one of the fastest growing countries in the world in terms of number of papers, and Turkey moved up the ladder very quickly from 45th in 1983 to 25th in 1999 to 18th in 2008 in the world, contributing to 1.56% of the overall scientific production in the world.

Figure 1

Number of papers and total number of publications with Turkish affiliations (1976–2015).

A considerable percentage of these publications were supported by TÜBİTAK’s support program when it was first initiated in 1993. However, the support program seems to have not kept up with the pace of increase of papers and the percentage of papers supported went down from 70% in early 2000s to below 30% in recent years (Table 2, Figure 2).

Figure 2

Number of papers listed in WoS with Turkish affiliations and supported by TÜBİTAK (1997–2015).

The detailed analysis of changes in TÜBİTAK’s support policies over the years is beyond the confines of this paper (see Tonta, 2017b). Instead, we concentrate on whether TÜBİTAK’s support program has actually played a role in the steep rate of increase of papers by Turkish researchers. The time path of the number of papers listed in the Web of Science (WoS) originating from Turkey between 1976 and 2015 is given below (Figure 3). The intervention point (1993) is marked on the graph. As there exists a trend of increase in the number of papers both before and after the intervention, we took the difference of the time series from the 1st level (d = 1) to make it stationary. Consequently, the auto-correlation function (ACF) and partial auto-correlation function (PACF) of the time series became static within the confidence intervals (Figure 4).

Figure 3

Time path of papers with Turkish affiliations (1976–2015).

Figure 4

Correlograms of autocorrelation (ACF) and partial autocorrelations (PACF) functions.

We then defined ARIMA (1, 1, 0) model for interrupted time series data and wanted to see the impact of TÜBİTAK’s support program in 1994, 1997, and 2003 (after one, four and 10 years of its start, respectively). The test statistic of the ARIMA model shows that the defined model is suitable for the time series data (X2 = 23.531, DF = 17, p = .133) (Table 3). The parameters of the ARIMA model (estimates, SE, t- and p- values) are given in Table 4. The ARIMA Model did not produce statistically significant results (coefficient = .153, SE = .170, t = .899, p = .375). The coefficient for “Time series” in Table 4 gives the slope of the regression line before the intervention (14.051), which is used to analyze the different time points by taking into account the existing trend in data before calculating the effect of the intervention. The coefficient for “Before/after Support Program” represents the slope of y- axis when x is equal to 0 (zero) and is used to measure the effect of the intervention in later time points. The coefficient for “Effect” (29.091)gives the difference between slopes before and after the intervention. By adding this difference to the value of pre-intervention slope (14.051), the value of the post-intervention slope (43.142) can be calculated (Interrupted, 2013).

ARIMA Model Parameters.

EstimateSEtSig.
# of papers# of papersNo transformationConstant ARLag 1-57.138334.811-.171.866
Model 1.153.170.899.375
Difference1
Time seriesNo transformationNumeratorLag 014.05129.910.470.642
Before/afterNo transformationNumeratorLag 011.258708.202.016.987
Support
Program
EffectNo transformationNumeratorLag 029.09136.715.792.434

In order to see the effect of the support program on the number of papers with Turkish affiliations, we continued with this model. The slopes of pre- and postintervention are the same for all analyses. It is possible to see the direct effect of the intervention on the number of papers with Turkish affiliations (Table 5). According to the model, an additional 564 papers were published in 1994 because of the support program. However, the effect of the support program is not statistically significant (p = .157). The delayed effect of the program has not been materialized in later years, either, as additional number of papers published due to the program were limited (651 papers in 1997, and 826 in 2003) and the effect is not statistically significant (p > .05). As the effect of the program has been negligible, the formula of the effect of the intervention is not given.

Values showing the delayed effect of TÜBİTAK’s support program.

YearPredicted increaseSEt-valuep-value
1994563.633390.0841.446.157
1997651.241431.1291.510.140
2003825.784571.2791.446.157
20151,174.941947.7611.240.224

Despite the fact that other types of contributions (non-papers) have been supported very little during the period of analysis, the rate of their increase is on a par with that of generously supported papers (see Figure 5). Slopes of linear regression lines of papers and non-papers are almost identical with corresponding R2 values (y = 738.01x – 1E + 06, R2 = 0.814 for papers; and y = 173.78x – 344912, R2 = 0.766 for non-papers). As a control group, the rate of continuous increase in other types of publications seems to confirm the results of the interrupted time series analysis. For instance, some 4,000–7,000 other types of publications have been published annually in recent years, of which only a few hundreds got supported. Yet, the number of other publications continues to increase regardless of support, suggesting that TÜBİTAK’s support program is probably not the main factor causing the increase in the number of papers with Turkish affiliations. The main finding of this paper is, to some extent, in line with the evidence that researchers with Turkish affiliations do not seem to attach too much importance to TÜBİTAK’s “cash-for-publication” program (Yuret, 2017).

Figure 5

Rate of increase of papers and non-papers.

Note: Scales for y axes on the left and right are different. y axis on the left represents the number of papers while the one on the right respresents the number of non-papers.

Limitations of the Study

It should be noted though that interrupted time series analysis has some limitations. The assumption that no other “event” or “events” occurred during the period of analysis that might have affected the time series data is one of them. For example, the prerequisite of having papers published in journals listed in citation indexes for academic promotion may have triggered this increase, as more than 90% of research in Turkey has been carried out in universities, and the number of academic personnel in universities has increased tremendously over the years. Moreover, in addition to the number of research personnel in universities, the number of papers may be increasing due to a number of other factors such as the number of researchers per 10,000 capita, and the share of R&D expenditures within the Gross National Product (GDP). As indicated earlier, even though some positive correlation between PRFSs and the number of papers has been observed, this may not necessarily point to a strong causality between the two. As was the case in Spain (Osuna, Cruz-Castro, & Sanz-Menéndez, 2011), the number of papers with Turkish affiliations continues to increase perhaps not because of TÜBİTAK’s support program but because of other factors such as the growth in and the maturity of universities’ research systems including academic personnel.

We should also note that the interrupted time series analysis tells us whether the intervention has had any significant effect on the dependent variable or not but it does not tell us what caused the increase in the number of papers if it was not the intervention. To find out this, we carried out a multiple regression analysis and observed fairly strong correlation between the number of papers with Turkish affiliations and the number of academic personnel as well as the number of supported papers. However, we decided not to report the results of the multiple regression analysis, as the Durbin-Watson statistic was rather small (0.921), probably indicating the existence of serial autocorrelation between variables and thereby making the results less reliable. This can to some extent be observed from Figure 2: the correlation between the number of papers with Turkish affiliations and the number of supported papers was positive and statistically significant between 1997 and 2006 whereas it was negative and not statistically significant between 2007 and 2015.

For a more definitive answer to the question of whether TÜBİTAK’s support program has had any effect on the increase in the number of papers with Turkish affiliations, a true control group is needed. In other words, the rate of increase of papers supported by TÜBİTAK needs to be compared with that of non-supported ones. Even such a comparison may not necessarily be sufficient to reveal the causality, should there be any, between TÜBİTAK’s support program and the steep increase in the number of papers with Turkish affiliations. For this, individual level data for both TÜBİTAK-supported and non-supported papers are needed to see if the increase is due to the increase of the productivity of: (1) the same researchers benefiting from TÜBİTAK’s support program; (2) more researchers responding to TÜBİTAK’s cash incentives; (3) researchers who have not sought TÜBİTAK support for their papers in the past at all; or (4) a combination of some or all of the above.

Conclusions

As part of TÜBİTAK’s support program, the authors of over 157,000 publications received more than 124 million Turkish Liras (in 2015 current prices, c. 35 million USD) as monetary support between 1997 and 2015. Yet, two thirds of all payments were less than 826 liras (or c. 230 USD). These “micropayments” might be one of the reasons why, according to the test results of the interrupted time series analysis, the program did not seem to have direct impact on the increase of the number of papers published by Turkish authors. It is likely that small amounts of payments were not much of an incentive for authors to publish more.

We should point out that the objective of the support program is not to increase the number of papers per se but to increase their impact and quality, as stated in the By-Law of TÜBİTAK’s support program (TÜBİTAK, 2016). Some authors may find the small payments satisfactory. Yet, if such small payments do not help achieve the program’s objectives, precautions should be taken to correct it. The support program seems to have functioned as a mechanism to transfer small amounts of payments to authors without any considerable improvement in the impact and quality of the papers. Transaction costs of such small payments should be borne in mind as well as the costs of missed opportunities of increasing the impact and quality of papers. For instance, it might perhaps be a better strategy to concentrate limited resources on a few high impact projects rather than to disperse them every year as “pocket money” to the authors of some 10,000 papers that appear mostly in journals with relatively low impact factors. Sustainability of the existing support program should also be considered and its impact should be monitored more often.

Such support programs should function as a leverage to speed up the scientific and economic development of countries.A thorough study as to why the support program did not seem to function as intended should be carried out. After a comprehensive review of existing support programs, new policies should be instituted to increase the impact and quality of scientific papers originating from Turkey, and TÜBİTAK’s support program should be redesigned accordingly.

Based on 25 years’ worth of payments data, this is perhaps one of the first large-scale studies showing that “cash-for-publication” policies or “piece rates” paid to researchers tend to have little or no effect on the increase of researchers’ productivity. The main finding of this paper has some implications for countries wherein publication subsidies are used as an incentive to increase the number and quality of papers published in international journals. They should be prepared to consider reviewing their existing support programs (based usually on bibliometric measures such as journal impact factors) and revising their reward policies.

eISSN:
2543-683X
Language:
English
Publication timeframe:
4 times per year
Journal Subjects:
Computer Sciences, Information Technology, Project Management, Databases and Data Mining