[go: up one dir, main page]

Showing posts with label Bibliometrics. Show all posts
Showing posts with label Bibliometrics. Show all posts

Tuesday, December 20, 2022

Changes in Authorship, Networks, and Research Topics in Ecosystem Services

I have a new paper, coauthored with Ida Kubiszewski, Bob Costanza, and Luke Concollato, which investigates the development of the field of ecosystem services over the last decade since the founding of the journal Ecosystem Services. This is an open access publication – my first in a so-called hybrid journal. We used the University College London read and publish agreement with Elsevier to publish the paper. ANU now has a similar agreement starting in 2023.

The paper updates Ida and Bob's paper published in 2012: "The authorship structure of ‘‘ecosystem services’’ as a transdisciplinary field of scholarship". In this paper, we update and expand that analysis and compare results with those we found in the previous analysis. We also analyse the influence that the journal Ecosystem Services has had on the field over its first 10 years. We look at which articles have had the most influence on the field (as measured by the number of citations in Ecosystem Services) and on the broader scientific literature (as measured by total citations). We also look at how authorship networks, topics, and the types of journals publishing on the topic have changed. 

Not surprisingly, there has been significant growth in the number of authors (12,795 to 91,051) and number of articles published (4,948 to 33,973) on ecosystem services since 2012. Authorship networks have also expanded significantly, and the patterns of co-authorship have evolved in interesting ways. The most prolific authors are no longer in as tight clusters as they were 10 years ago.

The network chart shows the coauthorship relations among the 163 most prolific authors – those authors who have published more than 30 articles in the field. Colors indicate continent: Yellow = North America, red = South America, blue = Europe, purple – Africa, green = Asia, and orange = Oceania. The greatest number of authors is in Europe and they almost all collaborate with other top authors. Only in Asia and to a lesser degree North America are there top authors who do not collaborate with other top authors.

Costanza et al. (1997) is the most influential article in terms of citations in the journal Ecosystem Services and "Global estimates of the value of ecosystems andtheir services in monetary units" by de Groot et al. (2012) is now the most cited article published in Ecosystem Services.

Ecosystem Services is now the most prolific publisher of articles on ecosystem services among all the journals that have published in the area. There are nine journals that are both on the list of the 20 journals cited most often in Ecosystem Services and on the list of the top 20 journals cited by articles published in Ecosystem Services: Ecosystem Services, Ecological Economics, Ecological Indicators, Science of the Total Environment, Land Use Policy, Journal of Environmental Management, PLoS One, Ecology and Society, and Environmental Science & Policy.

Sunday, August 7, 2022

Trends in RePEc Downloads and Abstract Views

For the first time in a decade, I updated my spreadsheet on downloads and abstract views per person and per item on RePEc.


The downward trends I identified ten years ago have continued, though there was an uptick during the pandemic, which has now dissipated. There was more of an increase in abstract views than in downloads in the pandemic.

Since the end of 2011 both abstract views and downloads per paper have fallen by about 80%. Total papers rose by around 260%, while total downloads fell 38% and total abstract views 27%. 

I'd guess that a mixture of the explanatory factors I suggested last time has continued to be in play.


Thursday, June 2, 2022

Confidence Intervals for Recursive Journal Impact Factors

I have a new working paper coauthored with Johannes König and Richard Tol. It's a follow up to my 2013 paper in the Journal of Economic Literature, where I computed standard errors for simple journal impact factors for all economics journals and tried to evaluate whether the differences between journals were significant.* In the new paper, we develop standard errors and confidence intervals for recursive journal impact factors, which take into account that some citations are more prestigious than others, as well as for the associated ranks of journals. We again apply these methods to the all economics journals included in the Web of Science.

Recursive impact factors include the popular Scimago Journal Rank, or SJR, and Clarivate's Article Influence score. We use Pinski and Narin's invariant method, which has been used in some rankings of economics journals

As simple impact factors are just the mean citations an article published in a journal in a given period receives in a later year, it is easy to compute standard errors for them using the formula for the standard error of the mean. But the vector of recursive impact factors is the positive eigenvector of a matrix and its variance does not have a simple analytical form.

So, we use bootstrapping to estimate the distribution of each impact factor. Taking all 88,928 articles published in 2014-18 in the economics journals included in the Web of Science, we resample from this dataset and compute the vector of recursive impact factors from the new dataset.** Repeating this 1,000 times we pick the 2.5% or 97.5% range of values for each journal to get a 95% confidence interval:

95% confidence intervals of the recursive impact factor, arithmetic scale (left axis) and logarithmic scale (right axis).

The graph repeats the same data twice with different scales so that it's possible to see some detail for both high- and low-ranked journals. Also, notice that while the confidence intervals for the highest ranked journals are quite symmetric, confidence intervals become increasingly asymmetric as we go down the ranks. 

The top ranked journal, the Quarterly Journal of Economics, clearly stands out above all others. The confidence intervals of all other journals overlap with those of other journals and so the ranks of these journals are somewhat uncertain.*** So, next we construct confidence intervals for the journals' ranks.

It turns out that there are a few ways to do this. We could just construct a journal ranking for each iteration of the bootstrap and then derive the distribution of ranks for each individual journal across the 1,000 iterations. Hall and Miller (2009), Xie et al. (2009), and Mogstad et al. (2020) show that this procedure may not be consistent when some of the groups (here journals) being ranked are tied or close to tied. The corrected confidence intervals are generally broader than the naive bootstrap approach.

We compute confidence intervals for ranks using the simple bootstrap, the Xie et al. method, and the Mogstad et al. method:

 

95% confidence intervals of the rank based on the recursive impact factor. The inner intervals are based on Goldstein’s bootstrap method, the middle intervals use Xie’s correction to the bootstrap, and the outer intervals follow Mogstad’s pairwise comparison.

The simple bootstrap under-estimates the true range of ranks, while it seems that the Mogstad et al. method might be overly conservative. On the other hand, Xie et al.' s approach depends on choosing a couple of "tuning parameters".  

All methods agree that the confidence interval of the rank of the Quarterly Journal of Economics only includes one. Based on the simple bootstrap, the remainder of the "Top-5'' journals are in the top 6 together with the Journal of Finance, while the Xie et al. method and the Mogstad et al. methods generally broaden estimated confidence intervals, particularly for mid-ranking journals. All methods agree that most apparent differences in journal quality are, in fact, mostly insignificant. We think that impact factors, whether simple or recursive should always be published together with confidence intervals.

* The latter exercises were a bit naive. As pointed out by Horrace and Parmeter (2017), we need to account for the issue of multiple comparisons. 

** Previous research on this topic resampled at the journal level, missing most of the variation in citation counts.

*** Overlapping confidence intervals don't neccessarily mean that there is no signficant difference between two means. Goldstein and Harvey (1995) show that the correct confidence intervals for such a test of the difference between two means are narrower than the conventional 95% confidence intervals. On the other hand, for multiple comparisons we would want wider confidence intervals.

Wednesday, May 12, 2021

Public Policy Schools in the Asia-Pacific Ranked

I have a new paper with my Crawford School colleague Bjoern Dressel published in Asia & the Pacific Policy Studies (open access). The data and figures for the article are on Figshare. Bjoern has been interested for a while in ranking public policy schools in the Asia-Pacific region.  But a comprehensive ranking seemed hard to achieve. Recently, I came across an article by Ash and Urquiola (2020) that ranks US public policy schools according to their research output and impact. Well, we thought, if they can rank schools just by their research output and not by their education and public policy impact then so can we 😀. Research is the easiest component to evaluate.

We compare the publication output of 45 schools with at least one publication listed in Scopus between 2014 and 2018, based on affiliations listed on the publications rather than current faculty. We compute the 5-Year impact factor for each school. This is identical to the impact factor reported for academic journals, but we compute it for a school rather than a journal. It is the mean number of citations received in 2019 by a publication published between 2014 and 2018. This can be seen as an estimate of research quality. We also report the standard error of the impact factor as in my 2013 article in the Journal of Economic Literature. If we treat the impact factor as an estimate of the research quality of a school then we can construct a confidence interval to express how certain or uncertain we are about that estimate. This graph shows the schools ranked by impact factor with a 90% confidence interval:

Peking and Melbourne are the two top-ranked schools but the point estimates have a very wide confidence interval. This is because their research output is relatively small and the variance of citations is quite large. The third ranked school – SGPP in Indonesia – only had two publications in our target period. After that there are several schools with much narrower confidence intervals. These mostly have more publications.


Here we can see the impact factors on the y-axis and the number of publications of each school on the x-axis. Three schools clearly stand out at the right: Crawford, Lee Kwan Yew, and Tsinghua. These schools are also top-ranked by total citations, which combines the quality and quantity variables. The three top schools account for 54% of publications and 63% of citations from the region.

In general, the elite schools are in China and Australia. Australia has three out of the top ten schools ranked by impact factor and total citations, despite its small population size. China, on the other hand has at least five schools ranked in the top ten across both rankings, which is remarkable given that many of these schools have been established only in the last 15 years (though linked to well-established research universities).

We found more schools that had no publications in Scopus in the target period. Perhaps in some cases they are too new, or faculty use their other affiliations, but clearly there is a lot of variation in research-intensiveness. Somewhat surprising is the low ranking of public policy schools in Japan and India – both countries with a considerable number of public policy schools, but none in the top ten schools when ranked by 5-year citation impact factor or total number of citations. 

One reason for the strong performance of the Chinese schools is that they focus to some degree on environmental issues, and particularly climate change, where citation numbers tend to be higher. We did not adjust for differences in citations across fields in this research, but this is something that future research should address.



Saturday, January 6, 2018

How to Count Citations If You Must

That is the title of a paper in the American Economic Review by Motty Perry and Philip Reny. They present five axioms that they argue a good index of individual citation performance should conform to. They show that the only index that satisfies all five axioms is the Euclidean length of the list of citations to each of a researcher's publications – in other words, the square root of the sum of squares of the citations to each of their papers.* This index puts much more weight on highly cited papers and much less on little cited papers than simply adding up a researcher's total citations would. This is a result of their "depth relevance" axiom. A citation index that is depth relevant always increases when some of the citations of a researcher's less cited papers are instead transferred to some of the researcher's more cited papers. In the extreme, it rewards "one hit wonders" who have a single highly cited paper, over consistent performers who have a more extensive body of work with the same total number of citations.

The Euclidean index is an example of what economists call constant elasticity of substitution, or CES, functions. Instead of squaring each citation number, we could raise it to a different power, such as 1.5, 0.5, or anything else. Perry and Reny show that the rank correlation between the National Research Council peer-reviewed ranks of the top 50 U.S. economics departments and the CES citation indices of the faculty employed in those departments is at a maximum for a power of 1.85:



This is close to 2 and suggests that the market for economists values citations in a similar way to the Euclidean index.

RePEc acted unusually quickly to add this index to their rankings. Richard Tol and I have a new working paper that discusses this new citation metric. We introduce an alternative axiom: "breadth relevance", which rewards consistent achievers. This axiom states that a citation index always increases when some citations from highly cited papers are shifted to less cited papers. We also reanalyze the dataset of economists at the top 50 U.S. departments that Perry and Reny looked at and a much larger dataset that we scraped from CitEc for economists at the 400 international universities ranked by QS. Unlike Perry and Reny, we take into account the fact that citations accumulate over a researcher's career and so junior researchers with few citations aren't necessarily weaker researchers than senior researchers with more citations. Instead, we need to compare citation performance within each cohort of researchers measured by the years since they got their PhD or published their first paper.

We show that a breadth relevant index that also satisfies Perry and Reny's other axioms is a CES function with exponent of less than one. Our empirical analysis finds that the distribution of economists across departments is in fact explained best by the simple sum of their citations, which is equivalent to a CES function with exponent of one, that favors neither depth nor breadth. However, at lower ranked departments – departments ranked by QS from 51 to 400 – the Euclidean index does explain the distribution of economists better than does total citations.


In this graph, the full sample is the same dataset that Perry and Reny used in their graph. The peak correlation is for a lower exponent – tau or sigma** – simply because we take into account cohort effects by computing the correlation for a researcher's citation index relative to the cohort mean.*** While the distribution across the top 25 departments is similarly to the full sample, with a peak at a slightly lower exponent that is very close to one, we don't find any correlation between citations and department rank for the next 25 departments. It seems that there aren't big differences between them.

Here are the correlations for the larger dataset that uses CitEc citations for the 400 universities ranked by QS:


For the top 50 universities, the peak correlation is for an exponent of 1.39 but for the next 350 universities the peak correlation is for 2.22. The paper also includes parametric maximum likelihood estimates that come to similar conclusions.

Breadth per se does not explain the distribution of researchers in our sample, but the highest ranked universities appear to weight breadth and depth equally, while lower-ranked universities do focus on depth, giving more weight to a few highly cited papers.

A possible speculative explanation of behavior across the spectrum of universities could be as follows. Lowest-ranked universities, outside of the 400 universities ranked by QS, might simply care about publication without worrying about impact. Having more publications would be better than having fewer at these institutions, suggesting a breadth relevant citation index. Our exploratory analysis that includes universities outside of those ranked by QS supports this. We found that breadth was inversely correlated with average citations in the lower percentiles.

Middle-ranked universities, such as those ranked between 400 and 50 in the QS ranking, care about impact; having some high-impact publications is better than having none and a depth-relevant index describes behavior in this interval. Finally, among the top-ranked universities such as the QS top 50 or NRC top 25, hiring and tenure committees wish to see high-impact research across all of a researcher's publications and the best-fit index moves towards. Here, adding lower-impact publications to a publication list that contains high-impact ones is seen as a negative.

* As monotonic transformations of the index also satisfy the same axioms, the simplest index that satisfies the axioms is simply the sum of squares.

** In the paper, we refer to an exponent of less than one as tau and an exponent greater than one as sigma.

*** The Ellison dataset that Perry and Reny use, uses Google Scholar data and truncates each researcher's publication list at 100 papers. With all working paper variants, it's not hard to exceed 100 items. This could bias the analysis in favor of depth rather than breadth. We think that the correlation computed for researchers with 100 papers or less only is a better way to test whether depth or breadth best explains the distribution of economists across departments. The correlation peaks very close to one for this dataset.

Thursday, November 9, 2017

Distribution of SNIP and SJR

My colleague Matt Dornan asked me about the distribution of the journal impact factors SNIP and SJR. Crawford School has used SNIP to compare journal rankings across disciplines. It is a journal impact factor that is normalized for the differences in citation potential in different fields. This makes it reasonable to compare Nature and the Quarterly Journal of Economics, for example. Nature looks like it has much higher impact using the simple Journal Impact Factor that just counts how many citations articles in a journal get. But taking citation potential into account, these journals look much more similar. SJR is an iterative indicator that takes into account how highly cited the journals which cite the journal of interest are. It is similar to Google's pagerank algorithm. It also can be compared across disciplines. SJR is more an indicator of journal prestige or importance rather than simple popularity. I've advocated SJR as a better measure of journal impact as some journals in my area have high citations but those are not in the really highly cited journals.

I assumed the distribution of these indicators was highly skewed with most journals having low impact. But I also assumed that the log of these indicators might be normally distributed as citations to individual papers is roughly log-normally distributed. It turns out that the log of SNIP is still a bit skewed, but not that far from normal:



On the other hand the log of SJR remains highly non-normal:


There is a small tail of high prestige journals and then a big bulk of low prestige journals and a huge spike at SJR = 0.1. It makes sense that it is harder to be prestigious rather than just popular, but still I am surprised by how extreme the skew is. The distribution appears closer to an exponential distribution than a normal distribution.

Friday, October 6, 2017

Impact Factors for Public Policy Schools

As part of our self-evaluation for the upcoming review of the Crawford School, I have been doing some bibliometric analysis. One thing I have come up with is calculating an impact factor for the School and some comparator institutions. This is easy to do in Scopus. It's the same idea as computing one for an individual or a journal, of course. I am using a 2016, 5 year impact factor. Just get total citations in 2016 to all articles and reviews published in 2011-2015. Divide by the number of articles. Here are the results with 95% confidence intervals:


The main difficulty I had was retrieving articles for some institutions such as the School of Public Affairs at Sciences Po. Very few articles came back for various variants of the name that I tried. I suspect that faculty are using departmental affiliations. I had a similar problem with IPA at LSE. So, I report the whole of LSE in the graph. It is easy to understand this metric in comparison to journal impact factors. As an individual metric the confidence interval will usually be large, though my 2016 impact factor was 5.9 with a 4.2 to 7.5 confidence interval. That's more precise than the estimate for SIPA.

Tuesday, September 5, 2017

Confidence Intervals for Journal Impact Factors

Is the Poisson distribution a short-cut to getting standard errors for journal impact factors? The nice thing about the Poisson distribution is that the variance is equal to the mean. The journal impact factor is the mean number of citations received in a given year by articles published in a journal in the previous few years. So if citations followed a Poisson distribution it would be easy to compute a standard error for the impact factor. The only additional information you would need besides the impact factor itself, is the number of articles published in the relevant previous years.

This is the idea behind Darren Greenwood's 2007 paper on credible intervals for journal impact factors. As he takes a Bayesian approach things are a little more complicated in practice. Now, earlier this year Lutz Bornmann published a letter in Scientometrics that also proposes using the Poisson distribution to compute uncertainty bounds - this time, frequentist confidence intervals. Using the data from my 2013 paper in the Journal of Economic Literature, I investigated whether this proposal would work. My comment on Bornmann's letter is now published in Scientometrics.

It is not necessarily a good assumption that citations follow a Poisson process. First, it is well-known that the number of citations received each year by an article, first increases and then decreases (Fok and Franses, 2007; Stern, 2014) and so the simple Poisson assumption cannot be true for individual articles. For example, Fok and Franses argue that for articles that receive at least some citations, the profile of citations over time follows the Bass model. Furthermore, articles in a journal vary in quality and do not all each have the same expected number of citations. Previous research finds that the distribution of citations across a group of articles is related to the log-normal distribution (Stringer et al., 2010; Wang et al., 2013).

Stern (2013) computed the actual observed standard deviation of citations in 2011 at the journal level for all articles published in the previous five years in all 230 journals in the economics subject category of the Journal Citation Reports using the standard formula for the variance
where Vi is the variance of citations received in 2011 for all articles published in journal i between 2006 and 2010 inclusively, Ni is the number of articles published in the journal in that period, Cj is the number of citations received in 2011 by article j published in the relevant period, and Mi is the 5-year impact factor of the journal. Then the standard error of the impact factor is √(Vi/Ni ).

Table 1 in Stern (2013) presents the standard deviation of citations, the estimated 5-year impact factor, the standard error of that impact factor, and a 95% confidence interval for all 230 journals. Also included are the number of articles published in the five year window, the official impact factor published in the Journal Citation Reports and the median citations for each journal.

The following graph plots the variance against the mean for the 229 journals with non-zero impact factors:



There is a strong linear relationship between the logs of the mean and the variance but it is obvious  that the variance is not equal to the mean for this dataset. A simple regression of the log of the variance of citations on the log of the mean yields:

where standard errors are given in parentheses. The R-squared of this regression is 0.92. If citations followed the Poisson distribution, the constant would be zero and the slope would be equal to one. These hypotheses are clearly rejected. Using the Poisson assumption for these journals would result in underestimating the width of the confidence interval for almost all journals, especially those with higher impact factors. In fact, only four journals have variances equal to or smaller than their impact factors. As an example, the standard error of the impact factor estimated by Stern (2013) for the Quarterly Journal of Economics is 0.57. The Poisson approach yields 0.2.

Unfortunately, accurately computing standard errors and confidence intervals for journal impact factors appears to be harder than just referring to the impact factor and number of articles published. But it is not very difficult to download the citations to articles in a target set of journals from the Web of Science or Scopus and compute the confidence intervals from them. I downloaded the data and did the main computations in my 2013 paper in a single day. It would be trivially easy for Clarivate, Elsevier, or other providers to report standard errors.

References

Bornmann, L. (2017) Confidence intervals for Journal Impact Factors, Scientometrics 111:1869–1871.

Fok, D. and P. H. Franses (2007) Modeling the diffusion of scientific publications, Journal of Econometrics 139: 376-390.

Stern, D. I. (2013) Uncertainty measures for economics journal impact factors, Journal of Economic Literature 51(1), 173-189.

Stern, D. I. (2014) High-ranked social science journal articles can be identified from early citation information, PLoS ONE 9(11), e112520.

Stringer, M. J, Sales-Pardo, M., Nunes Amaral, L. A. (2010) Statistical validation of a global model for the distribution of the ultimate number of citations accrued by papers published in a scientific journal, Journal of the American Society for Information Science and Technology 61(7): 1377–1385.

Wang, D., Song C., Barabási A.-L. (2013) Quantifying long-term scientific impact, Science 342: 127–132.

Tuesday, March 28, 2017

Cohort Size and Cohort Age at Top US Economics Departments

I'm working on a new bibliometrics paper with Richard Tol. We are using Glenn Ellison's data set on economists at the top 50 U.S. economics departments as a testbed for our ideas. I had to compute the size of each year cohort for one of our calculations, and thought this graph of the number of economists at the 50 departments in each "academic age" year was interesting:


There isn't as sharp a post-tenure drop-off in numbers as you might expect, given the supposed strict tenure hurdle these departments impose. But as we can see the cohorts increase in size up to year 5, which might be explained by post-docs and other temporary appointments, or people even moving up the rankings after a few years at a lower ranked department. So, as a result, the tenure or out year would be spread over a few years too. On the other hand, as the data were collected in 2011, the Great Recession might also explain lower numbers for the first few years.

A post-retirement drop-off only really seems to occur after 39 years. The oldest person in the study by academic age was Arnold Harberger.

Thursday, December 29, 2016

Ranking Economics Institutions Applying a Frontier Approach to RePEc data

Back in 2010 I posted that the RePEc ranking of economics institutions needed to be adjusted by size. Better quality institutions do tend to be bigger but as RePEc just sums up publications, citations etc rather than averaging them larger institutions also get a higher RePEc ranking even if they aren't actually better quality. In the post, I suggested using a frontier approach. The idea is that the average faculty member at Harvard perhaps is similar to one at Chicago (I haven't checked this), but because Harvard is bigger it is better. So, looking at average scores of faculty members might produce a misleading ranking.

A reader sent me an e-mail query about an updated version of this and I thought that was a good idea for a new post:


The chart shows the RePEc rank for 190 top-level institutions (I deleted NBER) against their number of registered people on RePEc. I drew a concave frontier by hand. How have things changed since 2010? The main change is the appearance of Stanford on the frontier. Also, the Federal Reserve is now listed as one institution, so the Minnesota Fed has dropped off the frontier. Dartmouth is now slightly behind the frontier and Tel Aviv looks like it has also lost a little ground. Otherwise, not much has changed.

Tuesday, July 12, 2016

Legitimate Uses for Impact Factors

I wrote a long comment on this blogpost by Ludo Waltman but it got eaten by their system, so I'm rewriting it in a more expanded form as a blogpost of my own. Waltman argues, I think, that for those that reject the use of journal impact factors to evaluate individual papers, such as Lariviere et al., there should be then no legitimate uses for impact factors. I don't think this is true.

The impact factor was first used by Eugene Garfield to decide which additional journals to add to the Science Citation Index he created. Similarly, librarians can use impact factors to decide on which journals to subscribe or unsubscribe from and publishers and editors can use such metrics to track the impact of their journals. These are all sensible uses of the impact factor that I think no-one would disagree with. Of course, we can argue about whether the mean number of citations that articles receive in a journal is the best metric and I think that standard errors - as I suggested in my Journal of Economic Literature article - or the complete distribution as suggested by Lariviere et al., should be provided alongside them.

I actually think that impact factors or similar metrics are useful to assess very recently published articles, as I show in my PLoS One paper, before they manage to accrue many citations. Also, impact factors seem to be a proxy for journal acceptance rates or selectivity, which we only have limited data on. But ruling these out as legitimate uses doesn't mean rejecting the use of such metrics entirely.

I disagree with the comment by David Colquhoun that no working scientists look at journal impact factors when assessing individual papers or scientists. Maybe this is the case in his corner of the research universe but it definitely is not the case in my corner. Most economists pay much, much more attention to where a paper was published than how many citations it has received. And researchers in the other fields I interact with also pay a lot of attention to journal reputations, though they usually also pay more attention to citations as well. Of course, I think that economists should pay much more attention to citations too.


Thursday, March 3, 2016

My Submission to Stern Review of the REF

The Stern Review of the REF (Research Excellence Framework) is the latest British government review of research assessment in the UK, following on from the Metric Tide assessment. I have just made a submission to the enquiry. My main comment in response to the first question (1. What changes to existing processes could more efficiently or more accurately assess the outputs, impacts and contexts of research in order to allocate QR? Should the definition of impact be broadened or refined? Is there scope for more or different use of metrics in any areas?) follows:

"I think that there is substantial scope for using bibliometrics in the conduct of the REF. In Australia the Australian Research Council uses metrics to assess natural science disciplines and psychology. Research that I have conducted with my coauthor, Stephan Bruns, shows that this approach could be extended to economics and probably political science and perhaps other social sciences. We have written a working paper presenting our results that is currently under review by Scientometrics.

The paper shows that university rankings in economics based on long-run citation counts can be easily predicted using early citations. The rank correlation between universities' cumulative citations received over ten years for economics articles published in 2003 and 2004 and citations received in 2003 to 2004 alone is 0.91 in the UK and 0.82 in Australia. We compare these citation-based university rankings with the rankings of the 2008 Research Assessment Exercise in the UK and the 2010 Excellence in Research assessment in Australia. Rank correlations are quite strong but there are differences between rankings based on this type of peer review and rankings based on citation counts. However, if assessors are willing to consider citation analysis to assess some disciplines as is the case for the natural sciences and psychology in Australia there seems no reason to not include economics in this set.

Previously, I published a paper, published in PLoS One showing that the predictability of citations at the article level is similar in economics and political science. This supports the view that metrics based research assessment can cover both economics and political science in addition to the natural sciences and economics.

I believe the REF review should seriously consider these findings in producing recommendations for a lighter touch future REF."

I also made briefer responses to some of their other questions. In particular:

5. How might the REF be further refined or used by Government to incentivise constructive and creative behaviours such as promoting interdisciplinary research, collaboration between universities, and/or collaboration between universities and other public or private sector
bodies?


"A major issue with the REF and the ERA in Australia is the pigeon-holing of research into disciplines, which might not match well the nature of the research conducted. This clearly will discourage publication in interdisciplinary venues that may not be as respected by mainstream reviewers. The situation is less acute in Australia where a single output can be allocated across different assessment disciplines, but I still think that assessment by pure disciplinary panels discourages interdisciplinary work in Australia. So, I imagine this is exacerbated in the UK.

7. In your view how does the REF process influence the development of academic disciplines or impact upon other areas of scholarly activity relative to other factors? What changes would create or sustain positive influences in the future?

Johnston et al. (2014) show that the total number of economics students has increased in UK more rapidly than the total number of all students, but the number of departments offering economics degrees has declined, particularly in post-1992 universities. Also, the number of universities submitting to the REF under economics has declined sharply with only 3 post-1992 universities submitting in the latest round. This suggests that the REF has driven a concentration of economics research in the more elite universities in the UK.

Johnston, J., Reeves, A. and Talbot, S. (2014). ‘Has economics become an elite subject for elite UK universities?’ Oxford Review of Education, vol. 40(5), pp. 590-609.

Tuesday, January 19, 2016

Influential Publications in Ecological Economics Revisited to be Published in... Ecological Economics

Our paper on the changes over the last decade in patterns of influence in ecological economics has been accepted for publication. Not very surprisingly the journal where it will be published is Ecological Economics. Elsevier have already sent me an e-mail saying that I should expect the proofs on 21 January! That is fast.

Friday, October 30, 2015

Google Scholar Matures

Since it was introduced in late 2004, Google Scholar has rapidly grown to become a widely used tool for finding and assessing the impact of academic literature. The database still suffers from noise relative to its competitors Scopus and Web of Science but it has broader coverage, especially in the social sciences and humanities and is open access. As the database developed, Google have periodically added new information sources to the database. This resulted in a rapid growth in estimated citations of articles in the early years. However, it now seems that the database has matured. The following graph shows the growth rate of citations to my research in the previous 12 months, measured monthly since 2009 for Google Scholar in blue and a bit more intermittently for Scopus in red. I have also fitted exponential trend lines to the two series:

Initially the growth rate of Google Scholar citations was very high and very erratic. But the month to month variation in the annual growth rate has reduced drastically over time. By contrast, the growth rate of Scopus citations has been much more consistent, with a slow rate of decline in the percentage growth rate over time. Interestingly, the two series have also converged to a common growth rate of 17-18% per year. So, it seems that Google's database is now as mature as Scopus is. This doesn't mean that Google is now as high a quality data source as Scopus is. It isn't. But large revisions to citation counts or additions of large new data sources seems to be a thing of the past.

Monday, September 28, 2015

Influential Publications in Ecological Economics Revisited

Back in 2004 I published with Bob Costanza, Chunbo Ma, Brendan Fisher, and Lining He a paper that looked at the most influential publications on the field of ecological economics. Then, about this time last year Gaël Plumecocq contacted me about participating in a session he was organizing at the European Society of Ecological Economics meeting in Leeds on "ecological economics understood as an epistemic community". I had the idea of revisiting our 2004 paper a decade later and seeing how the field had changed in the meantime. Eventually, Gaël also came on board our author team, contributing a textual analysis of the key themes in the influential papers. Gaël gave a presentation on the paper at ESEE and now we finally have a working paper version of our new paper on the web. The full author team includes: Bob Costanza, Rich Howarth, Ida Kubiszewski, Shuang Liu, Chunbo Ma, Gaël, and myself.

We downloaded from the Web of Science (WoS) information on all the papers published in Ecological Economics from 2004 to 2014 including the number of citations each received and the full reference list from all 2960 articles. We define outwardly influential papers as the 10% of articles published in the journal in each year from 2004 to 2014 that received the most citations in the Web of Science. The inwardly influential publications are all publications that received more than 15 citations in the journal in the period 2004-2014. For each of these publications we collected the total number of citations in the Web of Science, Google Scholar, and Ecological Economics. The inward influence data needed a lot of cleaning up, which was mainly done by Chunbo and Ida with my help.

Shuang produced these graphs of inward and outward influence using Tableau:




For inward influence, there are many publications at the bottom right that have been relatively much more influential across science as a whole than they have been in ecological economics. Publications towards the left have been mostly influential in ecological economics alone. By contrast, there is a stronger correlation between citations received in the journal and more broadly for the outwardly influential articles. One outlier is the Pimentel et al. paper on the costs of invasive species that is the most cited article (by WoS cites) ever published in ecological economics.

The theme analysis found, as we expected, that the most influential topic was ecosystem services and payments for ecosystem services, which received 25% of the citations of the influential publications. By contrast, sustainable development and foundations of ecological economics were the most influential topics prior to 2004.

We also followed up on my 2006 paper with Chunbo Ma by looking at the journals, which cite Ecological Economics the most and which are cited by Ecological Economics the most:


There have been quite dramatic changes in these lists with more than half the journals being new entrants. In general there has been an increase in citation links to interdisciplinary environmental science and environmental studies journals and a reduction in links to mainstream economics journals including environmental economics journals. No general interest economics journals are now on the top 20 inward and outward lists.

We think that these trends reflect a maturation of ecological economics as a transdisciplinary field.

Thursday, July 30, 2015

Scopus Adds More Article Level Metrics

Scopus has added a new set of article level metrics. I think the most interesting one is "Field-Weighted Citation Impact" which tells you how cited your article is relative to other similar articles. I think this metric has a big potential in tenure and promotion cases. Here is Scopus' explanation:


Field-weighted Citation Impact (FWCI) 

Field-Weighted Citation Impact is sourced directly from SciVal.


As defined in Snowball Metrics, Recipe Book/Field-Weighted Citation Impact Field-Weighted Citation Impact is the ratio of the total citations actually received by the denominator’s output, and the total citations that would be expected based on the average of the subject field. A Field-Weighted Citation Impact of:
  • *Exactly 1* means that the output performs just as expected for the global average.
  • More *than 1* means that the output is more cited than expected according to the global average. For example, 1.48 means 48% more cited than expected.
  • Less than 1 means that the output is cited less than expected according to the global average.
Field-Weighted Citation Impact takes into account the differences in research behaviour across disciplines. It is particularly useful for a denominator that combines a number of different fields, although it can be applied to any denominator.
  • Researchers working in fields such as medicine and biochemistry typically produce more output with more co-authors and longer reference lists than researchers working in fields such as mathematics and education; this is a reflection of research culture, and not performance.
  • In a denominator comprising multiple disciplines, the effects of outputs in medicine and biochemistry dominate the effects of those in mathematics and education.
  • This means that using non-weighted metrics, an institution that is focused on medicine will appear to perform better than an institution that specialises in social sciences.
  • The methodology of Field-Weighted Citation Impact accounts for these disciplinary differences.

Saturday, July 11, 2015

Papers from Google Scholar

One way that I keep up to date is to track the papers that cite me using Google Scholar alerts. This time I thought some of the papers were more interesting than usual, particularly the economic history papers. Well it's one way to produce a quick blogpost :)

Y Ren, D Parker, G Ren, R Dunn - Climate Dynamics, 2015
Abstract The spatial and temporal pattern of sub-daily temperature change in mainland
China was analysed for the period from 1973 to 2011 using a 3-hourly dataset based on 408
stations. The increase in surface air temperature was more significant by night between ...

H Nielsen
Abstract This paper examines the role of foreign trade in the consumption of primary energy
in the Czech Republic and to what extent adjustment for energy embodied in trade effects
the country's energy intensity curve. As opposed to previous studies, this article takes a ...

G Esenduran, E Kemahlıoglu-Ziya, JM Swaminathan
ABSTRACT In the last two decades, many countries have enacted product take-back
legislation that holds manufacturers responsible for the collection and environmentally
sound treatment of end-of-use products. In an industry regulated by such legislation, we ...

R Hölsgens, B Gales, JP Smits, F Notten
In this paper we analyze recent estimates of annual CO2 (carbon dioxide) emissions from
energy consumption in the Netherlands since 1800 alongside another emission to air
resulting from energy consumption: SO2 (sulfur dioxide). The new time series on CO2 can ...

E Ömer, M BAYRAK - Anemon Muş Alparslan Üniversitesi Sosyal Bilimler …, 2015
Özet Enerji; kullanım şekli, miktarı, bileşimi, yapısı ve mahiyetiyle ekonomik ve sosyal
gelişmişliğin temel ölçütlerinden biridir. Bir ülkede mevcut enerji arzının enerji talebini
karşılayamadığı durum olarak tanımlanan enerji açığı; büyüme ve kalkınma sürecinde, ...

R Hölsgens, C Ducoing, M Rubio, B Gales
Abstract The relationship between energy and capital is one of the most important
relationships of modern economic growth. Machines need energy to produce all the goods
we enjoy; energy without machinery is useless. However, the great majority of the ...

Z Guevaraa, JFD Rodriguesc, T Domingosb
Abstract Conventional energy input-output models were developed about 40 years ago and
have not been significantly improved since. These conventional models offer a limited
description of energy flows in the economy. This paper introduces a novel energy input- ...

M Amoah, O Marfo, M Ohene - Forests, Trees and Livelihoods, 2015
Firewood is the dominant fuel type used by rural households in Ghana. However, the
scarcity of firewood species has raised concerns about the sustainable use of this fuel type.
This study investigated the firewood consumption pattern, firewood species used by rural ...

B Deng, Y Li
Abstract: Efficiency Power Plant (EPP) promotes the use of energy-efficiency power plant
technology and energy efficient equipment, coupled with its low-input, zero pollution, zero
emissions and other advantages, has an important role in the control of energy ...

JD Urrutia, MLT Olfindo, R Tampis
Abstract: The researchers aim to formulate a mathematical model to forecast Exchange Rate of the Philippines from the 1st Quarter of 2015 up to the 4th Quarter of 2020 using
Autoregressive integrated Moving Average (ARIMA). The researchers used the data ...

Wednesday, May 6, 2015

Research Assessment Using Early Citation Information

A new paper with Stephan Bruns on carrying out research assessment like the UK REF and the Australian ERA using citations data rather than peer review. We did a lot of the work of processing the data (doing fancy things with R and manually checking names of universities in Excel) when I visited Stephan in Kassel in November.

The problem with research assessment as carried out in Britain and in the social sciences in Australia is that publications that have already passed through a peer review process are again peer reviewed by the assessment panels. This involves a significant workload for many academics who are supposed to read these papers as well as the effort a each university put into selecting the publications that will be reviewed. However, this second peer review though is inferior to the first. If instead citation based metrics were used the whole process could be done much faster and cheaper. In Australia the natural sciences and psychology are assessed using citation analysis. I think this can be extended to at least some other social sciences including economics.

UK REF panels can also put some weight on citations data in some disciplines including most natural sciences and economics, but only as a positive indicator of academic significance and in very much a secondary role to peer review. This represents a change from the previous RAE, which prohibited the use of citations data by panels. This paper provides additional evidence on the potential effectiveness of citation analysis as a method of research assessment. We hope our results can inform the future development of assessment exercises such as the REF and ERA.

One reason why citations analysis is less accepted in the social sciences than in the natural sciences is the belief that citations accumulate too slowly in most social sciences such as economics to be useful for short-term research assessmen.

My 2014 paper in PLoS ONE shows that long-run citations to articles in economics and political science are fairly predictable from the first few years of citations to those articles. However, research assessment evaluates universities rather than single articles. In this new paper, we show that rank correlations are greatly increased when we aggregate over the economics publications of a university and also when we aggregate publications over time. The rank correlation for UK universities for citations received till the end of 2004 (2005) by economics articles published in 2003 and 2004 with total citations to those articles received through 2014 is 0.91 (0.97). These are high correlations. Correlations for Australia are a bit lower.

Our results here show that at the department or university level citations definitely accumulate fast enough in economics in order to be able to predict longer run citation outcomes of recent publications. It's not true that citations accumulate too slowly in the social sciences to be used in research assessment.

On the other hand, the rank correlation between our early citations indicators and the outcome of research assessment exercises in the UK and Australia ranges from 0.67-0.76. These results suggest that citation analysis is useful for research assessment in economics if the assessor is willing to use cumulative citations as a measure of research strength, though there do appear to be some systematic differences between peer-review based research assessment and our citation analysis, especially in the UK. Part of the difference will emerge due to the differences between the sample of publications we selected to assess and the publications actually selected in the 2010 ERA and 2008 RAE.

Sunday, February 22, 2015

How Has Research Assessment Changed the Structure of Academia?

Does measuring something change it?  In quantum mechanics measurement disturbs what is being measured, which is referred to as the observer effect. The same is often true in social systems, especially of course when measurement is attached to rewards. The UK and Australia have been conducting periodical research assessment exercises - the REF and ERA. In the case of the UK, research assessment started almost three decades ago. In Australia, the first research assessment was only conducted in 2010 but the founding of the ARC in 1988 and its independence in 2001 are both milestones in the road to increased emphasis on competition in research in Australia.

Johnston et al. (2014) show that the total number of economics students has increased in UK more rapidly than the total number of all students, but the number of departments offering economics degrees has declined, particularly in post-1992 universities. Also, the number of universities submitting to the REF under economics has declined sharply with only 3 post-1992 universities submitting in the latest round. This suggests that the REF has driven a concentration of economics research in the more elite universities in the UK. BTW the picture above is of the Hotel Russell, which the Russell Group of British universities is named after.

Neri and Rodgers (2014) investigate whether the increased emphasis on research in Australia has had the desired effect in the field of economics. They investigate the output of top economics research by Australian academics from 2001 to 2010. By constructing a unique database of 26,219 publications in 45 top journals, they compare Australia’s output internationally, determine whether Australia’s output increased, and rank Australian universities based on their output. They find that Australia’s output, in absolute and relative terms, and controlling for differences in page size and journal quality, increased and, on a per capita basis, is converging to the levels of the most research-intensive countries. Finally, they find that the historical dominance of the top four universities is diminishing. The correlation between the number of top 45 journal articles published in 2005-2010 and the ERA 2012 ranking is 0.83 (0.78 for 2003-8 and ERA 2010).

References

Johnston, J., Reeves, A. and Talbot, S. (2014). ‘Has economics become an elite subject for elite UK universities?’ Oxford Review of Education, vol. 40(5), pp. 590-609.

Neri, F. and Rodgers, J. (2014). ‘The contribution of Australian academia to the world’s best economics research: 2001 to 2010’, Economic Record.