Higher Education Reform

Do Journal Rankings Give Short Shrift to the South?

October 31, 2017 1680

The views expressed in this blog post are not necessarily those of Colciencias, a Colombian governmental organization in charge of national science and technology policy.
***

Many research evaluation systems are built around vague notions of excellence. Excellent research is presented as that which advances the frontiers of science. Often, governments reward researchers who can show that their publications are “excellent”. The ultimate aim of this reward system is to support only the best research and researchers and to discourage “poor quality” research.

A common practice in many research evaluation systems is to judge the value of a work based on the journal in which it is published. These assessments are based on the assumption that research published in prestigious (“top”) journals is excellent, and therefore should be rewarded. The DORA declaration and the Leiden Manifesto have warned against this practice, yet journal-based evaluation systems continue to be used in many countries such as Spain, Brazil, Colombia, South Africa, and in influential global rankings such as the Shanghai Jiao Tong University’s or the THE World University Rankings. However, our recent research highlights the problems of relying on journal rankings for research assessment, and seriously questions this practice.

LSE-impact-blog-logo

This article by Diego Chavarro and Ismael Ràfols originally appeared on the LSE Impact of Social Sciences blog as “Research assessments based on journal rankings systematically marginalise knowledge from certain regions and subjects” and is reposted under the Creative Commons license (CC BY 3.0).

One study shows that journal-based research evaluation systems underestimate research that is relevant for important social, economic, and environmental issues, thus marginalizing some types of knowledge. For instance, passion fruit horticulture, oil-palm diseases, specific pathogens that attack red carnation, botanical studies of biodiversity, and Latin American business history are subjects published mainly in journals that score low in journal rankings. Since these topics do not fit journal-based ideas of excellence, they receive little governmental recognition and resources, and are not considered desirable from the perspective of public accountability. However, these issues are essential to the economic, environmental, and social development of regions such as Latin America.

Another study shows that the inclusion of journals in the most prestigious citation databases, the Web of Science (WoS), is not based on objective criteria. Specifically, the study shows that the country, language, and discipline of a journal influence the probability of inclusion, regardless of its editorial quality or scientific impact. For instance, Colombian journals have lower chances of being indexed by WoS compared to Spanish journals with similar scientific impact and editorial quality. These biases are reflected in WoS’s overall coverage of journals, which is focused on those produced in the USA and north and western Europe, on natural sciences, and in English language. In other words, some journals received a more positive assessment than others by WoS due to particularistic reasons such as country or disciplinary subject, which poses questions about the adequacy of using this data source to assess the quality of journals.


This post is based on the authors’ article, “Why researchers publish in non-mainstream journals: Training, knowledge bridging, and gap filling”, published in Research Policy (DOI: 10.1016/j.respol.2017.08.002); and the preprints “To What Extent is Inclusion in the Web of Science an Indicator of Journal ‘Quality’?” available on SSRN; and “Under-reporting research relevant to local needs in the global south. Database biases in the representation of knowledge on rice,” available on SemanticScholar.


Inclusion and exclusion of journals in databases affect the coverage of topics, as illustrated in another case study that explores extent of coverage on rice research in different databases. This study compared the coverage of publications in the databases WoS, Scopus, and CAB Abstracts (CABI). It showed that CABI has a much higher coverage of the issue “rice” (77 percent) than Scopus (60 percent) or WoS (43 percent). In terms of research subjects or topics, WoS and Scopus focus on molecular biology, traditional genetics, and industry-related consumption, whereas CABI focuses more on productivity, plant nutrition, plant characteristics, and plant protection. The foci of WoS and Scopus seem to be related to the research interests of seed companies and food industry, while the foci of CAB are more related to potential interests of local farmers and communities. In this case, research in journals indexed by WoS and Scopus seems to better cover the interests of industrial stakeholders than the interests of small, poorer farmers.

All these studies show the consequences of applying a narrow understanding of excellence to evaluate research regardless of the context, for example in terms of country, discipline, and language. In many countries, evaluation systems are based on one-size-fits-all policy instruments. As a result of this uniformity, these evaluations fail to properly assess research that is socially, environmentally, and economically relevant but not highly cited, or not published in English, or in the more prestigious journals. Evaluations of this type are blind to dimensions of research that are needed to face the challenges posed by sustainable development, such as eradicating poverty, mitigating the effects of climate change, and achieving peace and justice.

Although the evidence presented in this post is focused on journal-based evaluations, the insights also apply to other types of appraisal. For instance, there is an increasing interest in indicators derived from social media – such as tweets, Mendeley reads, etc. – as alternatives to citation counts. Although these are well-intentioned attempts to offer alternative indicators, they are not derived from an alternative understanding of academic knowledge production and communication. In the end, these indicators are based on the same “counting” concept of excellence as reputation or popularity (“the more, the better”), which makes them prone to the same pitfalls as citations-based rankings and assessments (hierarchies, concentration of recognition, marginalization).

A genuinely alternative understanding of academic knowledge production as an endeavour to benefit people and the environment should motivate different ways to conduct research assessment. Alternative research evaluations should be sensitive to context and account for diversity in scientific production (related to geography, language, discipline, and others) helping to overcome the deficiencies of conventional evaluation practice.

For the reasons above, we believe that more inclusive research assessments are needed to overcome ongoing marginalization of some peoples, languages, and disciplines in evaluation. Some efforts, such as the Dutch “new standard evaluation protocol” for the social sciences and the “Norwegian model” of research assessment that is sensitive to regional journals, have moved in this direction – trying to account for the value of various disciplines and local research. However, a creative and radical reform in research evaluation from research councils and universities is needed in many other countries that follow traditional quantitative methods.

Facing the challenges posed by sustainable development requires us to transform the concept of excellence from an elitist view, defined at a distance from society, to a more community-oriented, inclusive view, which encourages engagement. Such an understanding could produce research evaluations that support socially robust knowledge and are sensitive to the relevance of many diverse types of knowledge that are currently underestimated.


Diego Chavarro is an adviser to Colciencias, a Colombian governmental organisation in charge of national science and technology policy. His research focuses on research evaluation, knowledge production, and the contribution of science, technology, and innovation to sustainable development. Ismael Ràfols is a research fellow at Ingenio, a science policy institute at the Universitat Politècnica de València. He is also visiting professor at the Centre for Science and Technnology Studies at the University of Leiden. His research focuses on plural and inclusive use of S&T indicators for informing evaluation, funding and research strategies.

View all posts by Diego Chavarro and Ismael Ràfols

Related Articles

Gazan Publisher, Late Ukrainian Writer Receive Publisher Group’s Prix Voltaire Award
Bookshelf
December 6, 2024

Gazan Publisher, Late Ukrainian Writer Receive Publisher Group’s Prix Voltaire Award

Read Now
From the University to the Edu-Factory: Understanding the Crisis of Higher Education
Industry
November 25, 2024

From the University to the Edu-Factory: Understanding the Crisis of Higher Education

Read Now
Canada’s Storytellers Challenge Seeks Compelling Narratives About Student Research
Communication
November 21, 2024

Canada’s Storytellers Challenge Seeks Compelling Narratives About Student Research

Read Now
Ninth Edition of ‘The Evidence’: Tackling the Gender Pay Gap 
Communication
October 31, 2024

Ninth Edition of ‘The Evidence’: Tackling the Gender Pay Gap 

Read Now
The Conversation Podcast Series Examines Class in British Politics

The Conversation Podcast Series Examines Class in British Politics

Even in the 21st century, social class is a part of being British. We talk of living in a post-class era but, […]

Read Now
Emerson College Pollsters Explain How Pollsters Do What They Do

Emerson College Pollsters Explain How Pollsters Do What They Do

As the U.S. presidential election approaches, news reports and social media feeds are increasingly filled with data from public opinion polls. How […]

Read Now
Diving Into OSTP’s ‘Blueprint’ for Using Social and Behavioral Science in Policy

Diving Into OSTP’s ‘Blueprint’ for Using Social and Behavioral Science in Policy

Just in time for this past summer’s reading list, in May 2024 the White House Office of Science and Technology Policy (technically, […]

Read Now
0 0 votes
Article Rating
Subscribe
Notify of
guest

This site uses Akismet to reduce spam. Learn how your comment data is processed.

0 Comments
Newest
Oldest Most Voted
Inline Feedbacks
View all comments