Higher Education Reform

Playing the Game of REF

December 19, 2014 1215

Research assessment is only partly reliable as an indicator of the real quality of the work going on in higher education. It has a dual character. On one hand it is rooted in material facts and objective methods. Strong research quality and quantity should be and are rewarded in the UK Research Excellence Framework (REF), the results of which have just been published.

But the outcome is also shaped by the universities that select and fashion data for competitive purposes and the subject area panels that define research judged to be outstanding on a global scale.

The Conversation logo

This article by Simon Marginson originally appeared at The Conversation, a Social Science Space partner site, under the title “Game-playing of the REF makes it an incomplete census.” In turn, this is an extract of an article published on the IOE London blog

Total research activity can never be fully captured in performance data. Some things, such as citations in top journals, are easier to measure than others, such as the long-term impacts of research on policy and professional practice. Experienced players are best at gaming the system in their own interest.

A very strong overall REF performance signifies a large concentration of outstanding work. It is an unambiguous plus. All the same, precise league table positions in the REF, indicator by indicator, should be taken with a grain of salt.

Measuring ‘impact’

In the REF, the indicators for “impact”, which are new to the 2014 assessment, are the least objectively grounded and most vulnerable to manipulation. This is because of the intrinsic difficulty of measuring the changes to society, economy and policy induced by new knowledge, especially in the long-term, and because of the kind of crafted “impact-related” data that is collected during the REF assessment process. A sophisticated industry has already emerged to manufacture examples of the relevant “evidence” of impact.

At best, this gets everyone thinking about real connections with the users of research, which is one – though only one – of the starting points when producing the impact documentation. At worst, it leads to data that bears as much relation to reality as Soviet-era statements of output by Russian factories in response to government targets.

Inevitably, the universities most experienced and adept at managing their response to performance measures of all kinds will perform especially well in demonstrating proof of impact. There is also a “halo” effect, of the kind that affects all measures contaminated by prior reputation.

The REF indicators that are the most meaningful are those related to “output” quality, such as the grade-point average (GPA) of each university, and the proportion of researchers ranked as “world-leading”. These are grounded in considered judgements of real research work, by panels with significant expertise.

Is it getting better and better?

Yet the value of the output indicators in the REF, which include publication numbers, as measures of comparative quality, are subject to two caveats.

First, between the previous Research Assessment Exercise (RAE) in 2008 and the 2014 REF there has been a notable inflation of the proportion of UK research outputs judged to be “world-leading” (rated four-star) and “internationally excellent” (rated three-star).

In 2008, just 14 percent of research outputs were judged to be four-star and 37 percent were judged to be three-star, a total of 51 percent in the top two categories. In 2014, the proportion of the work judged to be outstanding had somehow jumped to 72 percent – 22 percent was judged to be four-star and another 50 percent judged to be three-star. This phenomenal improvement happened at a time when resources in UK higher education were constrained by historical standards.

While genuine improvement no doubt has occurred in at least some fields, the scale and speed of this improvement beggars belief. It reflects a combination of factors that generate boosterism. Higher education institutions have a vested interest in maximising their apparent quality. Subject area panels have a vested interest in maximising the world-class character for their fields. And UK higher education and its institutions are competing with other nations, especially the United States, for research rankings, doctoral students and offshore income.

The inflation of three and four-star research is a worrying sign of a system in danger of becoming too complacent about its own self-defined excellence. This is not the way to drive long-term improvement in UK research. Less hubris and more hard-nosed Chinese-style realism would produce better outcomes.

It would be better to rely less on self-regulation, enhance the role of external international assessors in judgements about what constitutes “world-leading” research and spotlight areas where improvement is most needed, rather than focusing attention solely on the areas where research is very strong.

The selectivity game

The second caveat is that universities can readily game the assessment of output quality, by being highly selective about whose work they include in the assessment. Including only the best researchers pushes up the average GPA and the proportion of research ranked as four-star. Those institutions that do this pay a financial price, in that their apparent volume of research is reduced – and their subsequent funding will fall. Nevertheless, it is good for reputation. That has many long-term spin-offs, including financial benefits.

While some universities have chosen to approach the REF on an inclusive basis, others have pursued highly tailored entries designed to maximise average output quality and impact. Just one example: Cardiff sharply reduced its number of full-time equivalent staff, from 1,030 in the 2008 RAE to only 738 in the 2014 REF, according to analysis by Times Higher Education. This lifted Cardiff’s quality rating, the grade-point average of its outputs, to sixth in the country, though in terms of the volume of high-quality research it appeared to fall from 15th in the UK to 18th in the Times Higher Education’s ranking.

As universities do not have to enter all the eligible staff for the REF, the data is an incomplete census of all research activity and does not compare like-with-like. In each field of research, the measures of performance compare universities that enter 80%-100% of their staff in that field, with universities that enter only 10 percent-20 percent of the eligible staff, rendering meaningless any comparison of average quality. This undermines the validity of the REF as a league table of system performance, though everyone treats it that way.

The trend to greater selectivity manifest in some, but not all, higher education institutions is no doubt one of the factors that has inflated the incidence of four-star and three-star rated research.The Conversation


Simon Marginson is professor of international higher education at the Institute of Education, University of London and a professorial associate of the Centre for the Study of Higher Education at the University of Melbourne. A PhD graduate from the University of Melbourne (1996) when he was awarded the Chancellor’s Prize for excellence in the doctoral thesis. Marginson’s work focuses on higher education systems and policy, especially international and global aspects of higher education. He was designated an Australian Research Council Australian Professorial Fellow in 2003. His current projects are focused on the global strategies of research universities in the Asia-Pacific (a comparison across 18 countries) and on relations between public and private sector research organizations in the knowledge economy (a comparison of Australia, the Netherlands and Korea), and he is also preparing a monograph on globalization and higher education.

View all posts by Simon Marginson

Related Articles

Canada’s Storytellers Challenge Seeks Compelling Narratives About Student Research
Communication
November 21, 2024

Canada’s Storytellers Challenge Seeks Compelling Narratives About Student Research

Read Now
Tom Burns, 1959-2024: A Pioneer in Learning Development 
Impact
November 5, 2024

Tom Burns, 1959-2024: A Pioneer in Learning Development 

Read Now
Alondra Nelson Named to U.S. National Science Board
Announcements
October 18, 2024

Alondra Nelson Named to U.S. National Science Board

Read Now
Lee Miller: Ethics, photography and ethnography
News
September 30, 2024

Lee Miller: Ethics, photography and ethnography

Read Now
‘Settler Colonialism’ and the Promised Land

‘Settler Colonialism’ and the Promised Land

The term ‘settler colonialism’ was coined by an Australian historian in the 1960s to describe the occupation of a territory with a […]

Read Now
Research Assessment, Scientometrics, and Qualitative v. Quantitative Measures

Research Assessment, Scientometrics, and Qualitative v. Quantitative Measures

The creation of the Coalition for Advancing Research Assessment (CoARA) has led to a heated debate on the balance between peer review and evaluative metrics in research assessment regimes. Luciana Balboa, Elizabeth Gadd, Eva Mendez, Janne Pölönen, Karen Stroobants, Erzsebet Toth Cithra and the CoARA Steering Board address these arguments and state CoARA’s commitment to finding ways in which peer review and bibliometrics can be used together responsibly.

Read Now
Paper to Advance Debate on Dual-Process Theories Genuinely Advanced Debate

Paper to Advance Debate on Dual-Process Theories Genuinely Advanced Debate

Sage 987 Impact

Psychologists Jonathan St. B. T. Evans and Keith E. Stanovich have a history of publishing important research papers that resonate for years.

Read Now
0 0 votes
Article Rating
Subscribe
Notify of
guest

This site uses Akismet to reduce spam. Learn how your comment data is processed.

0 Comments
Newest
Oldest Most Voted
Inline Feedbacks
View all comments