探花视频

Do national research assessment exercises still pass peer review?

As an international review of the UK鈥檚 REF begins even before the assessment panels have done their work, has the exercise鈥檚 reliance on rereading published papers finally had its day? Might it be time for metrics? Or something else entirely? Jack Grove looks around the world for options  

Published on
September 2, 2021
Last updated
September 2, 2021
A man with a microscope for a head as a metaphor for assessing research assessment
Source: Alfred Gescheidt/Getty

鈥淚t鈥檚 bureaucratic insanity with predictable results 鈥 the proliferation of journal articles that no one reads, science that is not reproducible and millions of pounds spent on pointless assessment.鈥

That is the verdict of Julia Lane on what she regards as the 鈥渃omplete madness鈥 of the UK鈥檚 research excellence framework (REF). 鈥淚t also does little to convince politicians to hand over more money for research,鈥 adds Lane, a labour economist at New York University, who was herself asked by President Obama to make the case for higher science spending, resulting in the Star metrics programme.

Of course, questions about the REF鈥檚 utility and cost, estimated at almost 拢250 million for the 2014 exercise, have been raging for years. But while they have been endlessly scrutinised by numerous reviews 鈥 most recently, the by Lord Stern of Brentford 鈥撀爐he REF鈥檚 critics claim that the resulting changes have amounted to little more than tinkering round the edges of a process used to distribute more than 拢1 billion a year in research block grants.聽With the submissions phase of the 2021 exercise now over and a major international review under way, many observers expect a far more radical rethink this time around.

Even those in charge of UK science would seem to have little love for the REF, at least in its current incarnation.聽Four years ago, Dame Ottoline Leyser, chief executive of UK Research and Innovation聽鈥撀爓hose Research England subdivision runs the REF聽鈥撀燼rgued that the exercise鈥檚 focus on individuals聽. Instead, universities should be required to submit a set number of 鈥渙utputs鈥 to each subpanel, regardless of who wrote them, based on the number of their academics who work in that particular research area. This reform, argued Leyser, in her previous capacity as the Royal Society鈥檚 policy lead, would remove the obligation universities feel to only recruit applicants 鈥渨ith the 鈥榬ight鈥 sort of outputs鈥, allowing them to take a broader view of scientific contribution.

探花视频

ADVERTISEMENT

Ministers and Whitehall mandarins, meanwhile, are likely to be more anxious than ever for cost savings in the wake of the pandemic, so the REF鈥檚 historical reliance on peer review could be replaced, at least in part, by bibliometrics. The appointment of an international advisory panel, rather than a UK insider, to review the exercise further opens the door to wholesale change.

That said, an international consensus on whether and how research evaluation should be carried out is far from apparent. The UK is not the only country that has been taking a hard look at whether its own framework is fit for purpose. In June, the Australian Research Council agreed to implement recommendations from a wide-ranging review of its Excellence in Research for Australia (ERA) exercise, ahead of the next iteration in 2023. Changes include a recalibration of its 鈥渨orld standard鈥 rating, after 90 per cent of Australian research in the most recent exercise in 2018 met or exceeded that standard: a 鈥渟carcely believable鈥 figure that critics said had devalued the exercise. ERA's dual approach of assessing science via citations and humanities by peer review is also claimed to聽unfairly penalise聽the latter.

探花视频

ADVERTISEMENT

In New Zealand, doubts have been raised over the transaction costs of the individual-focused, peer review-driven聽 (PBRF). Since its inception in 2002, the exercise has awarded at least half of its funding to two universities 鈥 Auckland and Otago, leaving others scrapping over what is left of the NZ$315 million (拢160 million) associated funding. Amid concerns that the exercise also undervalues locally focused research, a of the exercise was conducted in 2020, although the resulting changes, announced in July, were modest.

In Italy, meanwhile, some scholars say the bibliometrics-centred approach of its own exercise, introduced in 2010,聽has encouraged cheating by means of 鈥渃itation exchange clubs鈥, leading to a more insular academy and the mere illusion of improvement.

Nevertheless, no other country has sought to copy the REF. 鈥淲e discussed different types of research assessment for a whole day in Copenhagen 鈥 not one person said they wanted a REF-like system,鈥 recalls Gunnar Sivertsen, research professor at the Nordic Institute for Studies in Innovation, Research and Education in Oslo, Norway, on his involvement in Denmark鈥檚 talks on research assessment two years ago. Sweden 鈥 where, as in the Netherlands, each institution runs an internal research evaluation with the help of international experts 鈥 also rejected a UK-inspired evaluation system back in 2016, he adds.

Institutional self-evaluation, rewarded with government funding, is, for Sivertsen, a better way to encourage universities to pursue excellence in their respective strengths. 鈥淎ll countries make a distinction between research-intensive and ordinary universities and allocate research funding accordingly 鈥 do you really need a REF to justify how you do it?鈥 he asks.

A man and a woman stretch out their arms toward a red computerised grid
厂辞耻谤肠别:听
Alfred Gescheidt/Getty

This is one of the big questions that UKRI鈥檚 eight-strong , containing figures from industry, academia and overseas research bodies, is beginning to address. Its chair, New Zealand鈥檚 former chief scientist Sir Peter Gluckman, tells 探花视频 that the panel will not shy away from 鈥減rovocative鈥 questions about the REF鈥檚 future. 鈥淚 am not someone who is afraid to say the emperor has no clothes,鈥 says Gluckman, who believes that one of the big strengths of the REF 鈥 known in its pre-impact incarnation as the research assessment exercise 鈥 is that 鈥渋t has never been static and has always been open to frank review鈥.

However, those hoping for a complete overhaul may be disappointed, he suggests. 鈥淲e shouldn鈥檛 throw out the baby with the bathwater or ignore what has been achieved through the REF and other research practices, namely turning Britain into a science superpower under some difficult circumstances,鈥 he says. 鈥淭here are, of course, some logistical issues 鈥 some people say it is very bureaucratic, though some would argue that this comes from universities rather than the state itself.鈥

Gluckman鈥檚 panel has yet to get into the weeds of the framework and has instead focused on fundamental questions that will inform their thinking. One key issue is 鈥渨hat does excellence mean?鈥, he explains. 鈥淵ou could have a team that has a hypothesis that is very interesting and does a brilliant piece of work, but the results are disappointing and appear in a minor research journal. Another team might do some rather less impressive research that produces an outcome of fundamental importance that is published in Nature.鈥 A world-class system should have space for both outcomes because having a 鈥渃ulture that promotes excellent research activity, even when the results are not that exciting鈥 is vital, he says.

Leyser recently repeated her concern that despite this REF iteration鈥檚 reduction in the required number of outputs per submitted researcher from four to one, the exercise remains too focused on individuals. Echoing such sentiments, the review panel will also consider how team science can be more effectively supported, says Gluckman. 鈥淭here is no doubt that the nature of enquiry has changed,鈥 he explains. 鈥淲hile I accept there is a big difference between how genomics and humanities researchers operate, there is a much greater sense that, for many subjects, team-based, transdisciplinary research is becoming much more impactful when it comes to solving global and national problems. The Stern report recognised this problem in 2016 but I鈥檇 argue the world has moved on a lot since then.鈥

探花视频

ADVERTISEMENT

Gluckman emphasises that while there is no 鈥減erfect system鈥, whichever one is adopted 鈥渕ust be seen to be fair鈥 given that the future of departments and livelihoods are at stake. 鈥淎nd you cannot have that without transparency,鈥 he adds. 鈥淣o one would say it was fair if five experts came in from around the world and decided to give this much to Oxford and Cambridge, or this much to Bristol.鈥

Some might also see that comment as a coded signal that the primary role of peer review by UK academics will be retained, given the controversies that continue to swirl around the obvious and frequently touted alternative: metrics.

At present, some 400 subject experts, assessors and specialist advisers across four main panels and 34 subpanels are of outputs submitted by universities before the 31 March submission deadline 鈥 a 10-month effort that, in the 2014 iteration, was estimated to have cost 拢19 million in panellists鈥 time. A far larger sum 鈥 拢212 million, or about 拢4,000 per submitted researcher 鈥 was spent by universities collectively on the submission process, while 拢55 million was spent preparing some 7,000 impact statements, according to official estimates.

Adam Tickell, vice-chancellor of the University of Sussex, is leading a government-commissioned review of research bureaucracy that is due to report its interim findings in the autumn. Tickell, who will take over at the University of Birmingham in January, is aiming for 鈥渂old rather than timid鈥 recommendations on cutting red tape, and he concedes that 鈥渢he REF creates quite a burden for institutions in the run-up to the deadline鈥. However, the exercise comes around only every six or seven years, and 鈥渋f you compare the REF to individual [grant funding] rounds, it is much more efficient and much less burdensome,鈥 he argues.

But others fear that the official 拢250 million cost of the REF may be a considerable underestimate, while a significant number of academics insist that citation metrics would produce at a fraction of the cost.

鈥淚 just can鈥檛 see the added value of reassessing publications that have already been assessed by two to five experts in the field,鈥 says Anne-Wil Harzing, professor of international management at Middlesex University, who was able to create a REF ranking virtually identical to the 2014 outcomes over the course of two hours on a in 2017.

Nor is she the only academic to claim that metrics offer a much cheaper, quicker and more up-to-date snapshot of research quality that leads to similar distributions of research funding. But the landmark review of metrics carried out by James Wilsdon, Digital Science professor of research policy at the University of Sheffield,聽in the wake of the 2014 REF聽came down against any significant expansion in the use of bibliometrics聽because of widespread scepticism about their reliability and a sense that聽peer review remains the 鈥済old standard鈥 of research assessment.

However, the case for a 鈥渕ore streamlined, metrics-driven system that isn鈥檛 a nightmare to administer鈥 has become increasingly compelling since then, says Steve Fuller,聽Auguste Comte professor of social epistemology at the聽University of Warwick.聽For instance, the REF鈥檚 system of disciplinary subpanels doesn鈥檛 encourage cross-disciplinary research, Fuller contends. 鈥淭here is also the fact that each iteration makes less and less difference to the outcomes 鈥 things do not change that much so the return on investment is diminishing,鈥 he adds.

探花视频

ADVERTISEMENT

More broadly, Fuller wonders if the REF may have outlived its initial purpose when it was conceived in the 1980s to bring some accountability to a research sector where institutional laxity let some scholars produce almost nothing while outstanding young minds were forced to leave for the US. 鈥淭here was a lot of concern about Britain鈥檚 brain drain and general levels of research productivity, so the REF [known then as the research selectivity exercise] was partly about bringing in new people to create a research culture that was lacking in Britain,鈥 recalls Fuller, who himself arrived in the UK from the US in 1994 as part of a REF hiring round.

If team science is the new priority then an assessment involving bibliometrics would be a better way to do it, Fuller contends. 鈥淚f you look at those people with enormous h-indexes, it鈥檚 usually because they are working in teams 鈥 their citations reflect how they work across a number of groups,鈥 he says.

But framing the future of the REF as a face-off between bibliometrics and peer review is a trap that has hampered previous reviews, says Sivertsen, an adviser to several national research evaluation reviews across Scandinavia, who will give evidence to the Gluckman panel. 鈥淎cademics will always choose peer review, so the changes made are always limited,鈥 he says.

A man with a clockwork mechanism superimposed over his face
厂辞耻谤肠别:听
Alfred Gescheidt/Getty

Could the inclusion of non-bibliometric indicators offer a better way forward? Under New Zealand鈥檚 PBRF, 60 per cent of funds are allocated based on peer review assessments of individual researchers鈥 portfolios, but the remaining 40 per cent rides on institutional-level evaluations of the number of research degrees completed (25 per cent) and the amount of research income received from external sources (15 per cent).

鈥淥ne reason for having universities is that they generate the high-level conceptual thinking that a workforce needs, so research degree completions recognises this,鈥 says Roger Smyth, an independent tertiary education consultant and the former head of tertiary education policy in New Zealand鈥檚 Ministry of Education. 鈥淩esearch income from the private sector also gets a higher weighting, as does income generated from overseas,鈥 he adds, because this incentivises collaboration with industry and international institutions.

One frequent complaint about the PBRF from academic unions is the 鈥渆laborate CVs鈥 and evidence portfolios that each researcher must compile for submission, admits Smyth, who was involved in the exercise鈥檚 creation in the early 2000s, when he was a government adviser. Could a more team-based approach, as suggested by Leyser, solve such concerns? 鈥淢any people believe this would be a good thing, but the problem, as the latest review panel concluded, is that there is much more potential for game-playing,鈥 he argues. Universities might, for example, seek to hide their weaker research scientists by recording them in the history department, thereby skewing results, suggests Smyth.

That said, the impact of game-playing matters less in New Zealand, which distributes a far lower proportion of research income via its PBRF than the REF, says Smyth; Auckland鈥檚 annual budget of NZ$1.1 billion is almost four times the size of the entire national PBRF budget, for instance. 鈥淭here has always been anxiety about a British system because the financial consequences are so high, including the threat of closing some departments down and moving them to service teaching if they don鈥檛 perform,鈥 he explains. 鈥淭hat鈥檚 less of a risk with individual assessment, where you might have one or two excellent researchers of value to an institution, though their department isn鈥檛 outstanding.鈥

What about the Australian model? There has been "a lot of back and forth" in policy terms between UK and Australia over the years, 鈥渂ut they will now need to look at very different things鈥, says Gemma Derrick, senior lecturer in higher education at Lancaster University, who was part of the advisory board for the ERA review. Derrick predicts that the UK鈥檚 review will seek to broaden the variety of outputs that can be submitted聽鈥 adding software to the list, for instance聽鈥 and expand the REF鈥檚 scope to聽include聽research-related staff currently not eligible for submission. 鈥淟ooking beyond authorship to the wider research community in universities is a good idea but will also lead to more administrative paperwork and increase the REF鈥檚 burden,鈥 she notes.

Paul Wellings, who recently retired after a decade as vice-chancellor of the University of Wollongong, having previously led Lancaster for nine years, also sees different challenges for the REF and ERA, given their聽divergent emphases. 鈥淭he REF has focused on quality by building up islands of excellence and has been a strong driver of reputation, while the ERA has never been used for reputational advantage 鈥 it assesses research, but its impact is modest,鈥 he says.

For Wellings, the growing number of research outputs may pose a headache even for future ERAs, even though where citation analysis is not commonly used. 鈥淚鈥檓 not sure the Australian model is sustainable鈥ith the volume of research outputs growing 9 per cent every year,鈥 he says. Moreover, the massive loss of overseas fee income that has propped up Australian research activity could precipitate a rethink of the ERA鈥檚 stated role as a . 鈥淚t has become a managerial exercise that will be severely challenged over the next few years,鈥 he says.

With the Covid pandemic intensifying interest in the impact of universities鈥 research, Wellings also wonders whether UK policymakers will be quite so interested in a framework largely unknown to the public whose interest lies mainly in the institutional rankings it generates. 鈥淕overnments are starting to ask how to reset the economy and will wonder if knowledge exchange is more important than just measuring excellence in research,鈥 he says. In this respect, the nascent knowledge exchange framework (KEF) and Australia鈥檚 鈥渕ore conservative鈥 efforts in this area 鈥渕ay become a challenge to the REF and ERA鈥, he says.

Of course, the REF also measures research impact. But in pursuit of economic growth, governments seem increasingly willing to direct research funding towards priority areas. A case in point is the UK鈥檚 recently unveiled innovation strategy, which prioritises seven 鈥渟trategic technologies鈥, following up on former science minister David Willetts鈥 鈥eight great technologies鈥 initiative, launched in 2013. Such emphases could undermine the case for an assessment that is still primarily based on the excellence of publications and other 鈥渙utputs鈥, according to Wellings (60 per cent of REF 2021 scores depend on outputs, versus 25 per cent for impact and 15 per cent for research environment).

An increased emphasis on teaching 鈥 reflected in another recent UK initiative, the teaching excellence framework 鈥 could also undermine the REF and ERA, Wellings thinks. He notes that some universities seem willing to discard subject areas where they excel in the REF if undergraduate recruitment is weak or simply no longer aligns with their more vocational focus: 鈥淲e鈥檝e seen this in Australia where some institutions are happy to have some areas which are vocational and teaching-only, rather than having a strong research culture and attempting to drive their reputation through research,鈥 he says. 鈥淚t feels like one of those moments in history鈥hen institutions are having a larger reset as society asks if they are still serving its needs.鈥 This confluence of factors may be 鈥渇airly lethal to the REF鈥.

So what about Lane鈥檚 approach to tracking the impact of research spending?聽The country鈥檚 diverse sector, with its variety of missions and funding sources, has not lent itself to a national research assessment exercise. But then along came the Star metrics programme (鈥淪cience and Technology for America's Reinvestment: Measuring the Effects of Research on Innovation, Competitiveness, and Science鈥), a partnership between the federal government and research institutions that mined existing datasets to examine the effect of university funding not only on regional employment and corporate prosperity but also health and other socio-economic factors.

Lane and various collaborators then followed up with 聽(Universities: Measuring the Impacts of Research on Innovation, Competitiveness and Science), which takes efforts to track the impact of research spending even further. Lane鈥檚 approach goes far beyond the REF鈥檚 focus on the direct effects of the research itself, and, she argues, does a far better job than the REF of documenting the full impact of the national research budget. Umetrics, for instance, takes account of how many people鈥檚 salaries depend on the country鈥檚 $100 billion annual research budget (about 720,000). Its latest report found that nearly 31,000 PhD students who were funded by research grants went on to get jobs in the US, while, in their first three years after leaving university, 64 per cent of research-trained graduate students were employed in the private sector, where they made an average of $94,000 per year.

鈥淚deas are transmitted through people, not publications, so any evaluation of whether research funding is successful or not should centre on people,鈥 she says 鈥 hence her consistently harsh criticism of the REF.

Whether the UK will go in such a direction is open to question, particularly given that the KEF covers some of the elements that Lane focuses on. Moreover, while the review panel contains figures from Canada, Australia, the Republic of Ireland, South Africa and , as well as New Zealand, there is no US representative.

Gluckman promises that 鈥渁s an expert panel, all we can do is be honest, critical and provocative鈥. But he acknowledges that however honest, critical and provocative his report may be, its implementation will be entirely hostage to political fortune.聽Still, bold changes may still be ahead. The UK government鈥檚 recent , for instance, states that 鈥渁cross the whole sector, there is a strong business case for increasing the diversity of people and ideas and for working in partnership to drive progress based on what works鈥. And 鈥渇rameworks, assessment and incentives at an institutional level鈥 should encourage this.

探花视频

ADVERTISEMENT

鈥淚t is up to the UK government to determine what is done,鈥 Gluckman reflects. But if history is any guide, the UK academic community will have a great deal to say on the matter, too. And if the verdict is largely negative, it will take a determined government to press on regardless.

jack.grove@timeshighereducation.com

POSTSCRIPT:

Print headline:聽Assessing research assessment

Register to continue

Why register?

  • Registration is free and only takes a moment
  • Once registered, you can read 3 articles a month
  • Sign up for our newsletter
Please
or
to read this article.

Related articles

Reader's comments (4)

Very good article...I would suggest that going forward all higher education research must be problem solving. This means that all faculties (engineering, law, science and technology,, social sciences, creative arts and humanities, business studues, sports, medicine) must investigate issues in their respective disciplines that are pertinent to the resolution/reduction of a problem in their community/society. For example, social science faculty should investigate any one of the following issues with a view to reducing same: poverty in a community, child poverty or any other aspect of poverty, unemployment, juvenile delinquency, any form of crime, workplace productivity, staff performance appraisal or any other. Science/engineering/technology staff can investigate transitioning efforts to green alternative energy or any other engineering, scientific or technological issue that is currently unresolved. Creative Arts and humanities students/staff can investigate the impact of direct arts involvement in the development of emotional intelligence and compare such results with the impact of a lack of direct involvement. All research funding, assessments (by THES world university rankings, Athena Swan and others) and incentives going forward should be hinged or tied to this requirement for problem solving. In addition all such research should also result in the development and implementation of a problem solving project proposal to be undertaken by academic staff and collaborators who conducted such research. Then we shall see visible benefits to communities, societies and government and other funding agencies will also see tangible results being derived from their research investments/ contributions. Undergraduate and graduate teaching can then be tied to real world issues and students would then see the subject matter of their various courses coming alive. This approach renders it easier to establish strong linkages between research, teaching and community outreach.
With so much money involved, we need to be able to evaluate the Return on Investment. (ROI). In the absence of Dominic Cummings, I wonder if there will still be a higher allocation of funds to the business sector rather than money going mainly to academics. What can we learn from the way Covid 19 has been handled? If the problem that needed solving was "How can we stop millions of people being killed off and the answer was vaccines, does a partnership approach involving Government, University Research and Business seem the right way forward?
The REF is the most useless thing that was ever designed and foisted upon the academic community. It has encouraged game playing on a massive scale that has benefitted many rent seekers both in managerial positions and in academia. Look at the countries around the world whose universities are dominating research both theoretical and applied, that should tell anyone with half a brain cell that you don't need the REF to be making an impact for the benefit of all humankind. In one discipline many new comers are lining up to attract academics from the US to somehow get their papers into particular journals. This is what it has been reduced to. Not to mention the fractional contract foreign academics who are being funded by the tax payers of this country!
If you really want to know about the research environment. As part of the REF, among other things conduct an anonymous survey of the academics and other services directly involved in research. That will tell you more than any massaged metric ever will.

Sponsored

Featured jobs

See all jobs
ADVERTISEMENT