As preparations for the next research excellence framework gather momentum, I fear I鈥檓 in a dwindling minority who finds it discombobulating to hear colleagues confidently asserting that they have a handful of 3* or 4* papers, and looking down upon those with only 2*s.
Colleagues, the difference between a 2*, 3* and 4* is based on subjective judgements of vague criteria. My articles have been ranked internally as 3* or 4*, so this isn鈥檛 sour grapes, but there really is not a knowable distinction between 鈥 as Main Panel C puts it 鈥 鈥渜uality that is internationally excellent鈥 (3*) and 鈥渜uality that is recognised internationally鈥 (2*). The latter implies the former.
Nor is there a meaningful difference between research that is a 鈥渕ajor influence鈥 (4*) and research that is 鈥渓ikely to have a lasting influence鈥 (3*). Ditto. And even if there were, the judgement is going to be made by a REF panellist who is unlikely to be an expert in your sub-field and who will probably skim-read your article over breakfast on a flight.
That, of course, has always been the problem with the REF. But back in 1986, when the first forerunner exercise was run, it was excusable to think that the only way of splitting research funding between universities on the basis of some sort of demonstrable merit was through some bureaucratic process of this sort, however imperfect.
糖心Vlog
But, hey, times have moved on. There are now numerous internet-based data sources on academics' research performance 鈥 and many of them are free. Senior academics have used some of these to come up with rankings of universities and departments that are extremely closely correlated to those produced from the REF.
For instance, in 2017, Anne-Wil Harzing, professor of international management at Middlesex University, found only small differences between the REF rankings and those created using data from Microsoft Academic. Memorably, doing so took her just 鈥溾. And, in 2013, Dorothy Bishop, professor of developmental neuropsychology at the University of Oxford, the data from the REF鈥檚 precursor, the research assessment exercise, and found that departmental h-indices in psychology predicted the results 鈥渞emarkably well鈥. She suggested this may be true more broadly, too.
糖心Vlog
Meanwhile, Marcus Munafo, professor of biological psychology at the University of Bristol,听 that a 鈥減rediction market鈥 closely mirrored REF outcomes for chemistry departments. Prediction markets arrive at the probability of an outcome occurring based on individuals betting on what they believe the outcome will be.
I鈥檝e also had a go, with my colleague Barrie Craven. We found that university rankings compiled using ResearchGate, Google Scholar and Webometrics (which creates scores based on 鈥渓ink analysis鈥, looking at each university鈥檚 presence and impact on the web) were, again, extremely closely correlated with REF rankings compiled by both 糖心Vlog (based on quality) and Research Fortnight (based on quality and volume).
Importantly, two of the approaches described, prediction markets and Webometrics, have nothing to do with citation indices. These sometimes get a bad press from those, especially in the humanities, who are reluctant to let go of the REF; in the sciences, by contrast, the case is more accepted that citation by colleagues who presumably are experts in the relevant field is a better mark of quality than the approval of stressed, non-expert REF panellists. Either way, it is hard to endorse the conclusion of the 2015 government-sponsored that subjective judgement based on ambiguous criteria remains 鈥渢he least worst form of academic governance we have鈥 in the 21st聽century.
Let鈥檚 spell this out. The REF delivers data extremely slowly and infrequently, at great expense (the official estimate is 拢246 million) and with huge disruption to university life, resulting in rankings very similar but, arguably, inferior to those obtained simply and cheaply using a range of methods that don鈥檛 disrupt anyone.
糖心Vlog
Clearly the government is not going to replace the REF any time soon. An elephantine beast like this develops a life and purpose of its own, and loyalty to match. But there is a clear market opportunity for a sympathetic thinktank to create parallel league tables using the alternative, freely available resources. Because there are many of these, it would be easy to experiment to find an optimum combination of data that can鈥檛 be gamed and that offers no perverse incentives. A handful of supervised interns could easily handle it.
Regularly updated tables will be much more attractive to consumers of higher education 鈥 students and funders 鈥 than the quickly stale and out-of-date REF rankings. Hence, as Friedrich Engels might have put it, the demand for the REF will wither away. The interference of state power in research excellence will become superfluous. Universities will cease to see the need to participate. And, with that, the minister鈥檚 pen will easily do the needful and consign the REF to history.
All that universities would then need to do would be to make sure that their academics published high-quality research articles. Then they could stand back and let the private sector do the heavy lifting.
James Tooley is professor of education policy at Newcastle University.
糖心Vlog
Register to continue
Why register?
- Registration is free and only takes a moment
- Once registered, you can read 3 articles a month
- Sign up for our newsletter
Subscribe
Or subscribe for unlimited access to:
- Unlimited access to news, views, insights & reviews
- Digital editions
- Digital access to 罢贬贰鈥檚 university and college rankings analysis
Already registered or a current subscriber?









