When Peter Swinnerton-Dyer exchanged the vice-chancellor鈥檚 office at the University of Cambridge for the chairmanship of the University Grants Committee in 1983, he was puzzled to find that no one on the committee could give him an 鈥渋ntelligible鈥 explanation of how it allocated research funding.
Then, as now, teaching funding was distributed between universities predominantly on the basis of the number of students taught by each institution. But the committee, which consisted largely of senior academics, distributed the research funding passed to it by the Department of Education and Science according to a process 鈥渟hrouded in mystery鈥.
鈥淚 think in fact they looked at last year鈥檚 grant and adjusted it for [recent] circumstances,鈥 Swinnerton-Dyer, now 86, recalls. 鈥淯niversities could go up or down but they had to make an ad hoc case鈥 for increased funding on the basis of 鈥渟ome new development鈥.
However, the squeeze being applied to higher education funding by the Thatcher government in the 1980s quickly led the number theorist to conclude that such a lack of transparency would soon become untenable 鈥 not so much because of ministerial demands for more accountability, but because universities needed to be reassured that their shrinking allocations were determined fairly.
糖心Vlog
鈥淭here had always been rumblings of discontent but one could see the rumblings were bound to increase,鈥 Swinnerton-Dyer says.
The UGC鈥檚 鈥渧ery limited鈥 resources meant the new mechanism had to be one that 鈥渄idn鈥檛 demand very much effort on the part of the committee and, consequently, didn鈥檛 demand very much effort on the part of the universities鈥.
糖心Vlog
Swinnerton-Dyer鈥檚 solution was to establish a 鈥渞esearch selectivity exercise鈥, the first of which ran in 1986, to assess the quality of university research. This was the precursor to the research assessment exercises that 鈥 unbeknown to him 鈥 would ultimately become such a significant feature of academic life in the UK.
Today, with only two weeks to go until the census date (the cut-off point for including staff) for the 2014 research excellence framework, it is a subject much discussed. But just how different was the process in those early days?
The 1986 exercise certainly contrasted in scale. The UGC鈥檚 existing subject subcommittees were charged with assessing departments in their area of expertise, chiefly on the basis of just five research outputs 鈥 books, papers or patents 鈥 from the previous five years, on which the department in question would be 鈥渃ontent for its total output to be assessed鈥. This compares with four outputs per academic submitted today.
鈥淲e didn鈥檛 say 鈥榖est outputs鈥 because that would have led to all sorts of disputes from philosophers,鈥 Swinnerton-Dyer says.
Departments were also permitted to submit up to four pages of general description about their strengths, which could include industrial links as well as pure research prowess. Not everyone, however, followed this guidance. 鈥淥ne department totally ignored that and sent in 52 pages 鈥 we pulped the last 48,鈥 Swinnerton-Dyer reveals.
He would have been prepared to confine the selectivity exercise to the sciences, which accounted for the bulk of research spending and which he sensed would find it most acceptable. But, in the event, humanities departments were keen to be involved.
One subject committee was so confident that it already knew all it needed to know about each university鈥檚 departmental quality, it produced a provisional classification before it received any submissions and 鈥渨hen it got all the extra evidence it saw no reason at all to alter any of the classifications鈥. But, for Swinnerton-Dyer, the subject committees鈥 pre-existing depth of knowledge 鈥 gathered partly through their regular departmental 鈥渧isitations鈥 鈥 was a distinct strength of his design, and he laments that by the time of the second selectivity exercise in 1989, they, along with the UGC itself, had been abolished.
In its place was the Universities Funding Council, an essentially 鈥渘on-expert body鈥 chaired by former Cranfield Institute of Technology (now Cranfield University) vice-chancellor Lord Chilver. Swinnerton-Dyer became the inaugural chief executive.
糖心Vlog
Some academics complained that the first exercise had been biased towards larger departments and lacked consistency across subjects. However, judging by the scarcity of complaints he received, Swinnerton-Dyer concluded that universities were 鈥渆ssentially unworried鈥 by it. 鈥淭hey welcomed the more transparent allocation machinery and accepted that it must involve a research assessment process of some kind and [recognised that] the process was very undemanding.鈥
But the Committee of Vice-Chancellors and Principals 鈥 the forerunner of Universities UK 鈥 did have one 鈥渞eal鈥 complaint: that departments 鈥渉adn鈥檛 had the opportunity to put forward their full strength鈥. Wary of vice-chancellors complaining to ministers about his obstinacy, Swinnerton-Dyer permitted universities to submit two publications for every member of staff for the 1989 exercise 鈥 even though he thought it was 鈥渁 silly idea鈥. Information was also sought on a department鈥檚 total volume of publications; Swinnerton-Dyer had chosen not to collect this in 1986 on the grounds that it would 鈥渙nly encourage the production of low-quality papers鈥.
There were other major changes, too. Consultation with universities resulted in the replacement of 1986鈥檚 37 cost centres (subject areas) with 152 subject units of assessment, assessed by nearly 70 panels assembled for the purpose, significantly increasing the scale of the operation. Each panel would now use a new five-point scale to judge the quality of research. Necessarily 鈥渨oolly鈥 universal definitions for each point related to national and international standards. However, Swinnerton-Dyer emphasises that this did not make the quality of research in different subjects comparable on an absolute scale because, for instance, 鈥測ou can be world-leading in dentistry research but still be pretty mediocre鈥.
Lord Chilver also vetoed Swinnerton-Dyer鈥檚 plans to further develop assessment of departments鈥 industrial links, which he accepts were 鈥渘ot very satisfactory鈥 in the first selectivity exercise.
On funding, accounts of the history of the exercise differ. Swinnerton-Dyer recalls using the selectivity exercise to allocate the whole of the UGC鈥檚 and UFC鈥檚 research-related funding. However, other accounts, such as Roger Brown and Helen Carasso鈥檚 recent book, Everything for Sale? The Marketisation of UK 糖心Vlog, cite evidence that only about 40 per cent was allocated on the basis of panels鈥 assessments, with the rest being distributed according to staff and student numbers, research grant income and volume of external research contracts.
Bahram Bekhradnia, director of the 糖心Vlog Policy Institute, also recalls that the research selectivity exercise was not a 鈥減rimary instrument鈥 of resources allocation. As director of policy for the 糖心Vlog Funding Council for England (which replaced the UFC in 1992), Bekhradnia was charged with deciding how the research funding mechanism should respond to 1992鈥檚 abolition of the 鈥渂inary line鈥 between universities and polytechnics.
By then, Swinnerton-Dyer and all other senior figures associated with previous selectivity exercises had moved on and, rather than poring over the past, Bekhradnia devised the rules afresh for what was to be known as the research assessment exercise or RAE.
He concluded that allocating research funding partly on the basis of student numbers would no longer be tenable once the teaching-focused polytechnics joined the system. The large expansion in university numbers would not be matched by a proportional rise in the overall research budget so, if large transfers of funding from research-active staff in pre-1992 universities to largely teaching-focused staff in the post-1992s were to be avoided, it would be necessary to allocate all research funding on the basis of the RAE.
But this meant that the RAE had to become 鈥渕uch more robust and rigorous鈥 than the 鈥渆mbryonic鈥 system Bekhradnia had inherited. Principally, this entailed setting out the rules more clearly and checking outcomes more thoroughly, since 鈥渨hen so much was at stake, we were going to be subject to legal challenge if we didn鈥檛 do it right鈥.
Despite Bekhradnia鈥檚 best efforts, the 1992 results were indeed challenged in the courts by the now defunct Institute of Dental Surgery. The funding council won the case, but the judge was 鈥渦nimpressed鈥 with its processes and 鈥渨arned that administrative law was moving very fast in the direction of transparency and it was no good experts simply saying they used their expert judgement to decide whether this was a good or bad submission鈥.
Hence, panels in the subsequent RAEs were required to be even more explicit about their criteria and working methods. Rules about which staff could have work entered also had to be tightly stipulated given the decision to allow universities to submit only academics they deemed 鈥渞esearch active鈥: this was a response to the fact that many academics in former polytechnics focused exclusively on teaching.
Large research institutions were able to 鈥榟ide a very long tail [of poor work] and still get a 5*鈥欌hose in the tail were effectively being over-funded
The 1992 exercise was also the first to stipulate that eligible staff must be in post on a specific census date; previous exercises had assessed all staff in post for any part of the review period, sparking criticism that they were unduly retrospective given that staff often moved on. The resulting phenomenon of universities 鈥減oaching鈥 each other鈥檚 top research staff ahead of the census date was addressed in the 2001 exercise by allowing departments to share credit for academics who had moved during the assessment period. But Rama Thirunamachandran, director of research, innovation and skills at Hefce between 2002 and 2008, disapproved of this measure and repealed it for the 2008 RAE.
鈥淎lthough there was some staff movement around the time of each RAE it was not excessive,鈥 he argues. 鈥淭he key issue is that the results inform funding for the following six years. It was therefore important that funding followed individuals, rather than funding excellence relating to the past at someone鈥檚 former institution.鈥
Another one-off was 1992鈥檚 separate assessment scheme for pure and applied research, introduced to counter the accusation that the latter had previously been unfairly disadvantaged. According to Bekhradnia, this was dropped from subsequent exercises because most panels were unable to distinguish clearly between pure and applied work.
In 1996, Bekhradnia moved to end any suggestion that RAE success depended more on volume than on quality by no longer seeking any information about the volume of research produced 鈥 although he admits that this 鈥渟illy鈥 misconception lingers to this day.
The next major change to the RAE came in 2008. According to Thirunamachandran 鈥 who took up the post of vice-chancellor of Canterbury Christ Church University on 1聽October 鈥 the single grades that had hitherto been awarded to departments hid a 鈥渕ultitude of sins鈥.
Large research institutions were able to 鈥渉ide a very long tail [of lesser work] and still get a 5*鈥, he recalls. Since such departments received 5*-level funding for every academic they submitted, those in the tail were effectively being over-funded. Conversely, pockets of excellent researchers in low-graded departments were unable to access any quality-related funding. Thirunamachandran鈥檚 solution, which will be preserved for the 2014 research excellence framework, was to introduce more detailed 鈥渜uality profiles鈥, which indicate the proportion of each department鈥檚 research in each quality category, with funding awarded to high ratings wherever they are found.
This new system had the added advantage, for Thirunamachandran, of mitigating the concentrating effects of confining funding almost exclusively to the top two grades on the scale, which 鈥渋sn鈥檛 good for the dynamism of the system鈥.
However, Thirunamachandran鈥檚 efforts to cut the workload associated with the ever-expanding RAE 鈥 which, by now, had become a major concern 鈥 were not successful. A review he carried out with the late Sir Gareth Roberts in 2003 proposed an institutional opt-out for teaching-focused institutions in return for a base level of funding to sustain research capacity. However, this was rejected by the institutions, which, in Thirunamachandran鈥檚 estimation, judged that opting out of the 鈥渕ainstream competition for research funding鈥 would have sent the wrong message to potential funders and business partners.
Nor was the Treasury any more successful in its efforts to trim the cost of conducting the exercise, which in 2008 is estimated to have amounted to some 拢47 million across English universities. In 2006, then chancellor Gordon Brown announced 鈥 鈥渃ompletely out of the blue鈥, according to Bekhradnia 鈥 that the RAE鈥檚 reliance on peer review was to be replaced by an assessment that, especially in the sciences, was based on metrics such as citations, research income and postgraduate numbers.
糖心Vlog
Thirunamachandran did not regard such a system as plausible even in the sciences but, after convincing the chancellor that the 2008 exercise had proceeded too far under the old rules to be abandoned, he dutifully went about mapping out and consulting on Brown鈥檚 proposed system, before concluding that it was unworkable. However, as an inducement for the Treasury to agree to drop its proposal, Thirunamachandran offered instead to introduce an assessment of 鈥渋mpact鈥 to what he had renamed the research excellence framework. (The new name aimed to free it from some of the negativity that had grown up around the RAE.)
鈥淭he Treasury was interested in the not-unreasonable question of 鈥榳hat does 拢1.6聽billion a year in QR [quality-related] funding buy in practical, lay terms, economically, socially and culturally?鈥 That can only be articulated in terms of impact and, to some extent, it was a sort of unstated compact that if we could demonstrate and assess impact, the sort of mindless cry for metrics would subside from the Whitehall end,鈥 he says.
He regards that compact as a 鈥渄ecent鈥 one; it was his successor at Hefce, David Sweeney, who reached the controversial conclusion in 2011 that impact should count for 20聽per cent of total scores in the 2014 REF.
Thirunamachandran admits that the assessment of research 鈥渋mpact鈥 has added another element to the process of assessment but, because other elements have been simplified, he does not believe the effort required from universities and panellists has increased overall. He also argues that a significant amount of the REF-related workload is imposed by universities themselves, as it is 鈥渋n our nature to cross the t鈥檚 and dot the i鈥檚 because we want the best possible outcome for our university鈥.
Both Thirunamachandran and Bekhradnia point out that, per pound distributed, the RAE and REF are vastly cheaper than distributing the same sums entirely via grant applications to the research councils. They also say there is no doubt that the introduction of the RAE has led to an increase in the quality of UK research, measured against international comparisons. In Thirunamachandran鈥檚 view, this increase had also convinced successive governments to invest vastly more resources in university research.
鈥淭he RAE has put a focus on research that didn鈥檛 exist pre-1986, when research was a residual activity that a proportion of academics did. Now it is really core and mainstream,鈥 he says.
But he also acknowledges that this success has a downside: universities鈥 comparative neglect of teaching. Hefce has sought to address this with a variety of initiatives, such as the creation of the 糖心Vlog Academy, an organisation focused on the quality of university teaching. But, according to Bekhradnia, a satisfactory answer will always be difficult to find when research funding is competitive and teaching funding is not. Nor would it be sensible to allocate teaching funding competitively, for the 鈥渦tterly compelling鈥 reason that it would not be 鈥渆thically sustainable鈥 to say to a student: 鈥淵ou are going to study history at a university that is not very good at teaching history and, by the way, you are going to have fewer books in the library and less money per student and a worse staff-student ratio [than other institutions].鈥
Another of Bekhradnia鈥檚 regrets about the RAE is the 鈥済amesmanship鈥 that the necessity of stating the rules more explicitly has promoted. 鈥淭he ideal RAE would be one where people wake up and find they got a score but don鈥檛 know how they got it. If being explicit helps people to change their practices in a way that makes things better then that is desirable, but I doubt if that is the case,鈥 he says.
However, his objection is chiefly to institutions鈥 鈥渨asteful鈥 expense of energy and money on game-playing. Except in a few marginal cases, he doubts that universities鈥 efforts to skew the results to their advantage are successful.
Indeed, the increasing stability of results over successive RAEs has only heightened suggestions that it simply is not worth all the time and effort that goes into it, and the process has been likened to 鈥渁 Frankenstein鈥檚 monster鈥 that is out of control.
According to Swinnerton-Dyer, no one has ever seriously doubted that his 1986 scheme was 鈥渁dequate for getting the estimates roughly right鈥. This is not surprising, since, in his view, 鈥渋f the best five elements [of a department鈥檚 output] are good, the overall quality is probably good as well鈥.
For him, post-1986 iterations of the RAE 鈥渢ry to produce more information in more detail than a tolerable process could do鈥. He believes the rot really set in when vice-chancellors ceased to see the RAE as a funding mechanism and regarded it, instead, as a 鈥渇ree-standing assessment of research quality鈥, with the added advantage of being 鈥渦seful as a means to get rid of people not doing any research or to make them do more teaching鈥.
鈥淚f that is what vice-chancellors want, they can conduct their own internal processes, but, nationwide, I don鈥檛 think such an exercise is justified,鈥 Swinnerton-Dyer says. 鈥淚 don鈥檛 think you can run an assessment of individuals without an absolutely intolerable level of work.鈥
So what does he think of the design of the 2014 REF and its much-debated and controversial requirement for universities to demonstrate the impact of their research? Despite his enthusiasm for assessing industrial links, Swinnerton-Dyer dismisses impact assessment, describing it as 鈥渁 licence for lying鈥 鈥 since the evidence, in his view, is 鈥渦ncheckable鈥. And while the presenting of next year鈥檚 results through quality profiles will give a 鈥渟lightly more nuanced鈥 picture of departmental quality, he doubts whether the additional work involved in obtaining it is merited.
鈥淭he one question a modern civil servant fails to ask,鈥 says Swinnerton-Dyer despairingly, 鈥渋s 鈥榠s it worth the extra effort?鈥.鈥
A developmental process: how research assessment has changed
1986
The first 鈥渞esearch selectivity exercise鈥 launched by the University Grants Committee. Subjects were divided into 37 鈥渃ost centres鈥. For each one, universities were asked to submit five outputs 鈥 which could include patents 鈥 and up to four pages of general description of the unit鈥檚 research and industrial strength.
Assessments were made by the UGC鈥檚 subject subcommittees and results were scored on a four-point scale from 鈥渂elow average鈥 to 鈥渙utstanding鈥.
1989
The Universities Funding Council oversaw an exercise involving 152 subject units of assessment, evaluated by nearly 70 peer review panels. Units were asked to submit up to two publications for every member of staff. Information was also gathered on research student numbers and research income. Results were presented on a five-point scale related to national and international standards.
1992
The 糖心Vlog Funding Council for England ran the exercise, now called the RAE, on behalf of the four UK funding councils. For each 鈥渞esearch active鈥 member of staff in post on the census date, universities were asked to submit up to two publications and two other forms of public output. A letter grade in the results indicated the proportion of staff in each department submitted. An extra year was added to the assessment period for arts and humanities.
Separate assessments were planned for applied and basic research in science and engineering, but many panels struggled to differentiate the two.
About 2,800 submissions in 72 units of assessment were rated by 63 subpanels.
Submissions were audited for accuracy. No funding was given to departments assigned the lowest of the five grades.
1996
Universities filed up to four publications per academic submitted.
Sixty panels assessed work in 69 units of assessment produced over four years for the sciences and six for the humanities. Each sub-panel published its assessment criteria and working methods ahead of submissions and more outside assessors, including academics from abroad and people working beyond the sector, were recruited. Data on output volume were no longer sought.
Two extra levels were added to the now seven-point assessment scale: 3 was split into 3a and 3b and 5* was added at the top. Departments in the lowest two categories received no funding.
2001
Nearly 2,600 submissions were made to 69 units of assessment. Five 鈥渦mbrella groups鈥 of panel chairs in related disciplines met to try to achieve greater consistency and fairer assessment of interdisciplinary work. Submissions awarded top grades were reviewed by international experts. Reductions in the number of research outputs required were allowed for some staff in certain circumstances.
More feedback on results was given to vice-chancellors, and panels鈥 assessment of the strength of their disciplines was published.
Universities whose staff were poached ahead of the census date were permitted to submit two of their outputs, but all the funding still went to their new departments. Departments rated in the top two categories contained nearly 40 per cent of academics, compared with only 13 per cent in 1992.
Funding was withdrawn from departments rated 3b and the amount of funding for those rated 4 was steadily reduced.
2008
The 67 subpanels were overseen by 15 main panels. Explicit criteria were introduced for the assessment of applied, practice-based and interdisciplinary research.
Results were presented as 鈥渜uality profiles鈥, setting out the proportion of each department鈥檚 submissions that fell into five quality categories. Research in the top three categories was originally funded, but this has since been reduced to the top two.
糖心Vlog
2014
The renamed research excellence framework will incorporate four main panels overseeing 36 subpanels. It will incorporate an assessment of non-academic impact, accounting for 20聽per cent of the final marks. Treasury proposals for the exercise to be metrics-driven were abandoned. Figures collected by the higher education funding bodies indicate that UK universities plan to submit the research of some 54,300 academic staff for assessment in the REF, up from 52,400 in the 2008 RAE.
POSTSCRIPT:
Sources: ; by Valerie Bence and Charles Oppenheim (Journal of Educational Administration and History, 2005); Everything for Sale? The Marketisation of UK 糖心Vlog by Roger Brown and Helen Carasso (2013).
Register to continue
Why register?
- Registration is free and only takes a moment
- Once registered, you can read 3 articles a month
- Sign up for our newsletter
Subscribe
Or subscribe for unlimited access to:
- Unlimited access to news, views, insights & reviews
- Digital editions
- Digital access to 罢贬贰鈥檚 university and college rankings analysis
Already registered or a current subscriber?




