Research outputs do not make a university ranking

The Department of Higher Education and Training’s (DHET) recent Report on the Evaluation of the 2019 Universities’ Research Output has led to some excitement with various claims made about the positioning of certain universities in comparison to others.

The DHET’s research outputs policy encourages universities to engage in research. On an annual basis university research outputs are evaluated and generate research outputs subsidies, intended to provide operational funds to support and stimulate research activity.

Depending on the mandate and mission of an institution, we would expect to find different outcomes in terms of research productivity. The slavish, and sometimes perverse pursuit of “research productivity” can lead to skewed outcomes with quantity outplaying quality.

“Brute” data on publication output units is reflected in Table 1 on page 20 of the report. This table reports the quantity of publication output units for the system, based on the DHET’s proportional evaluation system. At face value it shows the “top six universities” to be the University of KwaZulu-Natal (UKZN), University of Johannesburg (UJ), University of Pretoria (UP), Stellenbosch University (SU), University of Cape Town (UCT) and the University of the Witwatersrand (Wits) in that order. These six institutions collectively produce 60% of the publication output units for the system. The next eight institutions, North West University (NWU), University of South African (Unisa), University of the Free State (UFS), University of the Western Cape (UWC), Rhodes University (RU), Nelson Mandela University (NMU), University of Limpopo (UL) and University of Fort Hare (UFH), in that order, collectively produce 30% and the remaining institutions, including all the Universities of Technology (UoTs), produce the final 10% of the system’s publication units.  

The majority of units (81.80%) are awarded for journal publications. Sorting the brute data on journal publication outputs from highest to lowest gives UKZN first position, followed by UP, UCT, UJ, SU and Wits. Journal articles in accredited journals are important research products and perhaps the best measure of research publication productivity.

The next highest proportion of publication output units are awarded for books (12.15%). Doing a similar sort, the top six are UJ, SU, UFS, UP, Wits and UCT, in that order. UKZN has dropped to eighth, replaced by UFS. Book publication units increased substantially after a change in policy in 2016. Incentivising scholarly books is important, however recognition of quality outputs needs work.

Finally, the smallest contribution proportionally is for conference publication units (6.05%).  The “top six” here are: UJ, NWU, SU, UCT, UP and Unisa. What is remarkable is that UJ has 23.20% of all publication units of this type, compared to NWU with 9.34%.  

The quality of research outputs in conference proceedings and gaming of the system in conference proceedings was highlighted in a DHET commissioned report, The Quality of South Africa’s Research Publications (CREST, 2019). Suspicious activity, including a university claiming outputs for 113 conference papers by a single author in 2017, was identified. The report recommends that a maximum number of conference publication units per author per annum should be implemented to discourage unethical behaviour. 

Table 1 therefore needs to be interpreted with caution. If comparisons are being made then other metrics need to be looked at and, in particular, normalised data needs to be considered.

Table 12 shows the publication units per capita. This considers the number of permanently employed academics in each institution. The same six universities are at the top of the list however the order is different. The highest number of publication units awarded per academic is recorded for UKZN (1.83), followed by SU (1.72), UJ (1.71), UP (1.68), UCT (1.64) and Wits (1.59). Smaller institutions such as RU and UFH move up the table.

But this still leaves out an important aspect of a university’s research activity — the production of research master’s and doctoral graduates, a core responsibility of research-focused universities. Table 13 considers the “weighted per capita research output” which considers publication and research graduate output units per permanent academic staff.  Here the top sixchange again. UKZN remains top with 3.62 per capita weighted research output units, followed by UP (3.61), SU (3.39), Wits (3.04), UCT (2.88) and RU (2.82).  This table would be most appropriate when attempting to “rank” institutions in terms of research productivity.

In Table 13 the top six, based on the proportion of permanent academics with doctoral qualifications are UP (69.70%) followed by Wits (66.00%), UCT (62.40%), UKZN (61.50%), UWC (59.90%), and RU (59.40%).  Interestingly SU is replaced by UWC in this list. One would expect a high level of correlation between staff with doctoral degrees and the per capita publication and weighted per capita research output.

Table 14 shows the ratio of doctoral graduates to doctorate staff members by university. This ratio could be interpreted as a measure of efficiency of staff in relation to supervision and doctoral student completion.  UKZN is again on top of the table with a ratio of 0.59, followed by SU (0.53), UP (0.47), Unisa (0.42), RU (0.41) and NWU (0.39). In this list Wits, with a ratio of 0.37 drops to position eight.

The report illustrates how research outputs of the system have improved over time. Certain universities are consistently in the top tier of the various lists. While all universities are to some extent involved in research activity, traditional universities are more research productive on the whole than comprehensive universities or UoTs.

One critique is that the report does not reveal any qualitative data about the system. For example, it does not report on impact; it says nothing about whether or not this is research that matters; there is no information on outputs produced through collaborative partnerships; or, what proportion is produced by researchers in conjunction with their doctorate or master’s students (a form of double dipping).

While the research outputs policy and subsidy system has effectively incentivised institutions to produce research, it has also had some unintended consequences with gaming the system.

A 2018 the Academy of Science of South Africa (ASSAf) publication, Recognising Individual Contributions to Collaborative Research shows that the proportional system implemented by the DHET fails to recognise the contribution of individuals within large research teams of greater than 100 authors. Changes in the practice of modern science in disciplines including physics, genomics, medicine and statistics is a global phenomenon and have resulted in individuals working within large teams publishing impactful research. The ASSAf report makes some suggestions on how this kind of research, which is highly important, impactful and really matters can be “counted”.

In addition, transdisciplinary research directed at societal challenges may need special incentivisation going forward. Such research addresses complex, complicated and intersectional challenges including, for example pandemics, unemployment, inequality, poverty and climate change. It not only requires the crossing of disciplinary borders and boundaries but also knowledge co-creation with other stakeholders in society including government, the private sector and civil society. The proportional evaluation system currently implemented is unlikely to recognise this sufficiently.

In conclusion, the research outputs report provides interesting data on research productivity, but is a partial story and should not be seen as some kind of university ranking system. The DHET’s commitment to working with the system to develop a research quality framework that can assist in instilling ethical research practices and driving the quality of research as opposed to quantity is welcome. The need to better understand the changing research practices across scientific fields and in particular to find ways of recognising high-impact research produced by large research teams will be important. The system must also consider how to factor in high-impact research that matters and leads to innovation.  

We make it make sense

If this story helped you navigate your world, subscribe to the M&G today for just R30 for the first three months

Subscribers get access to all our best journalism, subscriber-only newsletters, events and a weekly cryptic crossword.”

Related stories


Already a subscriber? Sign in here


Latest stories

Grilling for UK leader Boris Johnson after top ministers quit

The prime minister has faced lawmakers' questions after two of the most senior figures in his government resigned. The finance and health ministers said they could no longer tolerate the culture of scandal

Declare an ‘energy emergency’, says National Planning Commission

The commission said the goals of the National Development Plan, which it is charged with advancing, ‘cannot be achieved without energy security’

PPE looter Roshan Morar lands plum KZN treasury contract

Corruption Watch wants politically connected auditor blacklisted and appointment reversed

Debt-ridden Eskom faced with ‘complex issues’ in its just transition...

Shutting down Eskom’s coal plants and shifting to cleaner technologies will bring ‘cleaner air’ to the polluted Mpumalanga Highveld, says the utility’s environmental manager

press releases

Loading latest Press Releases…