The Italian miracle is an illusion

Improved citation performance amid falling funding is the result of gaming, say Alberto Baccini, Eugenio Petrovich and Giuseppe De Nicolao

October 7, 2019
The painted dome by Andrea Pozzo, in the Church of Saint Ignatius of Loyola in Rome
Source: iStock

In 2010, the Italian research system was profoundly changed by a landmark university reform.

In the name of efficient use of public money, research funding was cut, the recruitment of researchers was curtailed and a complex evaluation system was introduced to monitor and improve the quality of Italian research.

At the heart of that system are bibliometric indicators. They play a central role not only in the assessment of departments and research centres, but also in the recruitment, promotion and evaluation of individual researchers. For instance, researchers must now undergo a national scientific habilitation that cannot be passed unless their citations, h-index and output exceed the national median in their field.

On the surface, the reform appears to have had the desired effect. Notwithstanding the reduction in personnel and funding, Italian science achieved a kind of miracle, increasing both the volume and citation impact of its publications. On the latter measure, Italy overtook the US in 2012, rising to second place behind the UK among the G8 countries, according to Elsevier’s SciVal Analytics. In a 2016 report commissioned by the UK government, Elsevier predicted that, “based on current trajectories, [Italy is] set to overtake the UK in the near future”.

Meanwhile, a Nature editorial in August observed that “although Italy’s spending on research and development – at around 1.3 per cent of its gross domestic product – sits well below the EU average of 2 per cent, its research performance continues to improve”.

And, according to ANVUR, the Italian governmental research evaluation agency, “Italy is an international exemplar in the design and implementation of research evaluation exercises”. Its 2018 biennial report says that the “sophisticated use of multiple indicators” to “deal with the problem of comparison between disciplines” has meant that “in recent years Italian universities have shown a gradual convergence towards higher standards, both in teaching [and] research”.

However, bibliometric-centred policies have a dark side, which is increasingly being brought to light. Indicators invite gaming. Some of the highly cited researchers identified by Clarivate Analytics are, in fact, massive self-citers. And, in Italy, some recent studies document how, in some fields, self-citation has increased significantly since the reform.

Our study, “Citation gaming induced by bibliometric evaluation: A country-level comparative analysis”, published in September in Plos One, confirms that the recent sharp rise in Italian citation impact is, in essence, a mirage, generated by a massive national change in citation habits after the 2010 reform.

To illuminate the real situation, we introduce a new indicator called “inwardness”, defined as the ratio between the total number of a country’s self-citations and the total number of citations received by that country in the same time frame. It is able to track not only individual self-citers but also intra-national citation clubs.

We observe that, after 2009, most scientific fields in Italy saw their inwardness jump: a trend unique among European members of the G10. In 2016, Italy became – both globally and for a large majority of research fields – the country with the second highest inwardness and the second lowest rate of international collaborations, behind only the US.

The likely explanation is that the bibliometric targets have stimulated large numbers of Italian scientists to both increase their level of self-citation and join citation-exchange clubs.

So what are the lessons of this huge case of collective citation doping? It might be claimed that the perverse effects of bibliometric targets could be countered by adopting better indicators. However, excluding self-citations from calculations would just reinforce the incentive to create citation clubs – which, by their nature, are elusive.

In the end, there is no escape from Goodhart’s famous maxim that when a measure becomes a target, it ceases to be a good measure. Our results show that any new indicator is doomed to be matched by new forms of strategic adaptation, feeding an endless perverse cycle.

Miracles do not happen. There is no magic wand – bibliometric or otherwise – able to transform the scientific performance of a country while also cutting investment.

Alberto Baccini is a professor and Eugenio Petrovich is a postdoctoral researcher in the department of economics and statistics at the University of Siena. Giuseppe De Nicolao is a professor in the department of electrical, computer and biomedical engineering at the University of Pavia.

Related articles

Reader's comments (1)

Pigs don't get fatter just because you keep weighing them!

Have your say

Log in or register to post comments

Most commented

Mary Beard’s recent admission that she is a ‘mug’ who works 100 hours a week caused a Twitter storm. But how hard is it reasonable for academics to work? Who should decide? And should the mugs be obliged to keep quiet? Seven academics have their say

20 February

Sponsored