Foundation essay: This article on the open access and science by Björn Brembs is part of a series marking the launch of The Conversation in the UK. Our foundation essays are longer than our usual comment and analysis articles and take a wider look at key issues affecting society.
There is a looming crisis in science, and we must act now to prevent it.
Currently, the number of scientific papers retracted from a large database of thousands of biomedical journals is a mere 0.05%, a low rate. But recently this rate has been rising. That rise is so quick that, if the trend were to continue, as many scientific publications will have to be retracted as are being published by about 2045.
Usually, any problem at this scale will have a number of underlying causes. Science, like any enterprise, is a complex affair. In this special case, however, a large part of these causes can be traced to the reward structure that scientists have built around them to make this enterprise work. They are related to how scientists choose to publish their results and how fellow scientists reward this publishing behaviour.
Starting in 1960, the rising number of scholarly journals increased subscription costs for libraries too quickly. Institutional libraries and university faculties were faced with the problem of deciding which journals to read and subscribe to, keeping their budgets in mind. One solution, it seemed, was to rank journals according to how many citations each of its publication garnered on average.
Back then, only a few realised that a generation later their future colleagues would not only adjust their reading habits to fit these journal rankings, but that those reading habits would also translate into new publishing habits. After all, everyone wants their work to be read by as many people as possible.
Today, where you publish matters more than what you publish. This behaviour may vary from country to country and field to field, but it is quite widely known that tenure, promotion and grant funding all depend on where in the hierarchy of journals one’s research results have been published.
What makes it worse is this feedback loop: more citations means more attention. More attention, more citations. As long as space is scarce, more manuscript submissions mean higher rejection rates, for some journals the rejection rates can be as high as 90%. Like long lines in front of clubs, higher rejection rates mean more validation for the few researchers that actually manage to get published in these journals and hence more citations.
Not surprisingly, as researchers adjusted their behaviour, so did publishers. The top journals did not increase space for the increased demand. Instead, they spawned ever more low-ranking offspring to reinforce the ratchet and keep the demand high.
With the arrival of the internet, publishers went through a complete reversal of their role in knowledge dissemination. Before the internet, publishers would heed their profession, investing in printing presses and delivery systems for a wide and efficient system of dissemination to the public. Today, academic publishers have become the opposite. These corporations are now “hiders”: they invest in paywalls, hiding the research from the public that paid for it.
But perhaps this is all fine and well? After all, publishers select the very best science using professional editors and need to protect their investment.
Superficially, one may assume that such strong selection would only enable the most groundbreaking, methodologically sound research to clear such monumental hurdles. However, the data speak a different language. While quantifications of “novelty” and “importance” have been found to correlate significantly with journal rank, the correlation is so weak, that for every publication correctly published in a high-ranking journal, there is one that should not have been published there.
Conversely, such a large percentage of research findings that did not manage to get published in a top journal end up being recognised as deserving after being published in a lower journal, that the practical value of journal rank as an evaluation signal is negligible, despite the statistics. Even more damaging, however, quality-related metrics such as technical quality, methodological soundness, reproducibility and so on either failed to show any correlation with journal rank, or correlated negatively.
These data support the observation of more frequent scientific fraud scandals and the increasing range and scope of each case, mostly based on fraudulent work published in top journals.
Apparently, as the pressure mounts to publish in top journals, authors race to find a way to present their work as important and groundbreaking, no matter the actual content and quality. Science today has become a rat race between authors marketing their research and professional editors struggling to see through the fluff.
It is the perfect arrangement to bring out the worst in all participants: scientists fear for their livelihoods and do what is necessary to survive, with little regard to science. Publishers fear their business model and do what is necessary to survive, with little regard to science. And librarians are caught in the middle.
This straightforward analysis of the current state of affairs suggests a slightly more alarmist suspicion. Exponential effects are typical for social phenomena, where feedback is driving much of the dynamics of the system.
The already described, relatively fast feedback loop propped up by journal rank supports a second, slower loop: after a generation of rising pressure to publish in top journals, many of the now leading scientists excel at marketing their research to top journals. After all, this is how they were selected. It may of course happen that these researchers are also excellent scientists, but this was often enough not the selection criterion.
These scientists are now training the next generation of scientists in how to be successful in science. The anecdotal evidence suggests that - intentional or not - marketing one’s research to top journals may play a dominant role in this training. If this were indeed the case, it would explain the exponentially increasing retraction rates in the scientific literature and constitute cause for grave concern with respect to the looming crisis in science.
Of course, laziness, vanity, ambition, desperation, hubris, self-aggrandisement, sleep deprivation or bad time management are not going to disappear any time soon. Scientists are human too. The realistic task can thus not be to altogether eliminate shoddy, unreliable science, but to reform our infrastructure which currently appears to be purposefully designed to exacerbate these traits and bring out the worst in scientists.
Scientific discoveries are unreliable enough under the best of circumstances. Given the complexity of the matter, there surely is no need to worsen the situation. Instead, we must design an infrastructure that brings out the best in scientists and aligns their incentives with that of science and the public.
Given the current status quo, this is technically rather straightforward: one has to remove the incentives for scientists that encourage and reward marketing at the expense of science and instead encourage and reward actions of scientists that benefit both science and scientists.
Given that scholarly communication is at the heart of the scientific endeavour, this is where we ought to place most emphasis in our infrastructure reform. We need to abandon the traditional idea of journal publishing altogether in favour of an institution-based platform that allows every tax-payer, scientist or not, to access all the literature, data and software that is being paid for out of the public purse.
For the past two decades, corporate publishers have proven to have become exceedingly poor custodians of the public good that is research results. From their continued and often shamefully backstabbing and cunning resistance to change their business model to provide even simple read-only access to the scientific literature to those who paid for the research in the first place, to the current prohibition of text- and data-mining access for scientists to their own literature, corporate publishers have proven beyond the shadow of a doubt that they have forfeited every right to be perceived as partners of scientists. Instead, the evidence weighs heavy that many if not most of these corporations are now enemies of science.
Small versions open access platforms exist in many university libraries today, in various forms. They need to be combined using common standards. In less fortunate parts of the world, where research billions are not as abundant, governments felt forced, long before us, to realise how wastefully corporate academic publishing treats tax-funds.
These countries have very successfully been running a multi-national platform based in Brazil, SciELO. If implemented on a global scale and funded with only a single digit fraction of what corporate publishing costs today, this platform would eliminate the pernicious incentives currently in place and provide full open access to the public.
Recent initiatives like SHARE or the DPLA could serve as starting points for such developments in the rich countries. Provided that the reputation system built into this platform is guided by the same evidence-based reasoning and tested with the same scientific rigour we commonly apply to our experiments, this kind of reform would not only come for free, it would save billions in tax funds every year.
It is a collective disgrace for the entire scientific community that saved money and, potentially, saved lives are apparently not sufficient incentive to drive forceful and rationally designed reform right now. Instead, as in many other communities, vested interests are still keeping an upper hand. This must change now.