Stop the deluge of science research
The increasing pace of human discovery is a curse – we need to rethink what it means to publish the results of research
The rapid growth of scientific literature is often seen as
evidence, if evidence were needed, that the pace of human discovery is
accelerating. On the contrary, however, it is becoming a curse – one
that requires us to radically rethink what it means to publish the
results of research.
Relentlessly
– day after day, year after year – scientists are uncovering new facts
about the world. If anything, the startling rate at which this happens
appears to be increasing, but how would we know if such an impression
was true? One way is to look at the rate at which scientific papers are
published and these have indeed been appearing at an ever-increasing
rates for decades or even centuries.
As reported in a recent paper,
the rate of growth of cited (ie somewhat influential) scientific
publications has risen from less than 1% before the middle of the 18th
century to 2-3% in the first half of the 20th century, and 8-9% today.
That last figure is equivalent to a greater than doubling of scientific
output every decade. What better testament to human intellectual
progress and the power of the scientific method?
Not so fast. For a
start, not every publication is equal. When the author Theodore
Sturgeon first invoked his famous law ("90% of everything is crap") he
was talking about science
fiction, but he could equally well have been talking about science.
Estimates vary wildly, but probably between a quarter and a third of all research papers in the natural sciences go uncited. A much larger proportion is cited only by their own authors or by one or two others.
This
is not necessarily a sign of inadequate or wasteful research, but it
should give us pause. It is at least partly a result of the imperative
for researchers to publish in order to continue to secure funding and
employment, and the accompanying incentive to salami-slice research
results so as to secure the maximum number of publications from any
given piece of work. This in turn leads to the bane of every scientist's
existence: far too many papers to read in far too little time.
This
situation is only going to get worse, even in the absence of any
genuine increase in scientific activity. One reason is the rise of
author-pays open-access publishing. This has the benefit of allowing
anyone to read the research in question without paying the publisher,
but also has the disadvantage of giving publishers a strong commercial
incentive to issue as much content as possible.
In general I am a supporter of open access,
but subscription business models at least help to concentrate the minds
of publishers on the poor souls trying to keep up with their journals.
As
if this were not enough, proponents of open science (including me) are
proposing that researchers should start publishing all of their work –
complete with full data sets, comprehensive methods, negative results
and "failed" experiments.
In an age of bottomless digital
information this is technically achievable and has the potential to
greatly increase the transparency, impartiality and reproducibility of
research – particularly welcome at a time when science is going through
something of a crisis of confidence on these fronts. But it hardly promises a more manageable literature for those who are desperately trying to absorb all this information.
The
only practical solution is to take a more differentiated approach to
publishing the results of research. On one hand funders and employers
should encourage scientists to issue smaller numbers of more significant
research papers. This could be achieved by placing even greater
emphasis on the impact of a researcher's very best work and less on
their aggregate activity.
On the other they should require
scientists to share all of their results as far as practically possible.
But most of these should not appear in the form of traditional
scholarly papers, which are too laborious for both the author and the
reader to fulfil such a role. Rather, less significant work should be a
issued in a form that is simple, standardised and easy for computers to
index, retrieve, merge and analyse. Humans would interact with them only
when looking for aggregated information on very specific topics.
What
would such a publication look like? I don't know exactly, but we can
see signs in born-digital data publishing platforms such as Figshare (a portfolio company of Digital Science, which I run), Zenodo (an initiative from CERN, the European particle-physics laboratory) and Dryad (a not-for-profit membership organisation that supports the dissemination of research data).
It
is easy to forget that 350 years ago the scientific journal was itself
an innovation, enabled in large part by the emergence of the printing
press some 200 years earlier. Today our challenge is to create a new
kind of scientific publication, internet-enabled and fit for a data-rich
digital age.
Nenhum comentário:
Postar um comentário