Are the electronic literature databases making us forget about the older literature?

A short blog by Isis, a self-proclaimed domestic and laboratory goddess (and who are we to disagree?), reminded me that I have been meaning to write a blog about this issue for some time and that I should get on with it.

Isis mopes that articles not available electronically are just not worth the effort tracking down. At least not when it’s raining. Or something like that.

A few years ago, I noted an annoying trend in my research field, bioinformatics. The odd research paper–one or two too many for my tastes–re-invented or re-discovered old results and methods. These authors, presumably innocently, seemed unfamiliar with the earlier work. These papers seemed mostly from younger research groups, bioinformatics scientists who joined the field after the genome era. I put this down to older group leaders knowing the earlier literature. But then, again, maybe it’s the confirmation bias of a slightly older scientist…

After some ruminating, I wondered if the cause was PubMed and laziness.

PubMed is an on-line database of research paper abstracts and links to their original copies hosted by the NCBI. PubMed is a–if not the–main source of access to the scientific literature for many biologists.

There were two main things that I put these “re-inventions” down to. One I want to touch on in a later article as it bothers me: some researchers seem to incorrectly think that computational biology (aka bioinformatics) started with the genome projects, or at least that literature prior to that time can be ignored. This is peculiar to this field, but the general concept may apply to other fields that chugged along relatively unnoticed for years then suddenly become very popular and in demand.

The other was that the lack of presence of the older literature in the on-line databases: those who only used on-line databases and electronic copies of research articles would miss the older works.

It was quite a disturbing thought. Science relies on the continuity of the record. If that’s broken, a lot is lost.

PubMed, in trying to make the literature more accessible, may have–at least temporarily–encouraged “lazy” researchers to break the chain of knowledge.

This has since been alleviated by efforts to bring the back catalogues to electronic form. This will no doubt continue and in time we’ll all be in a stronger position. (Assuming some of the those grotty scans of early manuscripts are tidied up!)

Nevertheless, I can’t help but worry that the combination of easy access to the newer literature, the work involved locating the older literature and the slow, but sure, mothballing of the paper (card) and microfiche index systems encouraged researchers to consider old(er) papers irrelevant and this has persisted to establish a new norm.

Certainly it is tempting to ignore them. Research fields move a such a clip that it’s hard to find time to keep up, let alone dig up leads fifty or more years old. Hardly anyone seems to rummage the dusty corners of the library to locate their treasures anymore. OK, it’s time-consuming and carrying stacks of bound tomes works the arm muscles, but it suggests it’s thought that if it can’t be gotten on-line, it’s not worth the while.

Am I alone on this one? Or am I just getting older and fussier…?!


This of course justifies those review papers that give long views of a field, mostly written by those starting to sport white hairs :-)