There are an amazing number of scientific urban legends in the lab too that are passed down orally from senior to junior members. My personal favorite is that drinking MilliQ (ultra pure water) will cause the cells in your gut to explode [1]. How such a stupid idea ever got taken seriously by scientists who should know better has always amazed me.
[1] On the topic of drinking MilliQ water I have always encourage all the members of my lab to drink it. The reason why (apart from the taste which I quite like), is that it allows you to detect when you have bacterial contamination of your MilliQ carboy that will ruin your experiments.
Based on the fact that you didn't mention doing it, I am assuming you're missing out on a brilliant opportunity for hijinks here. Suggestion - take an alka-seltzer and some red food dye (I'm not sure if you can drip food coloring directly onto an alka-seltzer slowly enough that the dye is absorbed without causing the alka-seltzer to foam up, but probably you can experiment with something like this) and hide it in your mouth ahead of time. Loudly proclaim that it's harmless to drink MilliQ. Take a big beaker of MilliQ and drink it, allowing plenty of it to reach the alka-seltzer. Pause for half a second, start to look uneasy, grab your stomach, then start allowing blood-red foam to spill out of your mouth. Fall to the ground shaking and moaning.
Bonus points if it'a "demo" you do while teaching a lab class.
I has a colleague who used to like putting dry ice bombs in people’s tip waste containers. You would be sitting there pipetting away and them boom - tips all over the bench.
I've heard that handling ultra pure water, can cause acid like burns due to water drawing out chemicals from the skin. I heard it from a bio PHD, and can't find any evidence googling it now.
Actually most water is a dilute solution of carbolic acid with a pH around 5.5 because of the dissolved CO2 in the water.
This sometimes catches people out when they use pure water to store something like DNA which will breakdown over time at this pH. It is in theory why you should use a buffer like Tris/ETDA to store DNA. I have found that storing DNA for a short time in water is normally OK, but if you want to keep the DNA for more than a month it is a good idea to use a buffer.
Yup, I heard the gut-exploding-ddH2O in a lab and a girl in the class eventually did drink a little bit of it and was absolutely fine. A science champion, if you ask me.
I have had people try to physically stop me from drinking MilliQ water as they thought I was going to die in front of them!
I have always thought it would be a good business idea to sell MilliQ branded water for human consumption. The complete lack of taste is an interesting taste in itself - the closest I can describe it as is it is like drinking liquid air.
Yes you can detect contamination in MilliQ by taste. I don’t know the threshold, but I do know that if it has a taste then the water is contaminated enough to ruin your experiments.
No they are not dangerous. Pathogens are adapted to living in nutrient rich environments (i.e. our bodies) where their problem is avoiding our immune system. The bacteria that can grow in MilliQ water are those adapted to living in low nutrient environments and these bacteria are always harmless - except that is to your experiments :)
It is interesting that the paper does not actually cite the original paper. I guess it is not critical to the paper, but it does complete the story.
This does bring up a important point in that is old scientific publications in languages other than english. I have many times run into references in German, French or Russian that are only available as low quality scanned pdfs. As a monolingual English speaker these are completely inaccessible to me. I would be great to have all of them OCRed so at least we can run them through Google Translate.
>Perhaps, one should not only try to cite the original source, but also look for recent research on the same topic?
This is (or should be) done standardly by all scientists. It is actually the best way to learn about a field. Find the key paper (or at least the paper the field thinks is key) and then read all the papers that cite that key paper.
I am sure this has been done by somebody, but it would be great to have this citation link structure as a graph where you could click on any node and read the paper.
Yes they are, but there is no better way of finding those papers that can’t be found using keyword searches. I have found hundreds of critical papers over the years by using this technique that I would never have found in a thousand years using any other search method.
Take a closer look, Jannic. I did cite that page, created in 2010. The version I referred to was from 20 March 2014 (date also found in the list of references, in line with the requirements of this particular journal)
A small anecdote on this point. I first sent the article to a journal where the editors insisted that I say something about Bunge and the other German scientists involved in the early history. I refused, because it would have taken attention away from the main topic of the article, and sent the article to SSS instead.
Wow. This academic paper we just read on academic urban legends has an urban legend payload - I now believed there was a "misplaced decimal point" (i.e. typo) in a single source, that then became the source of an academic urban legend about spinach.
It'd be interesting if we could break down academic papers into stand-alone facts, which could be cited more succinctly than the paragraphs of prose that we do now. Ideally, this would look more like academic proofs, and it'd be easier to argue and discuss components of a given paper. I'm sure that a good amount of beautiful subtext would be lost, but might it be worth it to have a more accurate way of linking knowledge?
> It'd be interesting if we could break down academic papers into stand-alone facts, which could be cited more succinctly than the paragraphs of prose that we do now.
Perhaps there could be some kind of semantic syntax that would group together the factual claim and supporting information (data and references). For example,
<"e=mc^2",cite1,cite2,data1,data2>
The claims still could be included in prose, but also easily extracted, analyzed, manipulated, etc.
With a little work the claims could form a semantic web, with all that entails, and the validity of cites could be checked automatically when the document is opened or on a regular basis, updating the claim appropriately. For example, the UI could display the claim differently based on whether the cited source has been withdrawn, partially invalidated, superseded, confirmed, rarely cited, heavily cited, based on impact factor, if the source's own citations have been invalidated (e.g., claimC cites claimB which cites claimA; if claimA is invalidated, there should be some sort of indication in claimC -- consider that the invalidation of a seminal claimA would set off a massive chain reaction, as it should), etc.
But that all seems obvious and I suspect I'm not the first to think of it.
That approach has been tried and in the bioinformatics community years ago. Basically scanning published papers that mentioned genes relationship and extracting them in order to build a graph. Say someone publishes "T3SS1 induces cell death" so you can now add an edge to a graph "T3SS1->autofaghy" or something like that.
Please, no. The twitterification of academic papers would be terrible - all subtext and nuance would be lost. This is seen already with 'science journalism': a scientist reports an effect, couched with appropriate caveats, but a journalist then rips that effect free of any nuance or context and states it as fact. It can be highly (and sometimes intentionally) misleading.
Exactly, and it's not like papers are collections of 'facts' that can be extracted anyway. Papers are long arguments based on some data choices, some analysis choices, and followed by interpretations --- none of which are 'facts', and all of which are debatable.
While I can imagine that being amazing if that could be done well, but it seems like a really difficult problem to get right.
I would prefer an approach like PaperBricks caught on instead: http://arxiv.org/abs/1102.3523 -- where a single section of a current paper could be published by itself.
I think it would be both amazingly useful and could speed up the iteration rate of science substantially.
This could now be accomplished by providing an offset (say number of characters from the start) to where the information cited can be found.
What I find more of an issue is when trying to track down the source of some “fact” is going three or four papers up the citation chain and finding the paper at the end doesn’t actually say what the starting paper claims.
I am thinking in the exact opposite direction: stop this publication influence, citations SEO and plagiarism madness.
Get shit done and put it on your blog.
Normal people hear ideas, forget them, internalize them and regurgitate them as their own all the time. While you're scouring databases and calling libraries to try to find who's got access to the paper you're looking for because it looks like a primary source of what you're working on, you could be experimenting from secondary sources or even from textbooks if the thing is now mainstream enough.
OK what if the experimental setup is hard to replicate and I just make up attractive sounding claims for for my blog with flashy graphics and a folksy charming writing style?
what's the difference with today? You didn't fight for the author's order and you didn't spent months on publication issue, but your bullshit is still in the wild.
If you want to be in Nature today, you need attractive claims and flashy graphics already. And the peer review is just not working, they are favoring already famous people and ideas.
> If you want to be in Nature today, you need attractive claims and flashy graphics already. And the peer review is just not working, they are favoring already famous people and ideas.
Yeah, my first thought was, "wow, I always thought that about spinach!" But, that wasn't the point of the article at all. This was an engagingly written article that kind of read like a mystery whodoneit?
This article made me think of The Leprechauns of Software Engineering[1], a good book about the myths that are being passed as facts because of bad citations in Computer Science papers.
> The explanation for this phenomenon is usually that authors have lazily, sloppily, or fraudulently employed sources, and peer reviewers and editors have not discovered these weaknesses in the manuscripts during evaluation.
I've seen this sort of thing in the literature I'm familiar with (HF Radar). I don't think
it's laziness or fraud, but a version of the telephone game played as a kid. A 'fact' will
get stated, and then repeated over and over until some of the original meaning is lost and
it becomes gospel. This can happen because the fact is not crucial to the science in the
paper but always part of the introductions or statements about how the system works.
The statement I'm thinking about in particular is "if you have N receive antennas, you
can only find (simultaneously) N-1 signals". This is true if you are using covariance-based
methods such as Multiple Signal Classification (MUSIC). It got repeated and repeated
until the caveats (simultaneous, and MUSIC) are longer included. Of course, if you have
two orthogonal antennas you can distinguish two directions, perhaps not in complex situations!
[1] On the topic of drinking MilliQ water I have always encourage all the members of my lab to drink it. The reason why (apart from the taste which I quite like), is that it allows you to detect when you have bacterial contamination of your MilliQ carboy that will ruin your experiments.