Recently I received and email from Rachid Ounit pointing me to a new open access paper he had on a metagenomics analysis tool called CLARK. I asked him if he would be willing to write a guest post about it and, well, he did. Here is it:
Well, just got pointed to this paper: Metagenomic chromosome conformation capture (meta3C) unveils the diversity of chromosome organization in microorganisms | eLife by Martial Marbouty, Axel Cournac, Jean-François Flot, Hervé Marie-Nelly, Julien Mozziconacci, Romain Koszul. Seems potentially really interesting.
It is similar in concept and in many aspects to a paper we published in PeerJ earlier in the year (see Beitel et al., 2014 Beitel CW, Froenicke L, Lang JM, Korf IF, Michelmore RW, Eisen JA, Darling AE. (2014) Strain- and plasmid-level deconvolution of a synthetic metagenome by sequencing proximity ligation products. PeerJ 2:e415 http://dx.doi.org/10.7717/peerj.415.
Yet despite the similarities to our paper and to another paper that was formally published around the time of ours, this new paper does not mention these other pieces of work any where in the introduction as having any type of “prior work” relevance. Instead, they wait until late in their discussion:
Taking advantage of chromatin conformation capture data to address genomic questions is a dynamic field: while this paper was under review, two studies were released that also aimed at exploiting the physical contacts between DNA molecules to deconvolve genomes from controlled mixes of microorganisms (Beitel et al., 2014; Burton et al., 2014).
Clearly, what they are trying to do here is to claim that since they paper was submitted before these other two (including ours) was published, that they should get some sort of “priority” for their work. Let’s look at that in more detail. Their paper was received May 9, 2014. Our paper was published online May 27 and the other related paper by Burton et al. was published online May 22. In general, if a paper on what your paper is about comes out just after you submit your paper, while your paper is still in review, the common, normal thing to be asked to do is to rewrite your paper to deal with the fact that you were, in essence, scooped. But that does not really appear to be the case here. They are treating this in a way as “oh look, some new papers came out at the last minute and we have commented on them.” The last minute would be in this case, 6 months before this new paper was accepted. Seems like a long time to treat this as “ooh – a new paper came up that we will add a few comments about”.
But – one could quibble about the ethics and policies of dealing with papers that were published after one submitted one’s own paper. From my experience, I have always had to do major rewrites to deal with such papers. But maybe E-Life has different policies. Who knows. But that is where things get really annoying here. This is because it was May 27 when our FINAL paper came out online at PeerJ. However, the preprint of the paper was published on February 27, more than two months before their paper was even submitted. So does this mean that the authors of this new paper do not believe that preprints exist? It is pretty clear on the web site for our paper that there is a preprint that was published earlier. Given what they were working on – something directly related to what our preprint/paper was about, one would assume they would have seen it with a few simple Google searches. Or a reviewer might have pointed them to it. Maybe not. I do not know. But either way, our preprint was published long before their paper was submitted and therefore I believe they should have discussed it in more detail.
Is this a sign that some people believe preprints are nothing more than rumors? I hope not. Preprints are a great way to share research prior to the delays that can happen in peer review. And in my opinion, preprints should count as prior research and be cited as such. I note – the Burton group in their paper in G3 also did not reference our preprint in what I consider to be a reasonable manner. They add some comments in their acknowledgements
While this manuscript was in preparation, a preprint describing a related method appeared in PeerJ PrePrints (Beitel et al. 2014a). Note added in proof: this preprint was subsequently published (Beitel et al. 2014b).
Given that our preprint was published before their paper was submitted too, I believe that they also should have made more reference to it in their paper. But again, I can only guess that both the Burton and the Marbouty group just do not see preprints as being respectable scientific objects. That is a bad precedent to set and I think the wrong one too. And it is a shame. A preprint is a publication. No – it is not peer reviewed. But that does not mean it should not be considered part of the scientific literature in some way. I note – this new paper from the Marbouty group seems really interesting. Not sure I want to dig into it any deeper if they are going to play games with the timing of submission vs. published “papers” as part of how they are positioning themselves to be viewed as doing something novel.
Crossposting from microBEnet:
Some new papers that may be of interest to people:
- FOAM (Functional Ontology Assignments for Metagenomes): a Hidden Markov Model (HMM) database with environmental focus.
- Statistical Methods for Functional Metagenomic Analysis Based on Next-Generation Sequencing Data PhD thesis from Pookhao, Naruekamol
- Profile Hidden Markov Models for the Detection of Viruses within Metagenomic Sequence Data
- Parallel-META 2.0: Enhanced Metagenomic Data Analysis with Functional Annotation, High Performance Computing and Advanced Visualization
- MP3: A Software Tool for the Prediction of Pathogenic Proteins in Genomic and Metagenomic Data
Here is another post in my “Story Behind the Paper” series where I ask authors of open access papers to tell the story behind their paper. This one comes from Rogan Carr and Elhanan Borenstein. Note – this was crossposted at microBEnet. If anyone out there has an open access paper for which you want to tell the story — let me know.
We’d like to first thank Jon for the opportunity to discuss our work in this forum. We recently published a study investigating direct functional annotation of short metagenomic reads that stemmed from protocol development for our lab. Jon invited us to write a blog post on the subject, and we thought it would be a great venue to discuss some practical applications of our work and to share with the research community the motivation for our study and how it came about.
Our lab, the Borenstein Lab at the University of Washington, is broadly interested in metabolic modeling of the human microbiome (see, for example our Metagenomic Systems Biology approach) and in the development of novel computational methods for analyzing functional metagenomic data (see, for example, Metagenomic Deconvolution). In this capacity, we often perform large-scale analysis of publicly available metagenomic datasets as well as collaborate with experimental labs to analyze new metagenomic datasets, and accordingly we have developed extensive expertise in performing functional, community-level annotation of metagenomic samples. We focused primarily on protocols that derive functional profiles directly from short sequencing reads (e.g., by mapping the short reads to a collection of annotated genes), as such protocols provide gene abundance profiles that are relatively unbiased by species abundance in the sample or by the availability of closely-related reference genomes. Such functional annotation protocols are extremely common in the literature and are essential when approaching metagenomics from a gene-centric point of view, where the goal is to describe the community as a whole.
However, when we began to design our in-house annotation pipeline, we pored over the literature and realized that each research group and each metagenomic study applied a slightly different approach to functional annotation. When we implemented and evaluated these methods in the lab, we also discovered that the functional profiles obtained by the various methods often differ significantly. Discussing these findings with colleagues, some further expressed doubt that that such short sequencing reads even contained enough information to map back unambiguously to the correct function. Perhaps the whole approach was wrong!
We therefore set out to develop a set of ‘best practices’ for our lab for metagenomic sequence annotation and to prove (or disprove) quantitatively that such direct functional annotation of short reads provides a valid functional representation of the sample. We specifically decided to pursue a large-scale study, performed as rigorously as possible, taking into account both the phylogeny of the microbes in the sample and the phylogenetic coverage of the database, as well as several technical aspects of sequencing like base-calling error and read length. We have found this evaluation approach and the results we obtained quite useful for designing our lab protocols, and thought it would be helpful to share them with the wider metagenomics and microbiome research community. The result is our recent paper in PLoS One, Comparative Analysis of Functional Metagenomic Annotation and the Mappability of Short Reads.
|The performance of BLAST-based annotation of short reads across the bacterial and archaeal tree of life using the ‘top gene’ protocol. See the manuscript for full details. Figure and text adapted from: Carr R, Borenstein E (2014) Comparative Analysis of Functional Metagenomic Annotation and the Mappability of Short Reads. PLoS ONE 9(8): e105776|
To perform a rigorous study of functional annotation, we needed a set of reads whose true annotations were known (a “ground truth”). In other words, we had to know the exact locus and the exact genome from which each sequencing read originated and the functional classification associated with this locus. We further wanted to have complete control over technical sources of error. To accomplish this, we chose to implement a simulation scheme, deriving a huge collection of sequence reads from fully sequenced, well annotated, and curated genomes. This schemed allowed us to have complete information about the origin of each read and allowed us to simulate various technical factors we were interested in. Moreover, simulating sequencing reads allowed us to systematically eliminate variations in annotation performance due to technological or biological effects that would typically be convoluted in an experimental setup. For a set of curated genomes, we settled on the KEGG database, as it contained a large collection of consistently functionally curated microbial genomes and it has been widely used in metagenomics for sample annotation. The KEGG hierarchy of KEGG Orthology groups (KOs), Modules, and Pathways could then serve as a common basis for comparative analysis. To control for phylogenetic bias in our results, we sampled broadly across 23 phyla and 89 genera in the bacterial and archaeal tree of life, using a randomly selected strain in KEGG for each tip of the tree from Ciccarelli et al. From each of the selected 170 strains, we generated either *every* possible contiguous sequence of a given length or (in some cases) every 10th contiguous sequence, using a sliding window approach. We additionally introduced various models to simulate sequencing errors. This large collection of reads (totaling ~16Gb) were then aligned to the KEGG genes database using a translated BLAST mapping. To control for phylogenetic coverage of the database (the phylogenetic relationship of the database to the sequence being aligned) we also simulated mapping to many partial collections of genomes. We further used four common protocols from the literature to convert the obtained BLAST alignments to functional annotations. Comparing the resulting annotation of each read to the annotation of the gene from which it originated allowed us to systematically evaluate the accuracy of this annotation approach and to examine the effect of various factors, including read length, sequencing error, and phylogeny.
First and foremost, we confirmed that direct annotation of short reads indeed provides an overall accurate functional description of both individual reads and the sample as a whole. In other words, short reads appear to contain enough information to identify the functional annotation of the gene they originated from (although, not necessarily the specific taxa of origin). Functions of individual reads were identified with high precision and recall, yet the recall was found to be clade dependent. As expected, recall and precision decreased with increasing phylogenetic distance to the reference database, but generally, having a representative of the genus in the reference database was sufficient to achieve a relatively high accuracy. We also found variability in the accuracy of identifying individual KOs, with KOs that are more variable in length or in copy number having lower recall. Our paper includes abundance of data on these results, a detailed characterization of the mapping accuracy across different clades, and a description of the impact of additional properties (e.g., read length, sequencing error, etc.).
|A principal component analysis of the pathway abundance profiles obtained for 15 HMP samples and by four different annotation protocols.The different protocols are represented by color and shape. See the manuscript for full details. Figure and text adapted from: Carr R, Borenstein E (2014) Comparative Analysis of Functional Metagenomic Annotation and the Mappability of Short Reads. PLoS ONE 9(8): e105776|
Importantly, while the obtained functional annotations are in general representative of the true content of the sample, the exact protocol used to analyze the BLAST alignments and to assign functional annotation to each read could still dramatically affect the obtained profile. For example, in analyzing stool samples from the Human Microbiome Project, we found that each protocol left a consistent “fingerprint” on the resulting profile and that the variation introduced by the different protocols was on the same order of magnitude as biological variation across samples. Differences in annotation protocols are thus analogous to batch effects from variation in experimental procedures and should be carefully taken into consideration when designing the bioinformatic pipeline for a study.
Generally, however, we found that assigning each read with the annotation of the top E-value hit (the ‘top gene’ protocol) had the highest precision for identifying the function from a sequencing read, and only slightly lower recall than methods enriching for known annotations (such as the commonly used ‘top 20 KOs’ protocol). Given our lab interests, this finding led us to adopt the ‘top gene’ protocol for functionally annotating metagenomic samples. Specifically, our work often requires high precision for annotating individual reads for model reconstruction (e.g., utilizing the presence and absence of individual genes) and the most accurate functional abundance profile for statistical method development. If your lab has similar interests, we would recommend this approach for your annotation pipelines. If however, you have different or more specific needs, we encourage you to make use of the datasets we have published along with our paper to help you design your own solution. We would also be very happy to discuss such issues further with labs that are considering various approaches for functional annotation, to assess some of the factors that can impact downstream analyses, or to assist in such functional annotation efforts.
Interesting discussion yesterday with authors of new pathogen discovery papers. I will try to write more about this later but am heading out the door so this Storify will have to do for now.
Just got this email and thought it would be of interest to many out there.
Thank you for being a CAMERA user during its operation as a resource for
environmental genomics. During the past few years, CAMERA has been able to
offer a number of important community resources, including an exceptionally
well curated environmental genomic database, the ability for researchers to
deposit molecular sequence datasets with associated environmental
parameters (metadata), open access to computational resources to enable
metagenomic comparisons, educational resources, and helpdesk services.
These efforts have been funded through the Gordon and Betty Moore
Foundation (GBMF) Marine Microbiology Initiative and the National Science
Foundation to serve the needs of the marine microbiology community and
In particular, the CAMERA compute resources, which include large-scale
BLAST capabilities and other workflow-enabled analysis capabilities
(RAMMCAP), were generously supported by the GBMF, the San Diego
Supercomputer Center, the NSF XSEDE program, and commercial Cloud computing
resource providers (CODONiS).
Due to the termination of GBMF support, CAMERA can no longer accommodate
the computational needs of the community. Therefore, starting July 1, 2014,
CAMERA will begin to shut down the CAMERA portal and will no longer accept
any new workflow submissions. The results of workflows submitted by July 1,
2014 will be available to users through July 15th. Urgent requests for the
temporary use of CAMERA workflow resources beyond July 1, 2014 will be
considered on a case-by-case basis.
If you are a current or prior CAMERA user and would like to retrieve
personal data from the system, we strongly encourage you to do so now.
As announced earlier this year, CAMERA will continue to maintain free and
open access to its rich collection of curated data and metadata via the
CAMERA Data Distribution Center (DDC), which includes links to the Marine
Microbial Eukaryote Transcriptome Sequencing Project. In conjunction with
the portal shutdown, CAMERA will also no longer accept user data
submissions past July 1st (but data submissions currently in progress will
be completed and made available via the DDC).
Please contact us at email@example.com regarding matters pertaining to
the use of CAMERA.
Thank you so much for your support and participation!
Well, I am still really annoyed by this unbearable article on C-Net yesterday: IBM sees big opportunity in sequencing microbes by Daniel Terdiman. The article is about this “Sequencing the City” meeting organized by IBM that was on Tuesday and Wednesday. I talked at the meeting on Tuesday (I could not go on Wednesday). For more about my talk see: What to do when you realize the meeting you are speaking at is a YAMMM (yet another mostly male meeting)?. But I am not criticizing the meeting here. I am criticizing the article in C-Net which has many many flaws. For example consider:
According to James Kaufman, a research manager at the Almaden Research Center, the move to study metagenomics — the study of systems of micro-organisms — came from what he called a tipping point in big data. As more and more government-funded institutions study organisms and bacteria, they’ve collected more information about them, and submitted much of their work to centralized databases. “So there’s a growing library of genomes across the field of life,” Kaufman said. “That made possible metagenomics.”
What? Metagenomics has been around for a long time. Sure, many people in the field are taking advantage of so called big data, but there was no “tipping point” needed to launch the field. This is just completely misguided.
And then even worse
The result: We can now look at and understand whole ecosystems at the bacterial level. One example of how that manifests is what IBM refers to as the Human Microbiome Project. According to an IBM document, that’s about characterizing “microbial communities found at multiple human body sites to discover correlations between changes in the microbiome with changes in human health.”
So – there have been dozens of high profile papers from the Human Microbiome Project. There are hundreds of web pages with information about the project. It was started years and years ago. And the reporter quotes an “IBM document” to tell us what the Human Microbiome Project is? And even worse the reporter says “what IBM refers to as the Human Microbiome Project” like they ran it / designed it. Good that they refer to it as the Human Microbiome Project. You know why? Because that is what it is known as to all the other $(&@)(* people in the whole (%&# world.
The reporter goes on to write
This kind of work is not entirely new, but the scientists who will be gathering at IBM Research this week are grappling with one conundrum: they don’t know what they don’t know. So a big topic of conversation, and a big part of what IBM would like to see advanced, is “the ability to do metagenomics on the scale of a city or the world….That will depend on software services available in the cloud,” Kaufman said. “It has to be cheap, easy, and accessible from anywhere. That’s what we’re really good at.”
Once again making it seem like IBM is somehow leading this field. Not to pick on IBM here. I am glad they organized the meeting. But either the reporter just got handed a press release from IBM and wrote it up, or did not do any type of background research, or both. Sure IBM would like to see this. But so would lots of other people. Why make this all about IBM? There are so many people who have done interesting work in the area of “microbiology of the built environment” – why are none of them even discussed? What exactly is the point of this article if not to simply be a PR piece for IBM? Aaaaaarg.
UPDATE 5/9 Storify of some of the Tweets about the meeting
Doing another mini journal club here. Just got notified of this paper through some automated Google Scholar searches: Gnotobiotic mouse model of phage–bacterial host dynamics in the human gut
Full citation: Reyes, A., Wu, M., McNulty, N. P., Rohwer, F. L., & Gordon, J. I. (2013). Gnotobiotic mouse model of phage–bacterial host dynamics in the human gut. Proceedings of the National Academy of Sciences, 201319470.
The paper seems pretty fascinating at first glance. Basically they built on the Jeff Gordon germ free mouse model and introduced a defined set of cultured microbes that came from humans. And then they stages a phage attack on the system and monitored the response of the community to the phage attack.
|Figure 1 from Reyes et al.|
They (of course) also did a control – in this case with heat killed phage. And they compared what happened to the live phage. I love this concept as they are able to control the microbial community and then test dynamics of how specific phage affect that community inside a living host. Very cool.
I just love this paper … The human gut and groundwater harbor non-photosynthetic bacteria belonging to a new candidate phylum sibling to Cyanobacteria | eLife from the labs of Ruth Ley and Jill Banfield (
1st author is the co-first authors are Sara C. Di Rienzi and Itai Sharon). It represents a landmark study in something that has intrigued many microbial diversity / human microbiome researchers for many years. Early in the history of sequencing rRNA genes from human microbiome samples, researchers discovered something a bit weird – quite a few sequences were coming from what appeared to be close relatives of Cyanobacteria. This was weird because all known Cyanobacteria were thought to be photosynthetic and – well – there is not too much light in the human gut.
Now – one possible explanation for this was that these sequences were coming from photosynthetic bacteria but these bacteria were not residents of the human gut but came via consumable items (i.e., food and drink). Perhaps they were actually from chloroplasts of something in the diet (after all – chloroplasts are derived versions of cyanobacteria). This idea was discussed at many meetings I attended. But there was no evidence for this. Another possibility was that there was in fact some light in the human gut – leaking through from the outside or being produced from the inside. And perhaps this was enough to do a little photosynthesis. Sound crazy? Well, not so crazy after reports of photosynthesis in the deep sea. A third possibility was that these sequences were coming from residents of the human gut that were related to (or even within) cyanobacteria but were not photosynthetic. More detail on possible explanations are in this new paper and in some of the material cited therein.
Anyway – Ruth Ley has been discussing these unusual sequences for years and now in this paper her group and the group of Jill Banfield at Berkeley (along with some others) has used metagenomics and detailed assembly and phylogenetic analysis to reveal many new insights into these sequences. I could write much more about this. But, I think the paper really speaks for itself. And it is open access so anyone and everyone can check it out. And you should. It is wonderful.
|Fig 2 from Di Rienzi et al.|
UPDATED 10/9/2013 to correct that there were co-first authors
OK I have been dreaming about doing something like this for many years. One of the potentially most useful aspects of shotgun metagenomic data is that you get a sample of many/all members of a microbial community at once. And then in theory one could look across different species and taxa and ask – do they all have similar adaptations in response to some sort of environmental pressure. There have been a few papers on this over the last few years (e.g. check out this one from Muegge et al on Diet Driving Convergence in Gut Microbes). But this new paper is really the type of thing I have been hoping to see: Environmental shaping of codon usage and functional adaptation across microbial communities. Basically they looked at codon usage in organisms in different metagenomic samples and found major metagenome specific signatures, suggesting that different taxa were in essence converging on common codon usage.
The paper is definitely worth a look.