ArticlePDF AvailableLiterature Review

Abstract and Figures

Innovations in sequencing technologies have allowed biologists to make incredible advances in understanding biological systems. As experience grows, researchers increasingly recognize that analyzing the wealth of data provided by these new sequencing platforms requires careful attention to detail for robust results. Thus far, much of the scientific Communit’s focus for use in bacterial genomics has been on evaluating genome assembly algorithms and rigorously validating assembly program performance. Missing, however, is a focus on critical evaluation of variant callers for these genomes. Variant calling is essential for comparative genomics as it yields insights into nucleotide-level organismal differences. Variant calling is a multistep process with a host of potential error sources that may lead to incorrect variant calls. Identifying and resolving these incorrect calls is critical for bacterial genomics to advance. The goal of this review is to provide guidance on validating algorithms and pipelines used in variant calling for bacterial genomics. First, we will provide an overview of the variant calling procedures and the potential sources of error associated with the methods. We will then identify appropriate datasets for use in evaluating algorithms and describe statistical methods for evaluating algorithm performance. As variant calling moves from basic research to the applied setting, standardized methods for performance evaluation and reporting are required; it is our hope that this review provides the groundwork for the development of these standards.
This content is subject to copyright.
REVIEW
published: 07 July 2015
doi: 10.3389/fgene.2015.00235
Edited by:
Alexandre V. Morozov,
Rutgers University, USA
Reviewed by:
Claus J. Scholz,
University of Würzburg, Germany
Cuncong Zhong,
J. Craig Venter Institute, USA
*Correspondence:
Nathan D. Olson,
Biosystems and Biomaterials Division,
Material Measurement Laboratory,
National Institute of Standards
and Technology, 100 Bureau Dr.,
Gaithersburg, MD 20899, USA
nolson@nist.gov
Present address:
Jeffrey T. Foster,
Department of Molecular, Cellular,
and Biomedical Sciences, University
of New Hampshire, Durham, NH, USA
Specialty section:
This article was submitted to
Bioinformatics and Computational
Biology,
a section of the journal
Frontiers in Genetics
Received: 05 May 2015
Accepted: 22 June 2015
Published: 07 July 2015
Citation:
Olson ND,
Lund SP, Colman RE, Foster JT, Sahl
JW, Schupp JM, Keim P, Morrow JB,
Salit ML and Zook JM (2015) Best
practices for evaluating single
nucleotide variant calling methods
for microbial genomics.
Front. Genet. 6:235. doi:
10.3389/fgene.2015.00235
Best practices for evaluating single
nucleotide variant calling methods
for microbial genomics
Nathan D. Olson 1*, Steven P. Lund 2, Rebecca E. Colman 3, Jeffrey T. Foster 4,
Jason W. Sahl3,4, James M. Schupp 3, Paul Keim 3,4 , Jayne B. Morrow 1, Marc L. Salit 1,5 and
Justin M. Zook 1
1Biosystems and Biomaterials Division, Material Measurement Laboratory, National Institute of Standards and Technology,
Gaithersburg, MD, USA, 2Statistical Engineering Division, Information Technology Laboratory, National Institute of Standards
and Technology, Gaithersburg, MD, USA, 3Division of Pathogen Genomics, Translational Genomics Research Institute,
Flagstaff, AZ, USA, 4Center for Microbial Genetics and Genomics, Northern Arizona University, Flagstaff, AZ,
USA, 5Department of Bioengineering, Stanford University, Stanford, CA, USA
Innovations in sequencing technologies have allowed biologists to make incredible
advances in understanding biological systems. As experience grows, researchers
increasingly recognize that analyzing the wealth of data provided by these new
sequencing platforms requires careful attention to detail for robust results. Thus far, much
of the scientific Communit’s focus for use in bacterial genomics has been on evaluating
genome assembly algorithms and rigorously validating assembly program performance.
Missing, however, is a focus on critical evaluation of variant callers for these genomes.
Variant calling is essential for comparative genomics as it yields insights into nucleotide-
level organismal differences. Variant calling is a multistep process with a host of potential
error sources that may lead to incorrect variant calls. Identifying and resolving these
incorrect calls is critical for bacterial genomics to advance. The goal of this review is
to provide guidance on validating algorithms and pipelines used in variant calling for
bacterial genomics. First, we will provide an overview of the variant calling procedures
and the potential sources of error associated with the methods. We will then identify
appropriate datasets for use in evaluating algorithms and describe statistical methods
for evaluating algorithm performance. As variant calling moves from basic research to
the applied setting, standardized methods for performance evaluation and reporting are
required; it is our hope that this review provides the groundwork for the development of
these standards.
Keywords: next-generation sequencing, variant calling, single nucleotide variants, indel, performance metrics
Introduction
Next-generation sequencing (NGS) has transformed microbiology, making genomic analyses
possible for a broad range of species. However, converting millions of sequencing reads per sample
into meaningful data is not trivial, and genome assembly, sequence alignment, and variant calling
can all have substantial effects on results. Considerable effort has been spent addressing these issues
in human genomes, with a primary goal of finding genomic variations linked to human disease.
Variant calling from microbial genomes presents additional challenges, such as reference genome
Frontiers in Genetics | www.frontiersin.org July 2015 | Volume 6 | Article 2351
Olson et al. SNP evaluation review
selection, presence of rare variants within a culture, and use
of de novo genome assemblies for variant calling. Due to the
diversity of methods used to call variants from microbial genomes,
optimization of bioinformatics methods for specific organisms
and/or experiments is frequently required.
Variant calling can include single nucleotide polymorphisms
(SNPs), insertions and deletions (indels), and/or structural
variants. Here, we focus on SNPs and indels. Both SNP and indel
calling methods identify genome positions with polymorphisms
relative to a reference (for review, see Nielsen et al., 2011). SNP
and indel calling is achieved by either mapping reads directly to
the reference genome or generating a de novo genome assembly
from the reads and subsequently comparing the assembly to the
reference genome.
For SNP discovery, using raw reads can provide greater
resolution than using a genome assembly. With raw reads, both
the depth of coverage as well as the proportion of mixed alleles
can be quantified, in contrast to creating an assembly, in which
all coverage at a given locus is collapsed into a single base call.
When the raw reads are available they can be mapped back
to the assembly to obtain the coverage and allelic proportions.
However, lack of a closed reference genome may cause biases in
allelic proportions due to mapping errors. Updates to the sequence
alignment format (SAM/BAM) to accommodate storing de novo
sequence alignments can resolve this issue (Cock et al., 2015). For
reference-based SNP discovery, reads can fail to align to regions
of high divergence (Bertels et al., 2014) if the short read aligner is
too stringent. This can be resolved with some assembly-based SNP
discovery methods, as regions can be aligned from more divergent
sequences, allowing for the characterization of SNPs even from
different bacterial species within the same genus (Sahl et al., 2013).
Single nucleotide polymorphism identification is an important
method for bacterial comparative genomics. SNP-based analyses
have been used for outbreak attribution (Hendriksen et al., 2011),
phylogeography (Keim and Wagner, 2009), and genome-wide
association studies (GWAS; Nelson et al., 2014); SNPs have also
been used extensively in human GWAS applications (Cantor
et al., 2010). Multiple types of error, which must be explored
and accounted for to understand the evolution and relatedness
of bacterial genomes, affect SNP-based bacterial comparative
genomics.
Both similarities and differences exist between human and
microbial variant calling. The human reference genome is 1000
times larger than an individual microbial genome, and humans
are diploid whereas microbial genomes are generally haploid.
Therefore, the assumptions and statistics used in human variant
callers often are not optimal for microbial genomes. The smaller
number of mutations in a bacterial genome means that some
machine learning methods used to filter potential false positives
(FP) in humans will not work well for microbial genomes,
and different filtering thresholds are often needed depending
on the microbial sequencing application. In addition, some
microbial genomes have high mutation rates, so that they may
be heterogeneous, with only a small fraction of cells containing
a mutation. In this case, they are more similar to somatic variant
calling in human cancer cells. All of these differences mean that
validating variant callers for microbial sequencing should be done
FIGURE 1 | SNP calling workflow diagram. Horizontal boxes represent
steps in the workflow and arrows to the left indicate steps in the workflow
challenged with reference genomic DNA, and sequence data.
even if the same variant caller has been validated for human
sequencing.
Selecting the optimal variant calling method can depend on
the application, organism, and sequencing data (Koboldt et al.,
2009). Currently, there are no widely accepted guidelines for
evaluating SNP and indel calling methods. The lack of guidelines
has resulted in a diverse, inconsistent, and difficult-to-interpret
body of literature on SNP and indel calling method performance.
To help address this issue, we set out to provide general
guidelines for evaluating SNP and indel calling methods. We
briefly discuss SNP and indel calling procedures from NGS data
and describe the associated errors. In order to develop guidelines
for SNP and indel calling method evaluation, we discuss and
identify appropriate data for use in evaluation, present statistical
methods for evaluation, and methods for comparing variant
call sets.
SNP and Indel Calling Procedure
In order to draw meaningful conclusions from evaluation of
variant calling methods, the process used to identify the variants
must first be understood. This measurement process includes
sample processing, sequencing, mapping or de novo assembly,
followed by variant calling (Altmann et al., 2012,Figure 1).
Frontiers in Genetics | www.frontiersin.org July 2015 | Volume 6 | Article 2352
Olson et al. SNP evaluation review
Sample Processing and Sequencing
DNA is extracted from a bacterial culture, purified, and during
library preparation, adapters and unique barcodes (indices) are
added prior to sequencing.
The resulting library or pooled libraries are sequenced (for
reviews of NGS platforms see Metzker, 2010; Pabinger et al., 2013).
Base calls are automatically generated by the sequencing
platform. For each base call, algorithms assign a base quality score
(BQS), which is intended to reflect the probability that the base
was called correctly. BQS values are provided on the phred scale
(1–60), as the 10 logP, where P is the probability of an incorrect
base call (Ewing and Green, 1998; Pavlopoulos et al., 2013). For
example, a BQS of 20 indicates a 1 in 100 chance the base was
called incorrectly. BQS commonly inform variant calling software
when assigning variant quality scores (also typically expressed on
a phred scale).
Sequence Processing and SNP Calling
After the raw reads are generated by the sequencing platform,
reads can be mapped to a reference genome, de novo assembled,
or mapped and assembled in one simultaneous step (Zhong et al.,
2015). SNPs and indels relative to a reference are then identified
by comparison to a reference genome. Commonly used read
mapping and de novo assembly based approaches are discussed
in this review.
SNP Calling Following Read Alignment to a Reference
Mapping reads to a reference
To map reads to a reference genome, the location of each read,
relative to the reference genome, is predicted. To assess the
performance of short read mapping tools, studies have evaluated
different algorithms performance of short read alignment tools
(e.g., Fonseca et al., 2012; Schbath et al., 2012; Hatem et al.,
2013; Nagarajan and Pop, 2013). Reads are initially mapped and
then a series of steps including duplicate read removal, BQS
recalibration, and indel realignment are performed to refine the
read alignments prior to variant calling.
Most mapping programs assign phred scale mapping quality
scores to indicate the confidence that the reads are accurately
placed in relation to the reference (Gallagher and Desjardins,
2008). Mapping quality scores generated by different mapping
algorithms are generally not comparable to each other, but in
general, lower mapping quality scores are assigned to shorter reads
as well as reads mapping to multiple regions genome (e.g., repeat
regions).
The resulting mapped file, usually the binary version (BAM)
of a Sequence Alignment/Map (SAM) file, then undergoes addi-
tional processing to reduce variant calling errors. These include
duplicate read removal, BQS recalibration, and realignment
around indels. Duplicate reads originating from the DNA
sequence can arise as artifacts of the PCR step in the sequencing
library preparation process, or when the same fragment is read
twice (i.e., optical duplicate). As such, duplicates can lead to an
artificial increase in bases supporting variant calls, leading to an
erroneous increase in variant call confidence, andshould therefore
be removed either before or after the mapping step (DePristo et al.,
2011). For newer low cycle and “PCR-free” library preparation
methods, read duplication is less likely to occur.
Recalibrating BQS can help improve quality score accuracy and
thus, the accuracy of called variants (DePristo et al., 2011; Zook
et al., 2012). To recalibrate, the sequencing error rate for positions
in the reference genome that are known with high confidence are
compared to the sequencing platform assigned BQS. The BQS
for the entire data set are recalibrated based on differences in the
assigned BQS and the observed base call error rates for the known
positions.
Indels and other structural variants can lead to incorrect
sequence read mapping, causing false negative (FN) and
positive SNP and indel calls (Alkan et al., 2011). Therefore,
algorithms have been developed to increase the accuracy of
read mapping in these regions through read realignment (e.g.,
GATK IndelRealigner) or local de novo assembly (e.g., GATK
HaplotypeCaller), in turn reducing the number of incorrect
variant calls in these portions of the genome (Homer and Nelson,
2010; DePristo et al., 2011).
Calling SNPs using mapped read data
Finally, variant calling algorithms compare mapped reads to the
reference genome and identify potential variants. SNP and indel
calling algorithms vary in their approach to identifying candidate
variants (Altmann et al., 2012). Basic algorithms identify variants
based the on the number of high confidence base calls that
disagree with the reference bases for the genome position of
interest. More sophisticated algorithms commonly use Bayesian,
likelihood, or machine learning statistical methods that factor
parameters, such as base and mapping quality scores, to identify
candidate variants. See Pabinger et al. (2013) for a review of
different variant calling algorithms. The presumptive SNPs and
indels identified by the variant caller can be filtered using a
number of parameters associated with systematic errors discussed
in the next section, thereby reducing the number of FP variant
calls, but risking an increase in FN calls.
SNP Calling Using de novo Assemblies
De novo genome assembly
Many applications in comparative genomics, such as pan-genome
comparisons, operon structure determination, or genome synteny
in a population, require de novo genome assembly. For short
read de novo assembly, de Bruijn graph methods are typically
used (Chaisson and Pevzner, 2008), although overlap layout
consensus can also be used effectively (Loman et al., 2012). In both
methods, using paired-end and/or mate pairs can increase the
contiguity of an assembly by facilitating the generation of genomic
scaffolds, which link contigs based on insert length. For graph-
based methods, the choice of k-mer size (base string length) can
affect the contiguity and/or completeness of an assembly (Chikhi
and Medvedev, 2014). For some de novo assembly methods,
such as Velvet (Zerbino and Birney, 2008), a single k-mer
value is frequently chosen. For newer methods, such as SPAdes
(Bankevich et al., 2012) or IDBA (Peng et al., 2010), assemblies
from a range of k-mer values are merged, limiting the amount
of sequence lost from the assembly. SPAdes also incorporates the
Frontiers in Genetics | www.frontiersin.org July 2015 | Volume 6 | Article 2353
Olson et al. SNP evaluation review
FIGURE 2 | Cause-effect diagram indicating the sources of error associated with different steps in the variant calling measurement process. Note that
the SNP calling is performed using one of two methods, either read mapping or de novo assembly.
BayesHammer short read correction tool (Nikolenko et al., 2013),
which in combination with mapping reads back to the assembly to
fix errors, can reduce the error profile in downstream applications.
In addition, quality filtering to remove low quality regions from
raw data has been demonstrated to improve the overall quality of
genome assemblies (Del Fabbro et al., 2013).
Multiple bioinformatics pipelines have been published for de
novo assembly, yet significant performance variation has been
observed (Magoc et al., 2013). One of the most important
discordances among assemblers is the amount of the assembly
retained, based on benchmark comparisons using completed
genomes. As sequencing platforms that generate longer reads
become more widespread, completed bacterial genomes will
continue to be automatically generated (Koren and Phillippy,
2015), removing the limitations when using incomplete draft
assemblies. Until that time, short read assemblers should be
chosen based on their completeness of draft assembly to reduce
errors in SNP calling based on the presence or absence of
homologous genomic regions.
Calling SNPs using a genome assembly
SNPs can be identified from genome assemblies, however,
since coverage is 1x at each position in an assembly, spurious
SNPs cannot be filtered due to insufficient coverage, nor can
contaminating genomes be identified and subsequently removed.
For individual genes, SNPs are identified by extracting alignments
using BLASTN (Altschul et al., 1990) followed by pairwise
alignment of the SNPs. For whole genome assemblies, SNPs are
typically identified from whole genome alignments made with
software such as MUMmer (Kurtz et al., 2004), Mugsy (Angiuoli
and Salzberg, 2011), and Mauve (Darling et al., 2004). Software has
also been developed for the identification of SNPs from genome
assemblies for whole genome phylogenetics including kSNP
(Gardner and Hall, 2013) and parSNP (Treangen et al., 2014).
SNP identification using assemblies is useful when analyzing
individual genes, processing huge datasets, or if raw reads are
unavailable. However, when using assemblies for SNP discovery,
SNPs cannot be evaluated and verified with the underlying raw
read data.
Sources of Error and Mitigation Strategies
Understanding the types and sources of error associated with
a SNP and indel calling procedure will not only facilitate
evaluation of results but also enable the user to optimize method
performance. Several types of errors can impact the accuracy
of SNP and indel variant identification. These errors occur
during sample processing, the chemical and electronic processes
that occur during sequencing, as well as the bioinformatic
processing of sequence data: base calling, read mapping or de
novo assembly, and identification of SNP and indel variants
(Nielsen et al., 2011). The sources of error associated with
different steps in the measurement process are depicted in
Figure 2.
Frontiers in Genetics | www.frontiersin.org July 2015 | Volume 6 | Article 2354
Olson et al. SNP evaluation review
The errors occurring during the above processes can be random
or systematic in nature. Random errors occur in an unpredictable
manner, but given a large enough sample size may not lead
to inaccurate results. Systematic errors occur in a predictable
manner and if not accounted for, may lead to inaccurate results.
Additionally, systematic errors, when unaccounted for, often
result in a bias or a predictable difference between the true value
and a measurement result.
The relative impact of the error can vary based on the sample
preparation method, sequencing platform, or bioinformatic
analyses used. The following section discusses the types of errors
that may occur, along with possible mitigation strategies to
minimize errors associated with FP or negative variant calls.
Errors Associated With Sample Processing
As most NGS platforms require amplification of the genetic
material, random and systematic sequence errors associated
with sequence library preparation are primarily due to DNA
polymerase infidelity during replication. These replication errors
result in improper bases being inserted, deleted, or substituted
at any given position (Ross et al., 2013). Substitution errors are
random and subject to the fidelity of the DNA polymerases (1
error in 103–6 bases, Showalter and Tsai, 2002), but can occur at
a rate of up to 1,000–10,000 times lower than typical sequencing
error (1 in 102–3 bases McElroy et al., 2014). With sequencing
depth commonly greater than 20X, random polymerase errors
have minimal impact on variant calls (Cline et al., 1996; Metzker,
2010; Zook et al., 2012) and are usually only of concern when
detecting minor component variants. Most current sequencing
library preparation systems involve the use of high fidelity
polymerases, further minimizing polymerase errors. However,
homopolymers (sequence of identical bases, e.g., AAAA or
TTTT) and tandem repeats (adjacent repeated patterns of two
or more nucleotides) are known to experience higher replication
indel error rates. Indel replication error rates, depending upon
size and type of repeat array, can approach sequencing error
rates (Vogler et al., 2006). Both substitution and indel polymerase
errors accumulate linearly with increasing cycle number during
amplification and then exponentially if erroneous amplification
products become the primary template for subsequent cycles
(PCR duplicates), at which point they may significantly contribute
to variant call error (Kozarewa et al., 2009). Hence, minimizing
PCR amplification, or foregoing it altogether, will minimize
polymerase-introduced errors.
The use of paired-end library preparation methodology
can also reduce variant calling error. Standard paired-end
methodology (200–500 bp inserts between sequence reads) can
help reduce errors during read mapping. Overlapping paired-end
methodology (partial/complete overlap of forward and reverse
reads from the same molecule) can provide a filter for removing
random substitution error (i.e., only variants on both strands are
called), thereby improving the ability to detect rare substitution
variants by orders of magnitude (Schmitt et al., 2012; Chen-
Harris et al., 2013; Colman et al., 2015). An even more recent
advance is the circle sequencing” approach, in which templates
are circularized, copied via rolling circle amplification, and
sequenced. This approach has led to exceptionally low error rates
while maintaining relatively high sequencing yields (Lou et al.,
2013).
Errors Associated With Sequencing
Sequence generation errors are dependent on the sequencing
platform and can be both random and systematic in nature,
with the latter leading to local “hot spots” for high error rates
(Ross et al., 2013). These have been found to be as high as
6% error for Illumina and 50% for Roche-454 (McElroy et al.,
2014). The Illumina platform chemistry utilizes reversible dideoxy
terminator nucleotides labeled with fluorescent dyes that have
some degree of overlap in their excitation and emission spectra.
While the overlapping excitation spectra allow for the use of
only two lasers for all four nucleotides, the overlapping emission
spectra, particularly for the G and T channels, and may result
in base calling errors (Meacham et al., 2011). In fact, a number
of systematic sequencing errors have been described for the
various sequencing platforms (Ross et al., 2013). Some systematic
sequencing errors have distinct characteristics, such as strand bias,
which can be used to distinguish them from likely true variant
calls. Other systematic errors are not well characterized, thus are
more problematic, even with higher sequencing coverage (Shen
et al., 2010), and warrant further investigation.
Errors Associated With Sequence Processing
After the sequence data are generated, and as alluded to earlier,
additional errors may occur during the mapping of the sequence
reads to a reference genome or de novo assembly, and/or during
variant calling.
SNP Calling Errors When Mapping Reads to a
Reference
Read mapping errors
The two most common sources of true read mapping errors are
genomic duplication and structural variation (Alkan et al., 2011).
If multiple regions contain only slightly divergent sequences, then
reads can potentially map incorrectly, leading to FP variant calls
at those positions. Reads that map to duplicated” regions in
the reference genome are usually given a low mapping quality
score by the algorithm, which are typically filtered. As previously
mentioned, the use of paired-end libraries can reduce duplicate
mapping errors if one member of the read pair maps to a unique
location, thereby acting as an anchor for the proper mapping of
the paired read to the correct duplicated region. Filtering out
ambiguously mapped reads can mitigate FP variant calls due to
mapping errors, but may also remove correctly mapped reads with
true variants, potentially resulting in FN variant calls.
Alignment errors can also occur in regions containing small
indels or structural variation if reads are mapped to the correct
location but misaligned or allowed to extend into the structural
variant, resulting in FP or FN variant calls (Subramanian et al.,
2013). Optimizing the mapping criteria (e.g., k-mer size, number
of ambiguous bases per k-mer allowed, etc.) for the organism
under study can help mitigate the above errors. In general, highly
diverse regions of the genome are more prone to mapping and
alignment errors than lower diversity regions (Nielsen et al., 2011).
Frontiers in Genetics | www.frontiersin.org July 2015 | Volume 6 | Article 2355
Olson et al. SNP evaluation review
SNP calling errors using mapped reads
In an attempt to minimize variant calling errors, many variant
calling algorithms calculate statistics such as strand bias, base
quality rank sum, and neighboring base quality. In addition,
Bayesian statistics may be used to incorporate the mapping quality
scores assigned by the mapping algorithm (Li, 2011). These
statistics can be used to filter or remove FP variants, as discussed
below (McKenna et al., 2010; Meacham et al., 2011; Zook et al.,
2014). However, base call and mapping quality scores are not
always strongly associated with many systematic sequencing, local
alignment, or mapping errors (Dohm et al., 2008).
To remove likely FP variants, many variant callers often use
additional filters such as a minimum depth of coverage threshold,
base call frequency (e.g., >90% of calls at a position being
identical), masking of homopolymer and repetitive/duplicated
sequence regions, and trimming of poor quality bases from the
ends of reads. These filters can be hard filters with carefully chosen
cut-offs, or can be chosen by machine learning algorithms if
sufficient training data are available (e.g., human genomes). One
such algorithm is the GATK Variant Quality Score Recalibration
(DePristo et al., 2011). It is important to recognize that real
variants can be filtered out with FP, leading to higher FN rates.
To reduce FN, manual inspection of reads mapped to filtered
variants can be used to help identify true variants, thus reducing
the number of FN variant calls (Zook et al., 2014). Also of note,
strand bias can be defined in two ways: (1) a disproportionate
number of reads from only one strand mapping to the reference
or (2) a disproportionate number of reads with variant bases from
only one strand. The second definition is particularly useful for
identifying systematic sequencing errors caused by a particular
motif (e.g., the GGT to GGG error with Illumina chemistries,
which only occurs in reads in one sequencing direction; Meacham
et al., 2011; Ross et al., 2013).
Variant Calling Errors When Using de novo
Assemblies
De novo assembly errors
When using genome assemblies for variant detection, errors can
be introduced in multiple ways, including the intrinsic error rate
attributed to each sequencing platform. One approach to limit
the effects of the errors inherent with short read chemistries is to
use short read error correctors such as Musket (Bian et al., 2013)
and Hammer (Medvedev et al., 2011) prior to genome assembly.
Following assembly, these errors can be difficult or impossible
to identify (Baker, 2012). After the genome is assembled,
systematic errors can be corrected using bioinformatics tools
such as the PAGIT pipeline (Swain et al., 2012). The recently
published Pilon pipeline (Walker et al., 2014) can correct both
SNPs and short insertions/deletions, and can also identify and
fix incorrectly joined contigs. These best practices to reduce
assembly errors can reduce their effect on downstream SNP
applications.
Variant calling from de novo assemblies
Assembly errors are not the only form of error that can affect
SNP calling applications using genome assemblies. For example,
the whole genome alignment step can introduce errors. One of
the most commonly used whole genome alignment methods is
MUMmer (Delcher et al., 2002), which includes the nucmer
program for aligning nucleotide sequences (Kurtz et al., 2004).
By default, nucmer will align through large stretches of SNPs
potentially introduced by misjoined sequence, or through artifact
sequence introduced into the assembly (e.g., adapter sequence).
Although these artifacts can typically be identified and removed,
they can potentially introduce large numbers of erroneous
variants into an analysis. Mo difying nucmer parameters to specific
datasets can help diminish the background noise introduced by
these artifacts.
Another source of error for variant discovery in genome
assemblies is the incorporation of homopolymer stretches, which
are common to specific sequencing platforms (Loman et al., 2012).
The incorporation of homopolymers into genome assemblies
could generate incorrect variant calls, especially in the case of
indels. Therefore, indels that are composed of homopolymers
need to be verified with direct read mapping.
Additional errors may occur when calling variants that could be
part of a mixture of multiple alleles. During assembly, a mixture
of bases at a given position is collapsed down into a single base
call. Although some of these errors can be corrected with methods
discussed above, not all can be corrected and will serve as a source
of error in downstream applications.
General Guidelines for SNP Calling Method
Evaluation
Defining analytical requirements is critical to method evaluation.
The accuracy requirements of SNP and indel calling methods vary
by application. For example, a small number of inaccurate SNP
and indel calls may not alter phylogenetic inferences made for
samples when a large number of SNPs and indels are available
for analysis (Harris et al., 2013). However, for smaller total
numbers of variants, individual SNPs and indels can have a greater
impact on the phylogenetic interpretation. For example, for the
investigation of the 2011 E. coli O104:H4 outbreak in Germany,
only 19 SNPs were used to differentiate the isolates (Grad et al.,
2012). Similarly, isolates from the Bacillus anthracis Ames strain
and Haitian Vibrio cholerae outbreak investigations were based on
relatively few SNPs (Hendriksen et al., 2011; Rasko et al., 2011),
thus, accurate SNP calls were essential. Currently, the impact of
individual variant calls on a phylogenetic analysis in relation to the
total number of variants used to generate the analysis is unknown,
but warrants further exploration.
A few general principles can help guide method evaluation.
First, benchmarking datasets or samples used for evaluation
must be critically evaluated with credible accuracy and be
representative of the range of sample types used in the
study or application (Ellison and Williams, 2012). Whenever
possible, evaluation should be performed with data from
well-characterized genomic DNA reference materials (RMs)
or plasmid standards, if available and representative. Second,
there must be one or more metrics for evaluating algorithm
performance.
Frontiers in Genetics | www.frontiersin.org July 2015 | Volume 6 | Article 2356
Olson et al. SNP evaluation review
Selection of Material for use in Benchmarking
Selection of appropriate samples, datasets, and reference genome
is critical to robust characterization and evaluation of a variant
calling measurement process.
Benchmarking Samples
The use of well-characterized genomic RMs allows for evaluation
of the variant calling measurement process from sample
processing through variant calling (Figure 1). Furthermore,
to evaluate variant calling methods, the same data set must
be processed by each of the different analysis pipelines being
evaluated. The use of the same dataset is problematic because
many variant callers are specific to a single sequencing platform
or aligner, therefore evaluation of these variant callers using
datasets generated by other platforms or aligners would lead to
a biased representation of the caller’s performance. Therefore,
the only way to perform complete and adequate comparisons is
to use the same RM to generate data from different sequencing
platforms and chemistries that the variant calling method will
encounter.
A number of genomic RMs, along with respective reference
sequence data, are available or are being developed. The genome
sequences for currently available microbial genomic DNA RMs
have not yet been rigorously characterized, to the authors
knowledge. The National Institute of Standards and Technology
(NIST) is currently developing whole genome microbial and
human DNA NIST RMs (Zook et al., 20141). These microbial
RMs will provide valuable resources for challenging variant calling
algorithms using well-characterized data. The microbial whole
genome RMs are candidate NIST RMs 8375, 8376, 8377, and 8378.
The RM strains were selected based on relevance to food safety and
clinical settings and to represent a range in GC content. The four
microbial genomes are Salmonella enterica subsp., enterica serovar
Typhimurium LT2 (RM8375), a Staphylococcus aureus clinical
isolate MRSA strain (RM8376), a Pseudomonas aeruginosa clinical
isolate (RM8377), and a Clostridium sporogenes isolate (RM8378).
Another sample resource may be a set of plasmids containing
known variants. The development of plasmid sequencing controls
may be the best option for laboratory RM for rare variant
detection. Through a collaboration, the Translational Genomics
Research Institute (Flagstaff, Arizona) and Northern Arizona
University (Flagstaff, Arizona) have developed a plasmid control
using the pUC18 plasmid backbone, a widely used and commonly
available cloning vector. DNA replication in vivo based on a
plasmid represents the highest fidelity system for producing
a sequence standard. The pUC18 plasmid is 2.7 Kbp, and
DNA fragments containing SNPs or indels of interest can be
readily inserted or deleted. By mixing different SNP- or indel-
containing plasmids at known proportions, one can readily
evaluate sequencing reads of these known mixtures. Another
group successfully used plasmid controls to aid in the evaluation
of rare variant detection (Cushing et al., 2013). An added benefit
of plasmids as RMs is that their small size allows them be added
into sequencing runs as internal controls alongside other samples.
The error rate for plasmids within each sequencing run can then
1http://genomeinabottle.org
be directly measured. In fact, this is common practice with the
phiX genome in Illumina sequencing runs.
Benchmarking Datasets
Sequencing datasets can be used to evaluate the variant calling
measurement process, but do not aid in evaluation of sample
processing and sequencing components of the measurement
process (Figure 1).
Real sequencing data
Real data are the ideal source of sequence data for use in evaluating
variant calling algorithms. Sequence data for the NIST microbial
reference data are available through the NCBI sequence read
archive (BioProject Accession PRJNA252728). Current efforts are
underway to characterize the genome sequence of these RMs and
will be made publicly available for use in variant calling method
evaluation. While sequence data from well-characterized RMs
are ideal, these data are not always available or representative of
the intended use cases. When appropriate reference data are not
available, alternative sources of sequence data are needed.
One such alternative source is sequence data from isolates
sequenced on multiple sequencing platforms by multiple
sequencing centers. Sequencing manufacturers commonly use
Escherichia coli K-12 DH10B and MG1655 strains to benchmark
new sequencing chemistries and provide this data on their web-
sites. In addition, the Broad Institute (Cambridge, MA, USA),
Joint Genome Institute (Walnut Creek, CA, USA), and J. Craig
Venter Institute (Rockville, MD, USA) use E. coli K-12 MG1655for
quality control, and some of the quality control data are available
through the GenBank Sequence Read Archive (Interactive web
application for exploring datasets2). Multiple sequence datasets
are also available (archived in GenBank SRA) for Staphylococcus
aureus subsp. aureus TW20 and Bacillus subtilis strain 168.
However, it is important to keep in mind that sequence variation
may exist between stock cultures of the same strain at different
sequencing centers. This is the primary advantage of RMs, as
they are characterized for material homogeneity within a batch.
Although a well-characterized genome sequence may not be
currently available for these organisms, validated variant call sets
can be generated using approaches described in the next section.
Once a dataset has been selected, there are several approaches
for generating a genome sequence as the reference sequence data
source. One approach is to combine sequence data sets from
multiple sequencing platforms (Zook et al., 2014). Integrating
data from multiple platforms has two key advantages: biases
from each platform can be identified and down-weighted when
integrating the data and variant callers can be tested on data from
different sequencing platforms. Additionally, some laboratories
have used SNP arrays to confirm SNP calls from NGS platforms
(Goya et al., 2010; Subramanian et al., 2013). This approach
works for already identified and more easily characterized SNPs,
but is cost prohibitive for newly discovered SNPs. Furthermore,
arrays do not always work when there are SNPs neighboring
the variants of interest or in low quality mapping regions (Zook
et al., 2014). With a characterized genome sequence, a known
2https://github.com/nate-d-olson/snp_eval_shiny
Frontiers in Genetics | www.frontiersin.org July 2015 | Volume 6 | Article 2357
Olson et al. SNP evaluation review
set of variant calls can be further validated through comparison
of the characterized genome sequence to phylogenetically-related
organisms or through the insertion of simulated variants.
Simulated sequence data
Simulated sequencing datasets can be used to validate variant
calling methods when the available reference datasets do not
adequately represent the variant calling methods. A primary
advantage of using simulated data for method evaluation is that
the true variants are known. Additionally, simulated datasets can
be used in conjunction with reference datasets to test method
robustness. Numerous sequencing read simulators are available,
but differ in the error model (empirical or theoretical) used to
generate sequencing datasets. Additionally, some read simulators
do not introduce sequencing errors and thus can be used to define
a baseline for algorithm performance in the absence of such errors.
Empirically-based read simulator algorithms vary in
complexity. GenSim 1.0 (Engle and Burks, 1992) and MAQ
(Gallagher and Desjardins, 2008) assume a uniform or constant
error rate for all positions within a sequencing read. However,
error rates are known to increase toward the ends of sequencing
reads. Thus, more sophisticated error models have been
developed to incorporate position-specific error rates (Engle
and Burks, 1994; Richter et al., 2008; Holtgrewe, 2010; Jia et al.,
2013). A number of read simulators use error models with
sequence-specific error rates, for example homopolymers (Hu
et al., 2012; McElroy et al., 2012). The benefit to this approach
is that it more closely models sequence-specific errors in true
sequencing datasets compared to models that utilize uniform
or position-specific error rates. Quality scores from sequenced
datasets can also be used to model error rates (Frampton and
Houlston, 2012; Jia et al., 2013). Coverage bias and/or low
coverage in GC-rich regions are also incorporated into some
empirical error models (Frampton and Houlston, 2012; Hu et al.,
2012). Some read simulators provide algorithms that allow the
user to generate new error models, e.g., GemErr (McElroy et al.,
2012; Jia et al., 2013). Whether a default error model is used
or the user generates his/her own error model, a dataset with
ground truth base calls is necessary in order to generate accurate
error models, as inaccuracies in the reference genome would be
incorporated as sequencing errors in the model.
Sequencing read simulators can also use theoretical error
models based on the sequencing chemistry and sample
preparation, and the reaction detection methods (Myers,
1999; Hazelhurst and Bergheim, 2003; Richter et al., 2008; Angly
et al., 20123). The advantage of theoretical models is that they are
able to incorporate sources of error that are difficult to include in
empirical models.
In addition to modeling base call error rates, it is also important
to model BQS. However, some read simulators do not provide
quality scores, e.g., metaSim (Richter et al., 2008), celsim (Myers,
1999), and GenSim (Engle and Burks, 1992). Alternatively, fixed
quality values are generated by Grinder (Angly et al., 2012).
Correct bases are assigned a quality score of 30 and error positions
are assigned a score of 10. The pIRS by Hu et al. (2012) uses quality
3http://www.ebi.ac.uk/goldman-srv/simNGS/
scores from an existing sequence dataset. While this approach
is more representative of the distribution of quality scores in a
sequencing dataset, it assumes the quality scores are accurate.
The assigned BQS can also reflect the model uncertainty. For
example, flowsim (Balzer et al., 2010) uses Bayes theorem to
assign quality values for simulated base calls. Huang et al. (2011)
takes the opposite approach, modeling quality scores and using
those assigned quality scores for the base call error model. Using
an empirical approach to simulate BQS values, Li et al. (2008)
generated a position-specific quality score model for error and
non-error bases by modeling the quality score distribution using a
first order Markov chain. To fully evaluate SNP calling algorithms
using simulated sequence data, both sequences and quality scores
must simulate real sequence data as accurately as possible.
Finally, when using simulated datasets, a random number seed
value must be defined. Random numbers are used during the
generation of simulated reads and they require a seed number.
This seed number is used to produce the set of random numbers.
Reusing the same seed number results in the production of the
same set of random numbers. To enable reproducibility in the SNP
calling algorithm evaluation procedure, the user should rerun the
simulation using the same seed value, however, different seed
numbers should be used when generating replicate datasets.
Reference Genome Selection
The choice of the reference genome for SNP calling can bias
which SNPs are called, e.g., SNPs in genes not in the reference
genome will not be called, and these effects can be observed
after phylogenetic tree construction. This potential bias has
less of an effect in clonal bacteria, in which there are few
genomic variants among the clones, even when the clones are
compared to different species (Foster et al., 2009; Pearson et al.,
2009). When working with genetically diverse genomes, using
multiple references (Bertels et al., 2014) better reflects the diversity
in the number of SNPs and genomic complexity (e.g., repeat
regions and structural variants) the algorithm will face. While
accurate phylogenetic reconstruction of genetically diverse strains
requires using multiple reference genomes representing a range
of genomic similarity to the strains being compared, the same
biases and challenges exist for SNP calling. In fact, SNP calling
can be conducted against multiple references during validation
of the SNP discovery process (Shen et al., 2010; Pightling et al.,
2014).
Just as simulated whole genome sequence data can be
used to evaluate algorithm performance, reference genomes
with simulated mutations can also be used. The benefit to
using reference genomes with simulated mutations is that they
can provide a ground truth while using actual sequencing
reads. Reference sequence mutation algorithms can simulate
substitutions, as well as insertions and deletions. Some examples
of reference sequence modification tools able to introduce SNPs
into a reference sequence include: fakemut, part of the Maq
sequence mapping algorithm tool (Gallagher and Desjardins,
2008), GemHap from GemSim (McElroy et al., 2012), and
mutatrix4.
4https://github.com/ekg/mutatrix
Frontiers in Genetics | www.frontiersin.org July 2015 | Volume 6 | Article 2358
Olson et al. SNP evaluation review
TABLE 1 | Definitions of common performance metrics used in evaluating variant callers.
Metric Calculation Interpretation (Ideal) Alternative names
Accuracy TP+TN
TP+FP+TN+FN Ratio of correct calls to total calls and variants (1)
Specificity TN
TN+FP Non-variants not called as variants relative to the total non-variants (1)
Sensitivity TP
TP+FN True variants called relative to all variants (1) Recall, true positive rate (TPR), positive call rate
Precision TP
TP+FP True variants called relative to total calls (1) Positive predictive value (PPV)
False positive rate FP
TN+FP Non-variants called relative to the total non-variants (0)
Performance Metrics
A variety of performance metrics have been derived and/or
borrowed from other disciplines, including calculations of true
and FP and negative call rates, as well as more sophisticated
calculations that attempt to summarize algorithm performance in
a single value (Table 1). Graphical representation of these metrics,
such as precision versus recall plots and Receiver Operating
Characteristic (ROC) curves, can be very useful in visually
depicting the different performance aspects of various algorithms.
This section discusses common performance metrics, use of
replicates for determining metric uncertainty, as well as data
visualization approaches for interpreting performance metrics.
Selection and Calculation of Performance Metrics
Contingency tables, also referred to as confusion matrices, are
used to evaluate classifiers such as variant calling algorithms
(Figure 3). Two by two contingency tables present the relationship
between variant labels assigned by an algorithm and labels from a
declared truth set. The four basic values in a 2 ×2 contingency
table: true positive (TP), true negative (TN), false positive (FP),
and false negative (FN), are used to assess algorithm performance.
Of note, contingency tables can be functions of a parameter
value or threshold choices. For example, changing the variant call
quality score threshold used to determine variant call sets can alter
the resulting contingency table.
Several performance metrics have been developed to
characterize classifiers using contingency tables (Baldi et al.,
2000). Here, we discuss five of the most commonly used metrics
(Table 1). Accuracy is a single numeric summary of the four
values representing the ratio of the correct calls and non-calls
to the total number of variant and non-variant positions. The
other four metrics can be interpreted as conditional probability
statements relating the algorithms variant calls to a reference
variant call set. Specificity is the probability that a reference
non-variant is called as non-variant. Sensitivity is the probability
that a reference variant is called as a variant. Precision estimates
the probability that a variant call is truly a reference variant,
whereas the FP rate is the probability that a variant call is falsely
called in the sample.
Determining which metric to use is up to the user; however,
the authors suggest some or all of the commonly used metrics
presented here should be included when evaluating algorithms.
When determining which performance metrics to use, one
should consider the assumptions and requirements for the
performance metric (Table 1). For example, some metrics
depend on the proportions of positive and negative variants
FIGURE 3 | Contingency table. True Positive, False Positive, False
Negative, and True Negatives are defined based by the relationship between
variants called by the SNP calling algorithm and known differences between
the reference genome and the analyzed sample.
in the benchmark dataset, whereas sensitivity and specificity
are independent of these proportions. Another consideration
for selecting performance metrics is whether the purity or
completeness of a call set is more desirable. For example,
strict filtering criteria are used to identify high predominately
TP variant calls, however this may leave the set incomplete
and potentially misrepresentative of the population of variants.
Accuracy and precision provide insight into completeness,
whereas sensitivity and specificity measure completeness. In the
end no matter what metric is used, it must be properly interpreted
based on the application or context in which it is applied.
When comparing performance metrics across different variant
calling algorithms, it is important to consider each metric’s
associated uncertainty. The performance metric uncertainty
applied to variant calls from data may be quantified using
a bootstrapping protocol. For simulated datasets, repeated
simulations can be used to calculate performance metric
uncertainty. For example, Zeng et al. (2013) performed 10
replicate simulations and compared sensitivity, specificity, and
F-score [2 ×(precision ×recall)/(precision +recall)] for four
Frontiers in Genetics | www.frontiersin.org July 2015 | Volume 6 | Article 2359
Olson et al. SNP evaluation review
FIGURE 4 | Comparison of two variant calling algorithms using two of
the performance metrics in Table1, true positive rate and false positive
rate, calculated from contingency tables over a range of quality score
thresholds (left) and a fixed quality score threshold (right). Methods A
(red) and B (teal) indicate two different variant calling methods. Left: A
smoothing function (generalized additive model) was used to summarize the
contingency table metrics across the considered quality score interval. Red and
teal lines are smoothing functions, and the gray area represents the 95%
confidence interval. The vertical dashed line indicates the quality score cutoff
(Q=3195) for the static tables. Right: Boxplots are used to summarize the
performance metrics calculated from the contingency table value for the
replicate datasets at the defined cutoff value.
variant callers at three different coverage values (5X, 10X, 15X).
The desired performance metrics were then calculated for each
dataset and the resulting values were used to calculate the
performance metric uncertainty. Replicate simulations can be
generated for either the sample dataset or reference sequence.
If the sample dataset is replicated, the variant calling algorithm
performance can be evaluated for specific variant locations as
well as for overall performance. However, as variant detection can
be sequence context-dependent, using only one set of simulated
variants may lead to bias in favor of variant calling algorithms
better suited for the sequence context of that variant call set.
Resources for Comparing and Classifying Variant
Calls
A variety of tools have been developed for assessing performance
of variant calls with respect to benchmark variant calls for a
sample. Currently, these tools are mostly command line based
and optimized for human genomes, but many can be applied
to microbial genomes as well. The bcbio.variation tool5can
regularize vcf files, compare them to a benchmark call set, and
generate a variety of metrics, including sensitivity, specificity,
and genotyping error rate. The SMaSH benchmarking toolkit
(Talwalkar et al., 2013) can generate metrics for precision and
recall of mappers and variant callers. A novel aspect of SMaSH
is the calculation of uncertainty of precision and recall due to
imperfect benchmark call sets. Another tool, the vcflib library6,
can normalize many complex variants, generate ROC curves, and
perform comparisons across vcf files. The USeq VcfComparator
tool7can generate ROC curves and compare only variants inside
bed files. GATK also has tools to combine and compare variant
and genotype calls (McKenna et al., 2010; DePristo et al., 2011;
Van der Auwera et al., 2012). Finally, the vcfeval tool in RTGtools
can compare a vcf file to a benchmark dataset and produce ROC
5https://github.com/chapmanb/bcbio.variation
6https://github.com/ekg/vcflib
7http://useq.sourceforge.net/
Frontiers in Genetics | www.frontiersin.org July 2015 | Volume 6 | Article 23510
Olson et al. SNP evaluation review
FIGURE 5 | Scatter plot showing the relationship between two performance metrics for variant call sets. The individual data points are based on metrics
calculated from static contingency tables. The error bars represent the 95% confidence interval for each performance metric.
curves and lists of TP, FP, and FN variants (RTG paper under
review8). Two other new tools being developed by members of the
Global Alliance for Genomics and Health Benchmarking Team
are hap.py9and vgraph10, both of which are able to compare
complex variants using graph representations. Many of these tools
are being actively developed and their application to microbial
systems presents an opportunity for future research.
Finally, when a well-characterized set of variant calls is not
available for benchmarking, latent mixture models can be used
to predict rates for the truth table values (Kim and Speed, 2013;
Cantarel et al., 2014). Latent mixture models predict the model of
the underlying data (true variant call set) based on the responses
from multiple variant calling algorithms. Latent mixture models
treat the true mutation status as the latent or unknown variable.
The model can then be used to estimate the FP and FN rates
from which the TP and TN rates are calculated. While this
8http://realtimegenomics.com/products/rtg-tools/
9https://github.com/sequencing/hap.py
10https://github.com/bioinformed/vgraph
method has been frequently used to validate biomedical assays,
evaluation of algorithm performance relative to a known truth is
preferred, as the latent mixture model may be susceptible to an
unknown bias.
Visual Comparison of Performance Metrics
Graphical presentation of performance metrics can facilitate
the comparison and evaluation of algorithm performance.
The appropriate data visualization method depends on the
discreteness or continuity of the variables involved. For variant
algorithm evaluation, a single set of variants generated by an
algorithm provides an example of a discrete variable, and the
quality value assigned to each candidate variant provides an
example of a continuous variable. We analyzed 16 replicate
datasets using two different methods (A and B), generating 32
sets of variant calls. The variant call sets were then used to
demonstrate different methods for visually comparing variant
calling algorithms and their performance metrics discussed in
the previous section. The R code and input data used to
Frontiers in Genetics | www.frontiersin.org July 2015 | Volume 6 | Article 23511
Olson et al. SNP evaluation review
FIGURE 6 | Curves represent the relationship between the two performance metrics, in this case true positive and false positive rate. Replicate
samples are plotted individually to indicate the robustness in the relative performance of the two methods.
generate the figures are available at https://github.com/nate-d-
olson/snp_pipeline_evaluation.
Plotting the value of a performance metric as a function of
the chosen threshold value for the continuous variable can be
useful when comparing algorithm performance over a range of
cutoff values. Figure 4 depicts the relationship between algorithm
performance and variant quality values. The boxplots show how
method A has higher FP and TP rates compared to B for a single
threshold value. The dynamic plot with the smoothed data shows
the range of threshold values for which this relationship holds.
The range of threshold values for which a trend (e.g., method
A has a higher FP rate compared to method B) is observed is
an indication of the robustness of the trend. The ideal values
for FP and TP rates are 0 and 1 respectively. As method A has
higher values for both metrics compared to method B, a trade-
off between accepting FP vs. FN will determine which method to
use. Visualizing two metrics plotted against one another can more
clearly present the trade-off between the methods. To compare
method performance for two metrics with a fixed threshold, a
scatter plot can be used (Figure 5). When considering a range
of possible thresholds, one metric can be plotted as a function
of a second metric (Figure 6). A ROC curve is an example of
a comparison between two performance metrics over a range of
threshold values. Values of the threshold or additional metrics can
Frontiers in Genetics | www.frontiersin.org July 2015 | Volume 6 | Article 23512
Olson et al. SNP evaluation review
be indicated through variations in symbol or line color, width,
or pattern. When many metrics are presented, it may prove most
effective to plot each metric against a common variable, such as
threshold. The uncertainty of the metrics can be presented in a
number of ways. Here the uncertainty is represented qualitatively
through the comparison of the two methods relative performance
across a collection of 16 replicates. The uncertainty of the metrics
can also be presented as error bars representing the uncertainty of
one metrics given a fixed value for the second metric (marginal
distributions), or the combined distribution of both metrics (joint
distribution).
While ideal for many purposes, available reference variant call
sets may not be appropriate for a desired application. When an
appropriate benchmark variant call set is not available, Venn
diagrams can be use to compare variants called using different
methods. However, use of Venn diagrams for method comparison
should be supplemented with simulated data for the target
application. It is important to note, however, that variants called by
multiple algorithms are not necessarily true variants, as multiple
variant callers can be susceptible to the same biases.
Conclusions
When using genomic variants for phylogenetic analysis,
comparative genomics, or outbreak investigations, it is critical
to properly evaluate the variant calling method. Many sources
of error are associated with sequencing as well as variant calling.
To optimize the quality of the data used to generate the variant
calls, it is advised to minimize amplification during sequencing
library preparation, perform paired-end sequencing, remove
duplicate reads, realignment around indels, and perform BQS
recalibration.
A thorough evaluation of variant calling methods would
include: (1) The use of multiple datasets with known authoritative
variant call sets that represent the range of data the algorithm
will evaluate. The datasets used for evaluation could be real,
simulated, or a combination of both; (2) Replicates of different
sequence datasets or reference genomes to calculate performance
metric confidence intervals; (3) Performance metrics that aid
in evaluating algorithm performance. With method evaluation
performed in this manner, the user will be able to understand
positives and negatives of algorithms for the application of interest
and characterize the level of confidence in variant calls. We are
currently working to follow this review with the development of
complementary open source tools for use in evaluating variant
calling algorithm performance, with a focus on calculating and
presenting performance metrics.
Acknowledgments
The authors would like to acknowledge Dr. Nancy Lin for
feedback during the writing process. The Department of
Homeland Security (DHS) Science and Technology Directorate
sponsored the production of this material under Interagency
Agreement HSHQPM-13-X-00190 with NIST and HSHQDC-
10-C-00152 to TGen (Keim) and NAU (Foster).
References
Alkan, C., Coe, B. P., and Eichler, E. E. (2011). Genome structural variation
discovery and genotyping. Nat. Rev. Genet. 12, 363–376. doi: 10.1038/nrg2958
Altmann, A., Weber, P., Bader, D., Preuss, M., Binder, E. B., and Muller-Myhsok,
B. (2012). A beginners guide to SNP calling from high-throughput DNA-
sequencing data. Hum. Genet. 131, 1541–1554. doi: 10.1007/s00439-012-1213-z
Altschul, S. F., Gish, W., Miller, W., Myers, E. W., and Lipman, D. J. (1990). Basic
local alignment search tool. J. Mol. Biol. 215, 403–410. doi: 10.1016/S0022-
2836(05)80360-2
Angiuoli, S. V., and Salzberg, S. L. (2011). Mugsy: fast multiple alignment
of closely related whole genomes. Bioinformatics 27, 334–342. doi:
10.1093/bioinformatics/btq665
Angly, F. E., Willner, D., Rohwer, F., Hugenholtz, P., and Tyson, G. W. (2012).
Grinder: a versatile amplicon and shotgunse quencesimulator. NucleicAcids Res.
40, e94. doi: 10.1093/nar/gks251
Baker, M. (2012). De novo genomeassembly: what every biologist should know. Nat.
Methods 9, 333–337. doi: 10.1038/nmeth.1935
Baldi, P., Brunak, S., Chauvin, Y., Andersen, C. A. F., and Nielsen, H. (2000).
Assessing the accuracy of prediction algorithms for classification: an overview.
Bioinformatics 16, 412–424. doi: 10.1093/bioinformatics/16.5.412
Balzer, S., Malde, K., Lanzen, A., Sharma, A., andJonassen, I. (2010). Characteristics
of 454 pyrosequencing data—enabling realistic simulation with flowsim.
Bioinformatics 26, 420–425. doi: 10.1093/bioinformatics/btq365
Bankevich, A., Nurk, S., Antipov, D., Gurevich, A. A., Dvorkin, M., Kulikov,
A. S., et al. (2012). SPAdes: a new genome assembly algorithm and its
applications to single-cell sequencing. J. Comput. Biol. 19, 455–477. doi:
10.1089/cmb.2012.0021
Bertels, F., Silander, O. K., Pachkov, M., Rainey, P. B., and Van Nimwegen, E. (2014).
Automated reconstruction of whole-genome phylogenies from short-sequence
reads. Mol. Biol. Evol. 31, 1077–1088. doi: 10.1093/molbev/msu088
Bian, J., Liu, C., Wang, H., Xing, J., Kachroo, P., and Zhou, X. (2013). SNVHMM:
predicting single nucleotide variants from next generation sequencing. BMC
Bioinformatics 14:225. doi: 10.1186/1471-2105-14-225
Cantarel, B. L., Weaver, D., Mcneill, N., Zhang, J., Mackey, A. J., and Reese, J.
(2014). BAYSIC: a Bayesian method for combining sets of genome variants
with improved specificity and sensitivity. BMC Bioinformatics 15:104. doi:
10.1186/1471-2105-15-104
Cantor, R. M., Lange, K., and Sinsheimer, J. S. (2010). Prioritizing GWAS results: a
review of statistical methods and recommendations for their application. Am. J.
Hum. Genet. 86, 6–22. doi: 10.1016/j.ajhg.2009.11.017
Chaisson, M. J., and Pevzner, P. A. (2008). Short read fragment assembly of bacterial
genomes. Genome Res. 18, 324–330. doi: 10.1101/gr.7088808
Chen-Harris, H., Borucki, M. K., Torres, C., Slezak, T. R., and Allen, J. E. (2013).
Ultra-deep mutant spectrum profiling: improving sequencing accuracy using
overlapping read pairs. BMC Genomics 14:96. doi: 10.1186/1471-2164-14-96
Chikhi, R., and Medvedev, P. (2014). Informed and automated k-mer
size selection for genome assembly. Bioinformatics 30, 31–37. doi:
10.1093/bioinformatics/btt310
Cline, J., Braman, J. C., and Hogrefe, H. H. (1996). PCR fidelity of Pfu DNA
polymerase and other thermostable DNA polymerases. Nucleic Acids Res. 24,
3546–3551. doi: 10.1093/nar/24.18.3546
Cock, P. J. A., Bonfield, J. K., Chevreux, B., and Li, H. (2015). SAM/BAM format
v1.5 extensions for de novo assemblies. bioRxiv 1–3. doi: 10.1101/020024
Colman, R. E., Schupp, J. M., Hicks, N. D., Smith, D. E., Buchhagen, J. L., Valafar,
F., et al. (2015). Detection of low-level mixed-population drug resistance in
Mycobacterium tuberculosis using high fidelity amplicon sequencing. PLoS
ONE 10:e0126626. doi: 10.1371/journal.pone.0126626
Cushing, A., Flaherty, P., Hopmans, E., Bell, J., and Ji, H. (2013). RVD: a command-
line program for ultrasensitive rare single nucleotide variant detection using
targeted next-generation DNA resequencing. BMC Res. Notes 6:206. doi:
10.1186/1756-0500-6-206
Darling, A. C., Mau, B., Blattner, F. R., and Perna, N. T. (2004). Mauve: multiple
alignment of conserved genomic sequence with rearrangements. Genome Res.
14, 1394–1403. doi: 10.1101/gr.2289704
Del Fabbro, C., Scalabrin, S., Morgante, M., and Giorgi, F. M. (2013). An extensive
evaluation of read trimming effects on Illumina NGS data analysis. PLoS ONE
8:e85024. doi: 10.1371/journal.pone.0085024
Frontiers in Genetics | www.frontiersin.org July 2015 | Volume 6 | Article 23513
Olson et al. SNP evaluation review
Delcher, A. L., Phillippy, A., Carlton, J., and Salzberg, S. L. (2002). Fast algorithms
for large-scale genome alignment and comparison. Nucleic Acids Res. 30,
2478–2483. doi: 10.1093/nar/30.11.2478
DePristo, M. A., Banks, E., Poplin, R., Garimella, K. V., Maguire, J. R., Hartl,
C., et al. (2011). A framework for variation discovery and genotyping using
next-generation DNA sequencing data. Nat. Genet. 43, 491–498. doi: 10.1038/
ng.806
Dohm, J. C., Lottaz, C., Borodina, T., and Himmelbauer, H. (2008). Substantial
biases in ultra-short read data sets from high-throughput DNA sequencing.
Nucleic Acids Res. 36, e105. doi: 10.1093/nar/gkn425
Ellison, S. L. R., and Williams, A. (2012). Eurachem/CITAC Guide: Quantifying
Uncertainty in Analytical Measurement, 3rd Edn. ISBN 978-0-948926-30-3.
Available at: http://www.eurachem.org
Engle, M. L., and Burks, C. (1992). Artificially generated data sets for
testing DNA sequence assembly algorithms. Genomics 16, 286–288. doi:
10.1006/geno.1993.1180
Engle, M. L., and Burks, C. (1994). GenFrag 2.1: new features for more robust
fragment assembly benchmarks. Comput. Appl. Biosci. 10, 567–568. doi:
10.1093/bioinformatics/10.5.567
Ewing, B., and Green, P. (1998). Base-calling of automated sequencer traces using
phred. II. Error probabilities. Genome Res. 8, 186–194. doi: 10.1101/gr.8.3.175
Fonseca, N. A., Rung, J., Brazma, A., and Marioni, J. C. (2012). Tools for
mapping high-throughput sequencing data. Bioinformatics 28, 3169–3177. doi:
10.1093/bioinformatics/bts605
Foster, J. T., Beckstrom-Sternberg, S. M., Pearson, T., Beckstrom-Sternberg, J. S.,
Chain, P. S. G., Roberto, F. F., et al. (2009). Whole genome-based phylogeny
and divergence of the genus Brucella.J. Bacteriol. 191, 2864–2870. doi: 10.1128/
JB.01581-08
Frampton, M., and Houlston, R. (2012). Generation of artificial FASTQ files to
evaluate the performance of next-generation sequencing pipelines. PLoS ONE
7:e49110. doi: 10.1371/journal.pone.0049110
Gallagher, S. R., and Desjardins, P. R. (2008). Quantitation of DNA and RNA with
absorption and fluorescence spectroscopy. Curr. Protoc. Protein Sci. (Suppl. 52),
A.4K.1–A.4K21. doi: 10.1002/0471140864.psa04ks52
Gardner, S. N., and Hall, B. G. (2013). When whole-genome alignments
just won’t work: kSNP v2 software for alignment-free SNP discovery and
phylogenetics of hundreds of microbial genomes. PLoS ONE 8:e81760. doi:
10.1371/journal.pone.0081760
Goya, R., Sun, M. G. F., Morin, R. D., Leung, G., Ha, G., Wiegand, K. C., et al.
(2010). SNVMix: predicting single nucleotide variants from next-generation
sequencing of tumors. Bioinformatics 26, 730–736. doi: 10.1093/bioinformatics/
btq040
Grad, Y. H., Lipsitch, M., Feldgarden, M., Arachchi, H. M., Cerqueira, G.
C., Fitzgerald, M., et al. (2012). Genomic epidemiology of the Escherichia
coli O104:H4 outbreaks in Europe, 2011. Proc. Natl. Acad. Sci. U.S.A. 109,
3065–3070. doi: 10.1073/pnas.1121491109
Harris, S. R., Torok, M. E., Cartwright, E. J., Quail, M. A., Peacock,S. J., and Parkhill,
J. (2013). Read and assemblymetrics inconsequential for clinical utility of whole-
genome sequencing in mapping outbreaks. Nat. Biotechnol. 31, 592–594. doi:
10.1038/nbt.2616
Hatem, A., Bozdağ, D., Toland, A. E., and Çatalyürek, Ü. V. (2013). Benchmarking
short sequence mapping tools. BMC Bioinformatics 14:184. doi: 10.1186/1471-
2105-14-184
Hazelhurst, S., and Bergheim, A. (2003). SeqGen: a Tool for Creating Benchmarks
for EST Clustering Algorithms. Technical Report - on ESTSim. Available at:
http://gi.cebitec.uni-bielefeld.de/people/zsuzsa/ESTclust/ESTSim.pdf
Hendriksen, R. S., Price, L. B., Schupp, J. M., Gillece, J. D., Kaas, R. S., Engelthaler,
D. M., et al. (2011). Population genetics of Vibrio cholerae from Nepal in 2010:
evidence on the origin of the Haitian outbreak. MBio 2, e00157-00111. doi:
10.1128/mBio.00157-11
Holtgrewe, M. (2010). Mason—A Read Simulator for Second Generation Sequencing
Data. Technical Report TR-B-10-06. Institut für Mathematik und Informatik,
Freie Universität Berlin
Homer, N., and Nelson, S. F. (2010). Improved variant discovery through local re-
alignment of short-read next-generation sequencing data using SRMA. Genome
Biol. 11, R99. doi: 10.1186/gb-2010-11-10-r99
Hu, X., Yuan, J., Shi, Y., Lu, J., Liu, B., Li, Z., et al. (2012). pIRS: profile-
based Illumina pair-end reads simulator. Bioinformatics 28, 1533–1535. doi:
10.1093/bioinformatics/bts187
Huang, W., Li, L., Myers, J. R., and Marth, G. T. (2011). ART: a next-
generation sequencing read simulator. Bioinformatics 28, 593–594. doi:
10.1093/bioinformatics/btr708
Jia, B., Xuan, L., Cai, K., Hu, Z., Ma, L., and Wei, C. (2013). NeSSM: a next-
generation sequencing simulator for metagenomics. PLoS ONE 8:e75448. doi:
10.1371/journal.pone.0075448
Keim, P. S., and Wagner, D. M. (2009). Humans and evolutionary and ecological
forces shaped the phylogeography of recently emerged diseases. Nat. Rev.
Microbiol. 7, 813–821. doi: 10.1038/nrmicro2219
Kim, S., and Speed, T. (2013). Comparing somatic mutation-callers: beyond Venn
diagrams. BMC Bioinformatics 14:189. doi: 10.1186/1471-2105-14-189
Koboldt, D. C., Chen, K., Wylie, T., Larson, D. E., Mclellan, M. D., Mardis, E. R.,
et al. (2009). VarScan: variant detection in massively parallel sequencing
of individual and pooled samples. Bioinformatics 25, 2283–2285. doi:
10.1093/bioinformatics/btp373
Koren, S., and Phillippy, A. M. (2015). One chromosome, one contig: complete
microbial genomes from long-read sequencing and assembly. Curr. Opin.
Microbiol. 23C, 110–120. doi: 10.1016/j.mib.2014.11.014
Kozarewa, I., Ning, Z., Quail, M. A., Sanders, M. J., Berriman, M., and Turner, D.
J. (2009). Amplification-free Illumina sequencing-library preparation facilitates
improved mapping and assembly of (G+++C)-biased genomes. Nat. Methods
6, 291–295. doi: 10.1038/nmeth.1311
Kurtz, S., Phillippy, A., Delcher, A., Smoot, M., Shumway, M., Antonescu, C., et al.
(2004). Versatile and open software for comparing large genomes. Genome Biol.
5, R12. doi: 10.1186/gb-2004-5-2-r12
Li, H. (2011). A statistical framework for SNP calling, mutation discovery,
association mapping and population genetical parameter estimation
from sequencing data. Bioinformatics 27, 2987–2993. doi: 10.1093/
bioinformatics/btr509
Li, H., Ruan, J., and Durbin, R. (2008). Mapping short DNA sequencing reads and
calling variants using mapping quality scores. Genome Res. 18, 1851–1858. doi:
10.1101/gr.078212.108
Loman, N. J., Constantinidou, C., Chan, J. Z., Halachev, M., Sergeant, M.,
Penn, C. W., et al. (2012). High-throughput bacterial genome sequencing:
an embarrassment of choice, a world of opportunity. Nat. Rev. Microbiol. 10,
599–606. doi: 10.1038/nrmicro2850
Lou, D. I., Hussmann, J. A., Mcbee, R. M., Acevedo, A., Andino, R., Press, W.
H., et al. (2013). High-throughput DNA sequencing errors are reduced by
orders of magnitude using circle sequencing. Proc. Natl. Acad. Sci. U.S.A. 110,
19872–19877. doi: 10.1073/pnas.1319590110
Magoc, T., Pabinger, S., Canzar, S., Liu, X., Su, Q., Puiu, D., et al. (2013). GAGE-B:
an evaluation of genome assemblers for bacterial organisms. Bioinformatics 29,
1718–1725. doi: 10.1093/bioinformatics/btt273
McElroy, K., Thomas, T., and Luciani, F. (2014). Deep sequencing of evolving
pathogen populations: applications, errors, and bioinformatic solutions. Microb.
Inform. Exp. 4, 1. doi: 10.1186/2042-5783-4-1
McElroy, K. E., Luciani, F., and Thomas, T. (2012). GemSIM: general, error-model
based simulator of next-generation sequencing data. BMC Genomics 13:74. doi:
10.1186/1471-2164-13-74
McKenna, A., Hanna, M., Banks, E., Sivachenko, A., Cibulskis, K., Kernytsky,
A., et al. (2010). The genome analysis toolkit: a mapreduce framework for
analyzing next-generation DNA sequencing data. Genome Res. 20, 1297–1303.
doi: 10.1101/gr.107524.110
Meacham, F., Boffelli, D., Dhahbi, J., Martin, D., Singer, M., and Pachter, L. (2011).
Identification and correction of systematic error in high-throughput sequence
data. BMC Bioinformatics 12:451. doi: 10.1186/1471-2105-12-451
Medvedev, P., Scott, E., Kakaradov, B., and Pevzner, P. (2011). Error correction
of high-throughput sequencing datasets with non-uniform coverage.
Bioinformatics 27, i137–i141. doi: 10.1093/bioinformatics/btr208
Metzker, M. L. (2010). Sequencing technologies—the next generation. Nat. Rev.
Genet. 11, 31–46. doi: 10.1038/nrg2626
Myers, G. (1999). A dataset generator for whole genome shotgun sequencing. Proc.
Int. Conf. Intell. Syst. Mol. Biol. 202–210.
Nagarajan, N., and Pop, M. (2013). Sequence assembly demystified. Nat. Rev.Genet.
14, 157–167. doi: 10.1038/nrg3367
Nelson, C. L., Pelak, K., Podgoreanu, M. V., Ahn, S. H., Scott, W. K., Allen, A. S., et al.
(2014). A genome-wide association study of variantsasso ciated with acquisition
of Staphylococcus aureus bacteremia in a healthcare setting. BMC Infect. Dis.
14:83. doi: 10.1186/1471-2334-14-83
Frontiers in Genetics | www.frontiersin.org July 2015 | Volume 6 | Article 23514
Olson et al. SNP evaluation review
Nielsen, R., Paul, J. S., Albrechtsen, A., and Song, Y. S. (2011). Genotype and SNP
calling from next-generation sequencing data. Nat. Rev. Genet. 12, 443–451. doi:
10.1038/nrg2986
Nikolenko, S. I., Korobeynikov, A. I., and Alekseyev, M. A. (2013). BayesHammer:
bayesian clusteringfor error correction in single-cell sequencing. BMC Genomics
14(Suppl. 1):S7. doi: 10.1186/1471-2164-14-S1-S7
Pabinger, S., Dander, A., Fischer, M., Snajder, R., Sperk, M., Efremova, M., et
al. (2013). A survey of tools for variant analysis of next-generation genome
sequencing data. Brief. Bioinform. 15, 256–278. doi: 10.1093/bib/bbs086
Pavlopoulos, G., Oulas, A., Iacucci, E., Sifrim, A., Moreau, Y., Schneider, R., et al.
(2013). Unraveling genomic variation from next generation sequencing data.
BioData Min. 6, 13. doi: 10.1186/1756-0381-6-13
Pearson, T., Okinaka, R. T., Foster, J. T., and Keim, P. (2009). Phylogenetic
understanding of clonal populations in an era of whole genome sequencing.
Infect. Genet. Evol. 9, 1010–1019. doi: 10.1016/j.meegid.2009.05.014
Peng, Y., Leung, H. M., Yiu, S. M., and Chin, F. L. (2010). “IDBA—a practical
iterative de Bruijn graph De Novo assembler, in Research in Computational
Molecular Biology, ed. B. Berger (Berlin: Springer), 426–440.
Pightling, A. W., Petronella, N., and Pagotto,F. (2014). Choice of references equence
and assembler for alignment of Listeria monocytogenes short-read sequence
data greatly influences rates of error in SNP analyses. PLoS ONE 9:e104579. doi:
10.1371/journal.pone.0104579
Rasko, D. A., Worsham, P. L., Abshire, T. G., Stanley, S. T., Bannan, J. D., Wilson,
M. R., et al. (2011). Bacillus anthraciscomparative genome analysis in support of
the Amerithrax investigation. Proc. Natl. Acad. Sci. U.S.A. 108, 5027–5032. doi:
10.1073/pnas.1016657108
Richter, D. C., Ott, F., Auch, A. F., Schmid, R., and Huson, D. H. (2008). MetaSim: a
sequencing simulator for genomics and metagenomics. PLoS ONE 3:e3373. doi:
10.1371/journal.pone.0003373
Ross, M. G., Russ, C., Costello, M., Hollinger, A., Lennon, N. J., Hegarty, R., et al.
(2013). Characterizing and measuring bias in sequence data. Genome Biol. 14,
R51. doi: 10.1186/gb-2013-14-5-r51
Sahl, J. W., Gillece, J. D., Schupp, J. M., Waddell, V. G., Driebe, E. M., Engelthaler,
D. M., et al. (2013). Evolution of a pathogen: a comparative genomics analysis
identifies a genetic pathway to pathogenesis in Acinetobacter. PLoS ONE
8:e54287. doi: 10.1371/journal.pone.0054287
Schbath, S., Martin, V., Zytnicki, M., Fayolle, J., Loux, V., and Gibrat, J. F. (2012).
Mapping reads on a genomic sequence: an algorithmic overview and a practical
comparative analysis. J. Comput. Biol. 19, 796–813. doi: 10.1089/cmb.2012.0022
Schmitt, M. W., Kennedy, S. R., Salk, J. J., Fox, E. J., Hiatt, J. B., and Loeb, L. A.
(2012). Detection of ultra-rare mutations by next-generation sequencing. Proc.
Natl. Acad. Sci. U.S.A. 109, 14508–14513. doi: 10.1073/pnas.1208715109
Shen, Y., Wan, Z., Coarfa, C., Drabek, R., Chen, L., Ostrowski, E. A., et al. (2010). A
SNP discovery method to assess variant allele probability from next-generation
resequencing data. Genome Res. 20, 273–280. doi: 10.1101/gr.096388.109
Showalter, A. K., and Tsai, M.-D. (2002). A reexamination of the nucleotide
incorporation fidelity of DNA polymerases. Biochemistry 41, 10571–10576. doi:
10.1021/bi026021i
Subramanian, S., Di Pierro, V., Shah, H., Jayaprakash, A. D., Weisberger, I., Shim,
J., et al. (2013). MiST: a new approach to variant detection in deep sequencing
datasets. Nucleic Acids Res. 41, e154. doi: 10.1093/nar/gkt551
Swain, M. T., Tsai, I. J., Assefa, S. A., Newbold, C., Berriman, M., and Otto,
T. D. (2012). A post-assembly genome-improvement toolkit (PAGIT) to
obtain annotated genomes from contigs. Nat. Protoc. 7, 1260–1284. doi:
10.1038/nprot.2012.068
Talwalkar, A., Liptrap, J., Newcomb, J., Hartl, C., Terhorst, J., Curtis, K., et al.
(2013). SMaSH: a benchmarking toolkit for human genome variant calling.
Bioinformatics 30, 2787–2795. doi: 10.1093/bioinformatics/btu345
Treangen, T. J., Ondov, B. D., Koren, S., and Phillippy, A. M. (2014). The
Harvest suite for rapid core-genome alignment and visualization of thousands
of intraspecific microbial genomes. Genome Biol. 15, 524. doi: 10.1186/s13059-
014-0524-x
Van der Auwera, G. A., Carneiro, M. O., Hartl, C., Poplin, R., Del Angel,
G., Levy-Moonshine, A., et al. (2012). From fastQ data to high-confidence
variant calls: the Genome Analysis Toolkit Best Practices Pipeline. Curr. Protoc.
Bioinformatics 43, 11.10.1–11.10.33. doi: 10.1002/0471250953.bi1110s43
Vogler, A. J., Keys, C., Nemoto, Y., Colman, R. E., Jay, Z., and Keim, P. (2006).
Effect of repeat copy number on variable-number tandem repeat mutations in
Escherichia coli O157:H7. J.B acteriol. 188, 4253–4263. doi: 10.1128/JB.00001-06
Walker, B. J., Abeel, T., Shea, T., Priest, M., Abouelliel, A., Sakthikumar, S.,
et al. (2014). Pilon: an integrated tool for comprehensive microbial variant
detection and genome assembly improvement. PLoS ONE 9:e112963. doi:
10.1371/journal.pone.0112963
Zeng, F., Jiang, R., and Chen, T. (2013). PyroHMMvar: a sensitive and accurate
method to call short indels and SNPs forIon Torrentand 454 data. Bioinformatics
29, 2859–2868. doi: 10.1093/bioinformatics/btt512
Zerbino, D. R., and Birney, E. (2008). Velvet: algorithms for de novo short
read assembly using de Bruijn graphs. Genome Res. 18, 821–829. doi:
10.1101/gr.074492.107
Zhong, C., Yang, Y., and Yooseph, S. (2015). GRASP: guided reference-based
assembly of short peptides. Nucleic Acids Res. 43, e18. doi: 10.1093/nar/gku1210
Zook, J. M., Chapman, B., Wang, J., Mittelman, D., Hofmann, O., Hide, W.,
et al. (2014). Integrating human sequence data sets provides a resource of
benchmark SNP and indel genotype calls. Nat. Biotechnol. 32, 246–251. doi:
10.1038/nbt.2835
Zook, J. M., Samarov, D., Mcdaniel, J., Sen, S. K., and Salit, M. (2012). Synthetic
spike-in standards improve run-specific systematic error analysis for DNA and
RNA sequencing. PLoS ONE 7:e41356. doi: 10.1371/journal.pone.0041356
Conflict of Interest Statement: Opinions expressed in this paper are the authors’
and do not necessarily reflect the policies and views of the Department of Homeland
and Security (DHS), National Institute of Standards and Technology (NIST), or
affiliated venues. Certain commercial equipment, instruments, or materials are
identified in this paper only to specify the experimental procedure adequately. Such
identification is not intended to imply recommendation or endorsement by the
NIST, nor is it intended to imply that the materials or equipment identified are
necessarily the best available for the purpose. Official contribution of NIST; not
subject to copyrights in USA.
Copyright © 2015 Olson, Lund, Colman, Foster, Sahl, Schupp, Keim, Morrow, Salit
and Zook. This is an open-access article distributed under the terms of the Creative
Commons Attribution License (CC BY). The use, distribution or reproduction in other
forums is permitted, provided the original author(s) or licensor are credited and that
the original publication in this journal is cited, in accordance with accepted academic
practice. No use, distribution or reproduction is permitted which does not comply with
these terms.
Frontiers in Genetics | www.frontiersin.org July 2015 | Volume 6 | Article 23515
... The SNPs located within or close to an annotated gene were aggregated into one single haplotype, providing a haplotype allele frequency (haplotype allele frequency). More than 20 million variants were identified between the cultivar and wild form donor. Variant detection of next-generation sequencing data is exceptionally error-prone (Nielsen et al. 2011;Olson et al. 2015); therefore, variant loci not reported in the Ensembl online database were removed (Cunningham et al. 2019). This approach retained about 3.6 million SNPs from pooled SNPs calling. ...
Article
Full-text available
Sustainable food production for a growing world population will pose a central challenge in the coming decades. Organic farming is among the feasible approaches to achieving this goal if the yield gap to conventional farming can be decreased. However, uncertainties exist to which extend-and for which phenotypes in particular-organic and conventional agro-ecosystems require differentiated breeding strategies. To answer this question, a heterogeneous spring barley population was established between a wild barley and an elite cultivar to examine this question. This initial population was divided into two sets and sown one in organic and the other in conventional managed agro-ecosystems, without any artificial selection for two decades. A fraction of seeds harvested each year was sown the following year. Various generations, up to the 23th were whole-genome pool-sequenced to identify adaptation patterns towards ecosystem and climate conditions in the allele frequency shifts. Additionally, a meta-data analysis was conducted to link genomic regions' increased fitness to agronomi-cally related traits. This long-term experiment highlights for the first time that allele frequency pattern difference between the conventional and organic populations grew with subsequent generations. Further, the organic-adapted population showed a higher genetic heterogeneity. The data indicate that adaptations towards new environments happen in few generations. Drastic interannual changes in climate are manifested in significant allele frequency changes. Particular wild form alleles were positively selected in both environments. Clustering these revealed an increased fitness associated with biotic stress resistance, yield physiology, and yield components in both systems. Additionally, the introduced wild alleles showed increased fitness related to root morphology, developmental processes, and abiotic stress responses in the organic agro-ecosystem. Concluding the genetic analysis, we demonstrate that breeding of organically adapted varieties should be conducted in an organically managed agro-ecosystem, focusing on root-related traits, to close the yield gap towards conventional farming.
... However, variant calling is challenging. Errors can occur in each step of variant identification, including sequencing errors, incorrect alignment to the reference genome (2), or erroneous variant calling (3). ...
Preprint
Full-text available
A critical step in the analysis of WGS data is variant calling. Despite its importance, variant calling is prone to errors. Our study investigated the association between incorrect SNP and variant quality metrics and nucleotide context. In our study, incorrect SNPs were defined in twenty Holstein-Friesian cows by comparing their SNPs genotypes identified by whole genome sequencing with the IlluminaNovaSeq6000 and the EuroGMD50K genotyping microarray. The data set was divided into the correct set of SNPs (666,333 SNPs) and the incorrect set of SNPs (4,557 SNPs). The training data set consisted of only the correct SNPs, while the test data set contained a balanced mix of all the incorrectly and correctly called SNPs. An autoencoder was constructed to identify systematically incorrect SNPs that were marked as outliers by a one-class support vector machine and isolation forest algorithms. The results showed that 59.53% (±0.39%) of the incorrect SNPs had systematic patterns, with the remainder being random errors. The frequent occurrence of the CGC trimer was due to mislabeling a call for C. Incorrect T instead A call was associated with the presence of T in the neighboring downstream position. These errors may arise due to the fluorescence patterns of nucleotide labelling.
... Using the simple predictive model that the destruction of the operon structure and alteration of the 5′ flanking non-coding regions, including promoters and transcriptional regulators, can interfere with gene expression, this study has shown that genomic variations, specifically SVs, can alter gene expression, including some virulence genes ( Figure 5B and Table 2). Interestingly, deletion of LuxI/LuxR-type QS systems observed in the strain Zagreb may also be indirectly altering the expression of many other genes, including those of virulence factors, toxins, and biofilm components (74). ...
Article
Full-text available
Glanders is a highly contagious and life-threatening zoonotic disease caused by Burkholderia mallei (B. mallei). Without an effective vaccine or treatment, early diagnosis has been regarded as the most effective method to prevent glanders transmission. Currently, the diagnosis of glanders is heavily reliant on serological tests. However, given that markedly different host immune responses can be elicited by genetically different strains of the same bacterial species, infection by B. mallei, whose genome is unstable and plastic, may result in various immune responses. This variability can make the serodiagnosis of glanders challenging. Therefore, there is a need for a comprehensive understanding and assessment of how B. mallei genomic variations impact the appropriateness of specific target antigens for glanders serodiagnosis. In this study, we investigated how genomic variations in the B. mallei genome affect gene content (gene presence/absence) and expression, with a special focus on antigens used or potentially used in serodiagnosis. In all the genome sequences of B. mallei isolates available in NCBI’s RefSeq database (accessed in July 2023) and in-house sequenced samples, extensive small and large variations were observed when compared to the type strain ATCC 23344. Further pan-genome analysis of those assemblies revealed variations of gene content among all available genomes of B. mallei. Specifically, differences in gene content ranging from 31 to 715 genes with an average of 334 gene presence-absence variations were found in strains with complete or chromosome-level genome assemblies, using the ATCC 23344 strain as a reference. The affected genes included some encoded proteins used as serodiagnostic antigens, which were lost due mainly to structural variations. Additionally, a transcriptomic analysis was performed using the type strain ATCC 23344 and strain Zagreb which has been widely utilized to produce glanders antigens. In total, 388 significant differentially expressed genes were identified between these two strains, including genes related to bacterial pathogenesis and virulence, some of which were associated with genomic variations, particularly structural variations. To our knowledge, this is the first comprehensive study to uncover the impacts of genetic variations of B. mallei on its gene content and expression. These differences would have significant impacts on host innate and adaptive immunity, including antibody production, during infection. This study provides novel insights into B. mallei genetic variants, knowledge which will help to improve glanders serodiagnosis.
... Only positions in the core genome, covered by all query genomes, are used to establish a set of core SNPs [109]. An SNP distance matrix is then computed for all pairwise combinations, enabling subsequent phylogenetic analysis [110]. In contrast to conventional Multilocus Sequence Typing (MLST), which examines genetic similarity based on only seven genes, core genome MLST assesses thousands of gene regions or alleles. ...
Article
Full-text available
Recent advancements in sequencing technology and data analytics have led to a transformative era in pathogen detection and typing. These developments not only expedite the process, but also render it more cost-effective. Genomic analyses of infectious diseases are swiftly becoming the standard for pathogen analysis and control. Additionally, national surveillance systems can derive substantial benefits from genomic data, as they offer profound insights into pathogen epidemiology and the emergence of antimicrobial-resistant strains. Antimicrobial resistance (AMR) is a pressing global public health issue. While clinical laboratories have traditionally relied on culture-based antimicrobial susceptibility testing, the integration of genomic data into AMR analysis holds immense promise. Genomic-based AMR data can furnish swift, consistent, and highly accurate predictions of resistance phenotypes for specific strains or populations, all while contributing invaluable insights for surveillance. Moreover, genome sequencing assumes a pivotal role in the investigation of hospital outbreaks. It aids in the identification of infection sources, unveils genetic connections among isolates, and informs strategies for infection control. The One Health initiative, with its focus on the intricate interconnectedness of humans, animals, and the environment, seeks to develop comprehensive approaches for disease surveillance, control, and prevention. When integrated with epidemiological data from surveillance systems, genomic data can forecast the expansion of bacterial populations and species transmissions. Consequently, this provides profound insights into the evolution and genetic relationships of AMR in pathogens, hosts, and the environment.
... Most of the variant callers are tuned for human samples and there are very few "truth-sets" available for microbes [14]. Moreover, a smaller number of mutations along with high mutation rates exist in certain bacterial genomes as compared to humans, hence, validating variant callers for bacterial genome sequences is essential [17]. Among read aligners-NovoAlign [18] , Bowtie2 [19] , BWA-MEM [20] , SOAP3-dp [21] , CUSHAW3 [22] , MOSAIK [23] are the most widely used. ...
Conference Paper
Full-text available
Analyzing the data deluge generated by next generation sequencing technologies is a challenging task. Variant calling aids in characterizing phenotype of an organism. The objective of this work is to compare popular variant calling protocols available in terms of their accuracy for human and microbial genomes. Three popular read aligners viz. BWA-MEM, Bowtie2 and NovoAlign were used for the human genome benchmark dataset of Genome in a Bottle (GIAB). Variant calling was performed using four popular variant callers viz. samtools-bcftools, VarScan2, GATK-HC and DeepVariant. For microbial genomes, BWA-MEM was used for reference mapping followed by variant calling using above-mentioned tools (excluding DeepVariant) along with FreeBayes and PILON. The predicted variants were assessed in terms of true positives, false positives and false negatives. Consensus approach was used for identifying SNPs (predicted by more than one variant calling pipeline) in the false positive dataset that satisfied the threshold criteria to be included as true positives. Precision and Recall were measured for evaluating each of the variant calling pipelines. Variant calling pipeline using BWA-MEM for read mapping along with either VarScan2 or samtools-bcftools predicted optimal number of single nucleotide polymorphisms (SNP) for high coverage human datasets. For Indel calling, GATK-HC with any aligner was preferred over VarScan2 and samtools-bcftools. Variant calling using bcftools, Pilon and GATK-HC with default parameters for heterogeneous microbial datasets revealed optimal performance. SNP calling with minimum errors was evident when $\sim\mathbf{95}\%$ of the samples from a population was taken into cognizance for variant prediction. Predicted variants were validated by comparing with gold standard data which aided in improved identification of actionable high confidence variants, reduced prediction of false positives and this approach has proved to be valuable for prioritization of actionable variants.
... Variant calling (single-nucleotide polymorphisms [SNPs] and Indels) is a crucial technique for comparing genomes, as it can determine nucleotide differences between organisms (Olson et al. 2015). SNPs and indels can identify differences in nucleotides position relative to a reference genome (Nielsen et al. 2011). ...
Article
Full-text available
Cyanobacteria have a strong potential to compete well under elevated temperatures. Understanding how they acclimate and evolve under climatic stressors can help us accurately predict their response to forecasted future conditions. However, it is unclear whether increased temperature results in microevolution and/or changes in gene expression. This is the first study to investigate how long‐term exposure under increased temperature influences cyanobacterial genomes. Here, we cultivated three strains of Microcystis aeruginosa (M10, M11, and M12) under two temperature conditions, ambient (22°C) and high‐temperature (26°C) for 2 yr and subsequently sequenced the full genomes. The six genomes were then compared to a reference genome and analyzed for single‐nucleotide polymorphisms, from which the mutation rate was calculated to see if temperature influenced the prevalence of gene changes. Furthermore, we investigated how temperature impacted the gene expression of six genes involved in thermal tolerance and heat shock response. We found that M. aeruginosa exposure to high temperatures demonstrated a stronger expressional response with genes associated with heat shock and thermal tolerance due to exposure to elevated temperature. Although the functionality of many genes encoding for the carbon concentrating mechanisms, nutrient metabolism and secondary metabolites were unaffected, temperature could be a possible driver of genetic change due to enhanced mutation rates. Yet, differing patterns in M10 exposed to high temperatures suggests strain specifics components are also a factor. These patterns suggest changes in plasticity, which would allow for M. aeruginosa to respond rapidly to changes in temperature and to be resilient to environmental change.
Article
Whole‐genome sequencing is widely used to investigate population genomic variation in organisms of interest. Assorted tools have been independently developed to call variants from short‐read sequencing data aligned to a reference genome, including single nucleotide polymorphisms (SNPs) and structural variations (SVs). We developed SNP‐SVant, an integrated, flexible, and computationally efficient bioinformatic workflow that predicts high‐confidence SNPs and SVs in organisms without benchmarked variants, which are traditionally used for distinguishing sequencing errors from real variants. In the absence of these benchmarked datasets, we leverage multiple rounds of statistical recalibration to increase the precision of variant prediction. The SNP‐SVant workflow is flexible, with user options to tradeoff accuracy for sensitivity. The workflow predicts SNPs and small insertions and deletions using the Genome Analysis ToolKit (GATK) and predicts SVs using the Genome Rearrangement IDentification Software Suite (GRIDSS), and it culminates in variant annotation using custom scripts. A key utility of SNP‐SVant is its scalability. Variant calling is a computationally expensive procedure, and thus, SNP‐SVant uses a workflow management system with intermediary checkpoint steps to ensure efficient use of resources by minimizing redundant computations and omitting steps where dependent files are available. SNP‐SVant also provides metrics to assess the quality of called variants and converts between VCF and aligned FASTA format outputs to ensure compatibility with downstream tools to calculate selection statistics, which are commonplace in population genomics studies. By accounting for both small and large structural variants, users of this workflow can obtain a wide‐ranging view of genomic alterations in an organism of interest. Overall, this workflow advances our capabilities in assessing the functional consequences of different types of genomic alterations, ultimately improving our ability to associate genotypes with phenotypes. © 2024 The Authors. Current Protocols published by Wiley Periodicals LLC. Basic Protocol : Predicting single nucleotide polymorphisms and structural variations Support Protocol 1 : Downloading publicly available sequencing data Support Protocol 2 : Visualizing variant loci using Integrated Genome Viewer Support Protocol 3 : Converting between VCF and aligned FASTA formats
Preprint
Full-text available
Background Cucurbita pepo L. cultivars display high morphological traits variation. In addition, C. pepo faces numerous threats, such as viral and fungal infections, which significantly influence crop cultivation. Recent genomic advancements improved the understanding of genetic diversity and stress responses in this crop. We investigated genetic variations related to plant morphology and quality traits. Additionally, the inclusion of both powdery mildew (PM) and Zucchini yellow mosaic virus (ZYMV) susceptible and tolerant varieties facilitated the examination of genetic diversity concerning biotic stress. Results The sequencing of eight Cucurbita pepo varieties produced an average of 40 million raw reads with a coverage of reference genome ranging from 22 to 40X. More than 4.7 million genomic variants were identified in all genomes. Based on admixture and PCA analysis, the eight C. pepo genotypes resulted organized in two clusters belonging to Cocozelle and Zucchini groups, with “Whitaker” separated from the rest of the accessions. Genes involved in pathways related to gibberellin regulation, leaf development, and pigment accumulation resulted highly affected by variation suggesting that the diversity observed among varieties in plant and fruit morphology could be related to variants identified in such genes. Each variety showed its own set of genetic differences. The genomic comparison of 381e, 968Rb and SPQ allowed the identification of variants in chromosome regions affecting response to Zucchini yellow mosaic virus (ZYMV) and powdery mildew (PM). Variants in key genes associated with resistant traits were identified, suggesting potential pathways and mechanisms involved in biotic stress response and plant immunity. Conclusions Genetic variations affecting morphology and fruit quality in C. pepo emphasize their significance for breeding efforts. Furthermore, the genomic comparison of 381e, 968Rb and SPQ highlighted variants in chromosomal regions influencing zucchini's response to PM and ZYMV. These findings could pave the way for more targeted and effective genetic improvement strategies, thereby potentially leading to increased agricultural productivity and quality.
Article
The emergence of multidrug resistance tuberculosis (MDR-TB) is caused by Mycobacterium tuberculosis (MTB) adaptation to survive in the presence of antibiotic, that were contributed by mutations in the MDR-associated genes. Previous research has indicated that the gene expression knockdown of fhaA leads to an accumulation of peptidoglycan (PG) precursors at the bacillary septum and poles, which suggest a possible deficiency in PG biosynthesis. Consequently, the cell wall becomes resistant to antibiotics, leading to multidrug resistance (MDR). In this study, bioinformatics analyses were performed on MDR-TB isolates from 24 clinical samples to search for novel mutations that contribute to antibiotic resistance. We found out a potential deletion of nucleotides encoding 6 amino acids in all 12 samples, particularly in fhaA gene (RV0020c). Our subsequent structural analysis shows that the deletion is at the position 243-248, causing conformational change of the native FhaA protein. We postulated that the deletion will potentially cause the loss of its binding affinity to MviN (precursor) and STPK (protein kinase), resulting in the inhibition and blockage of the peptidoglycan polymerization, causing MDR in MTB. In the future, experimental validation is necessitated to substantiate the association of these mutations with MDR.
Article
Full-text available
A breeding programme of aromatic vanilla, dating back to 1944, was conducted in Ambohitsara, Antalaha, SAVA (Sambava, Antalaha, Vohemara, Andapa) – Madagascar. Imported, local, wild and cultivated vanillas were used as progenitors and thousands of hybrids were generated. However, this germplasm has not undergone any genetic evaluation, and it appears that these valuable genetic resources have been dispersed or lost after the end of the programme (2000). This study aims to investigate the genetic diversity and structure of rescued genotypes currently held in a local collection in Antalaha. Double digest restriction associated-site (RAD)-seq (ddRAD)-seq protocol was applied, providing 865 million read sequences from 56 accessions. The ddRAD sequences have been deposited to the SRA archive of NCBI. From the data, 23,701 filtered concordant common Single Nucleotide Polymorphisms (SNPs) were identified using the three widely used tools (Stacks, BCFtools, Genome Analysis ToolKit - GATK) for short-read library sequencing. These SNPs were used for germplasm evaluation. Clustering analysis segregated samples into five genetic groups: Vanilla planifolia , Vanilla pompona , hybrid Tsitaitra, Vanille Banane and the phenotype Tsivaky. Our analysis revealed distinct subgroups within V. pompona and Tsitaitra, emphasizing the importance of further characterization to accurately reflect the genetic diversity and facilitate better utilization of these accessions in future research and germplasm management. The presence of private alleles in all groups (from 487 to 2866) indicated that populations were diverging and represented a large gene pool that could be useful for future breeding efforts. The genetic data obtained from this study offers valuable insights into the genetic diversity and structure of the vanilla population, with potential applications in breeding and conservation efforts.
Article
Full-text available
The wide availability of whole-genome sequencing (WGS) and an abundance of open-source software have made detection of single-nucleotide polymorphisms (SNPs) in bacterial genomes an increasingly accessible and effective tool for comparative analyses. Thus, ensuring that real nucleotide differences between genomes (i.e., true SNPs) are detected at high rates and that the influences of errors (such as false positive SNPs, ambiguously called sites, and gaps) are mitigated is of utmost importance. The choices researchers make regarding the generation and analysis of WGS data can greatly influence the accuracy of short-read sequence alignments and, therefore, the efficacy of such experiments. We studied the effects of some of these choices, including: i) depth of sequencing coverage, ii) choice of reference-guided short-read sequence assembler, iii) choice of reference genome, and iv) whether to perform read-quality filtering and trimming, on our ability to detect true SNPs and on the frequencies of errors. We performed benchmarking experiments, during which we assembled simulated and real Listeria monocytogenes strain 08-5578 short-read sequence datasets of varying quality with four commonly used assemblers (BWA, MOSAIK, Novoalign, and SMALT), using reference genomes of varying genetic distances, and with or without read pre-processing (i.e., quality filtering and trimming). We found that assemblies of at least 50-fold coverage provided the most accurate results. In addition, MOSAIK yielded the fewest errors when reads were aligned to a nearly identical reference genome, while using SMALT to align reads against a reference sequence that is ,0.82% distant from 08-5578 at the nucleotide level resulted in the detection of the greatest numbers of true SNPs and the fewest errors. Finally, we show that whether read pre-processing improves SNP detection depends upon the choice of reference sequence and assembler. In total, this study demonstrates that researchers should test a variety of conditions to achieve optimal results.
Article
Full-text available
Undetected and untreated, low-levels of drug resistant (DR) subpopulations in clinical Mycobacterium tuberculosis (Mtb) infections may lead to development of DR-tuberculosis, potentially resulting in treatment failure. Current phenotypic DR susceptibility testing has a theoretical potential for 1% sensitivity, is not quantitative, and requires several weeks to complete. The use of "single molecule-overlapping reads" (SMOR) analysis with next generation DNA sequencing for determination of ultra-rare target alleles in complex mixtures provides increased sensitivity over standard DNA sequencing. Ligation free amplicon sequencing with SMOR analysis enables the detection of resistant allele subpopulations at ≥0.1% of the total Mtb population in near real-time analysis. We describe the method using standardized mixtures of DNA from resistant and susceptible Mtb isolates and the assay's performance for detecting ultra-rare DR subpopulations in DNA extracted directly from clinical sputum samples. SMOR analysis enables rapid near real-time detection and tracking of previously undetectable DR sub-populations in clinical samples allowing for the evaluation of the clinical relevance of low-level DR subpopulations. This will provide insights into interventions aimed at suppressing minor DR subpopulations before they become clinically significant.
Article
Full-text available
The rapid development of next generation sequencing (NGS) technology provides a novel avenue for genomic exploration and research. Single nucleotide variants (SNVs) inferred from next generation sequencing are expected to reveal gene mutations in cancer. However, NGS has lower sequence coverage and poor SNVs detection capability in the regulatory regions of the genome. Post probabilistic based methods are efficient for detection of SNVs in high coverage regions or sequencing data with high depth. However, for data with low sequencing depth, the efficiency of such algorithms remains poor and needs to be improved. A new tool SNVHMM basing on a discrete hidden Markov model (HMM) was developed to infer the genotype for each position on the genome. We incorporated the mapping quality of each read and the corresponding base quality on the reads into the emission probability of HMM. The context information of the whole observation as well as its confidence were completely utilized to infer the genotype for each position on the genome in study. Therefore, more probability power can be gained over the Bayes based methods, which is very useful for SNVs detection for data with low sequencing depth. Moreover, our model was verified by testing against two sets of lobular breast tumor and Myelodysplastic Syndromes (MDS) data each. Comparing against a recently published SNVs calling algorithm SNVMix2, our model improved the performance of SNVMix2 largely when the sequencing depth is low and also outperformed SNVMix2 when SNVMix2 is well trained by large datasets. SNVHMM can detect SNVs from NGS cancer data efficiently even if the sequence depth is very low. The training data size can be very small for SNVHMM to work. SNVHMM incorporated the base quality and mapping quality of all observed bases and reads, and also provides the option for users to choose the confidence of the observation for SNVs prediction.
Article
Full-text available
Like a jigsaw puzzle with large pieces, a genome sequenced with long reads is easier to assemble. However, recent sequencing technologies have favored lowering per-base cost at the expense of read length. This has dramatically reduced sequencing cost, but resulted in fragmented assemblies, which negatively affect downstream analyses and hinder the creation of finished (gapless, high-quality) genomes. In contrast, emerging long-read sequencing technologies can now produce reads tens of kilobases in length, enabling the automated finishing of microbial genomes for under $1000. This promises to improve the quality of reference databases and facilitate new studies of chromosomal structure and variation. We present an overview of these new technologies and the methods used to assemble long reads into complete genomes. Copyright © 2014 The Authors. Published by Elsevier Ltd.. All rights reserved.
Article
Full-text available
Whole-genome sequences are now available for many microbial species and clades, however existing whole-genome alignment methods are limited in their ability to perform sequence comparisons of multiple sequences simultaneously. Here we present the Harvest suite of core-genome alignment and visualization tools for the rapid and simultaneous analysis of thousands of intraspecific microbial strains. Harvest includes Parsnp, a fast core-genome multi-aligner, and Gingr, a dynamic visual platform. Together they provide interactive core-genome alignments, variant calls, recombination detection, and phylogenetic trees. Using simulated and real data we demonstrate that our approach exhibits unrivaled speed while maintaining the accuracy of existing methods. The Harvest suite is open-source and freely available from: http://github.com/marbl/harvest. Electronic supplementary material The online version of this article (doi:10.1186/s13059-014-0524-x) contains supplementary material, which is available to authorized users.
Article
Full-text available
Protein sequences predicted from metagenomic datasets are annotated by identifying their homologs via sequence comparisons with reference or curated proteins. However, a majority of metagenomic protein sequences are partial-length, arising as a result of identifying genes on sequencing reads or on assembled nucleotide contigs, which themselves are often very fragmented. The fragmented nature of metagenomic protein predictions adversely impacts homology detection and, therefore, the quality of the overall annotation of the dataset. Here we present a novel algorithm called GRASP that accurately identifies the homologs of a given reference protein sequence from a database consisting of partial-length metagenomic proteins. Our homology detection strategy is guided by the reference sequence, and involves the simultaneous search and assembly of overlapping database sequences. GRASP was compared to three commonly used protein sequence search programs (BLASTP, PSI-BLAST and FASTM). Our evaluations using several simulated and real datasets show that GRASP has a significantly higher sensitivity than these programs while maintaining a very high specificity. GRASP can be a very useful program for detecting and quantifying taxonomic and protein family abundances in metagenomic datasets. GRASP is implemented in GNU C++, and is freely available at http://sourceforge.net/projects/grasp-release. © The Author(s) 2014. Published by Oxford University Press on behalf of Nucleic Acids Research.
Article
This unit describes how to use BWA and the Genome Analysis Toolkit (GATK) to map genome sequencing data to a reference and produce high-quality variant calls that can be used in downstream analyses. The complete workflow includes the core NGS data processing steps that are necessary to make the raw data suitable for analysis by the GATK, as well as the key methods involved in variant discovery using the GATK.