PrecisionFDA
Truth Challenge

Engage and improve DNA test results with our community challenges

The precisionFDA Truth Challenge ran from April 26, 2016 to May 26, 2016, giving participants the unique and exciting opportunity to test their NGS pipelines on an uncharacterized sample (HG002) and publish results on precisionFDA for subsequent evaluation against newly-revealed "truth" data.
This web page discusses the results and lists the awards and recognitions handed out by the precisionFDA team. Due to novelties related to both the truth data and the comparison methodology, these results are only a first cut at conducting an evaluation. In the sections below, we present some lessons learned during the process of setting up the challenge, determining the truth, and using a new comparison methodology; and we finally invite the community to further explore these results and provide insight for the future.
Community Challenge Awards
Highest
SNP Performance
in the precisionFDA Truth Challenge
Awarded to
Verily Life Sciences
  • Ryan Poplin
  • Mark DePristo
  • Verily Life Sciences Team
Highest
SNP Recall
in the precisionFDA Truth Challenge
Awarded to
Sentieon
  • Rafael Aldana
  • Hanying Feng
  • Brendan Gallagher
  • Jun Ye
Highest
SNP Precision
in the precisionFDA Truth Challenge
Awarded to
Kinghorn Center
for Clinical Genomics
  • Aaron Statham
  • Mark Cowley
  • Joseph Copty
  • Mark Pinese
Highest
INDEL Performance
in the precisionFDA Truth Challenge
Awarded to
Sanofi-Genzyme
  • Deepak Grover
Highest
INDEL Recall
in the precisionFDA Truth Challenge
Awarded to
Sanofi-Genzyme
  • Deepak Grover
Highest
INDEL Precision
in the precisionFDA Truth Challenge
Awarded to
Sentieon
  • Rafael Aldana
  • Hanying Feng
  • Brendan Gallagher
  • Jun Ye
Must-read Introductory Remarks

We are very excited to see growing numbers of challenge participants! For this challenge we received 35 entries, not only from many who participated in the previous challenge, but also from several first time participants. Some participants showcased their developing new methods, while others decided to use existing methods and see how they perform in this particular environment. We are extremely grateful to all participants for their willingness to share their results and to contribute to this effort.

This second precisionFDA challenge was conducted in collaboration with the Genome in a Bottle (GiaB) consortium, which provided the "truth" data, as well as with the Global Alliance for Genomics and Health (GA4GH), which provided best practices and software for conducting the comparison of participants' entries to the truth data. In fact, this effort represents the first time that this new truth data has been investigated (and it's only an approximation of the truth, hence sometimes we say "truth" instead of truth), and the first time that this new comparison methodology has been applied at scale across the vast number of submitted entries.

Our evaluation against the HG002 truth data has produced several metrics (such as f-score, recall and precision) across different variant types (such as SNP and indels), subtypes (such as insertions or deletions of specific size ranges) and genomic contexts (such as whole genome, coding regions, etc.), leading to thousands of computed numbers per challenge entry. We've chosen six of them (f-score, recall and precision, across SNPs and across indels, in the whole genome) to use as the basis for handing out awards and recognitions.

These results are by no means the final word. Given the originality of the truth data and the comparison methodology, we expect the community to further conduct analyses and contribute to improvements in the benchmarking methodology, the correctness of the truth data, and the definition of comparison metrics.

We would like to acknowledge and thank all of those who participated in the precisionFDA Truth Challenge. As with the previous challenge, we hope that everyone will feel like a winner.

Overview of results

The following table summarizes the challenge entries, and the results of the comparison against the HG002 truth data.

We've given each entry a unique label, comprised of the name of the submitting user as well as a short mnemonic keyword representing the pipeline. (As with the previous challenge, these keywords are merely indicative of each pipeline's main component, hence somewhat subjective; for a more faithful description of each pipeline, refer to the full text that accompanied each submission by following the label links). Each submitted entry consisted of two VCFs, corresponding to the variants called on the HG001/NA12878 and HG002/NA24385 datasets respectively. Links to these files can be found in the "Datasets" tab of the table.

The entries are sorted alphabetically. You can click on any column header to re-sort the table according to that column.

Label Submitter Organization SNP-Fscore SNP-recall SNP-precision INDEL-Fscore INDEL-recall INDEL-precision
raldana-dualsentieon Rafael Aldana   et al. Sentieon 99.9260 99.9131 99.9389 99.1095 98.7566 99.4648
bgallagher-sentieon Brendan Gallagher   et al. Sentieon 99.9296 99.9673 99.8919 99.2678 99.2143 99.3213
mlin-fermikit Mike Lin DNAnexus Science 98.8629 98.2311 99.5029 95.5997 94.8918 96.3183
jmaeng-gatk Ju Heon Maeng Yonsei University 99.6144 99.4608 99.7686 99.1098 99.0216 99.1981
ckim-dragen Changhoon Kim Macrogen 99.8268 99.9524 99.7015 99.1359 99.1574 99.1143
ckim-gatk Changhoon Kim Macrogen 99.6466 99.4788 99.8150 99.2271 99.1551 99.2992
ckim-isaac Changhoon Kim Macrogen 98.5357 97.1616 99.9494 95.8099 93.7006 98.0163
ckim-vqsr Changhoon Kim Macrogen 99.2866 98.6511 99.9303 99.2541 99.0614 99.4476
ltrigg-rtg2 Len Trigg RTG 99.8749 99.8935 99.8562 99.2539 98.8759 99.6347
ltrigg-rtg1 Len Trigg RTG 99.8754 99.8921 99.8587 99.0160 98.3355 99.7061
hfeng-pmm1 Hanying Feng   et al. Sentieon 99.9496 99.9227 99.9766 99.3397 99.0289 99.6526
hfeng-pmm2 Hanying Feng   et al. Sentieon 99.9416 99.9254 99.9579 99.3119 99.0152 99.6103
hfeng-pmm3 Hanying Feng   et al. Sentieon 99.9548 99.9339 99.9756 99.3628 99.0161 99.7120
jlack-gatk Justin Lack NIH 99.7200 99.9393 99.5016 98.6899 98.8138 98.5664
astatham-gatk Aaron Statham   et al. KCCG 99.5934 99.2091 99.9807 99.3424 99.2404 99.4446
qzeng-custom Qian Zeng LabCorp 99.4966 99.2413 99.7533 96.8316 96.8703 96.7929
anovak-vg Adam Novak   et al. vgteam 98.4545 98.3357 98.5736 70.4960 69.7491 71.2591
eyeh-varpipe ErhChan Yeh   et al. Academia Sinica 99.4670 99.9638 98.9751 92.5779 91.3854 93.8021
ciseli-custom Christian Iseli   et al. SIB 97.7648 98.8356 96.7169 83.5453 82.5314 84.5844
ccogle-snppet* Christopher Cogle   et al. CancerPOP
asubramanian-gatk Ayshwarya Subramanian   et al. Broad Institute 98.9379 97.9985 99.8954 98.8418 98.5404 99.1451
rpoplin-dv42 Ryan Poplin   et al. Verily Life Sciences 99.9587 99.9447 99.9728 98.9802 98.7882 99.1728
cchapple-custom Charles Chapple   et al. Saphetor 99.8448 99.8832 99.8063 99.1388 98.8448 99.4346
gduggal-bwafb Geet Duggal   et al. DNAnexus Science 99.7820 99.8619 99.7021 96.9474 95.5004 98.4390
gduggal-bwaplat Geet Duggal   et al. DNAnexus Science 98.8646 98.0471 99.6958 92.6621 87.0843 99.0034
gduggal-bwavard Geet Duggal   et al. DNAnexus Science 99.3249 99.0431 99.6083 87.3464 87.1769 87.5166
gduggal-snapfb Geet Duggal   et al. DNAnexus Science 99.2501 99.8026 98.7037 92.2602 90.5733 94.0112
gduggal-snapplat Geet Duggal   et al. DNAnexus Science 99.0030 98.6815 99.3266 76.4210 69.0418 85.5664
gduggal-snapvard Geet Duggal   et al. DNAnexus Science 99.0871 98.9341 99.2406 83.0264 83.4429 82.6139
ghariani-varprowl Gunjan Hariani   et al. Quintiles 99.3496 99.8685 98.8361 87.2025 87.3272 87.0781
jpowers-varprowl Jason Powers   et al. Q2 Solutions 99.5004 99.5447 99.4561 86.4885 85.2886 87.7226
dgrover-gatk Deepak Grover Sanofi-Genzyme 99.9456 99.9631 99.9282 99.4009 99.3458 99.4561
egarrison-hhga Erik Garrison   et al. - 99.8985 99.8365 99.9607 97.4253 97.1646 97.6874
jli-custom Jian Li   et al. Roche 99.9382 99.9603 99.9160 99.3675 99.0788 99.6580
ndellapenna-hhga Nicolas Della Penna ANU 99.8818 99.8118 99.9519 97.3838 97.0938 97.6756
Label HG001 VCF HG002 VCF
raldana-dualsentieon Challenge201605_Sentieon_dualmap_HG001_final.vc... Challenge201605_Sentieon_dualmap_HG002_final.vc...
bgallagher-sentieon Sentieon DNAseq Gallagher HG001.vcf Sentieon DNAseq Gallagher HG002.vcf
mlin-fermikit HG001.fermikit.raw.vcf.gz HG002.fermikit.raw.vcf.gz
jmaeng-gatk HG001-NA12878-50x_illumina_WGS_TruthChallenge.R... HG002-NA24385-50x_illumina_WGS_TruthChallenge.R...
ckim-dragen DRAGEN_HG001-NA12878-50x.hard-filtered.vcf.gz DRAGEN_HG002-NA24385-50x.vcf.gz
ckim-gatk GATK_HG001-NA12878.Filtered.Variants.vcf.gz GATK_HG002-NA24385.Filtered.Variants.vcf.gz
ckim-isaac ISAAC_HG001-NA12878_all_passed_variants.vcf.gz ISAAC_HG002-NA24385_all_passed_variants.vcf.gz
ckim-vqsr VQSR_HG001-NA12878.recalibrated_variants.vcf.gz VQSR_HG002-NA24385.recalibrated_variants.vcf.gz
ltrigg-rtg2 RTG-HG001-NA12878-50x-AVR.vcf.gz RTG-HG002-NA24385-50x-AVR.vcf.gz
ltrigg-rtg1 RTG-HG001-NA12878-50x-GQ20.vcf.gz RTG-HG002-NA24385-50x-GQ20.vcf.gz
hfeng-pmm1 Sentieon-HG001-PMM1.vcf.gz Sentieon-HG002-PMM1.vcf.gz
hfeng-pmm2 Sentieon-HG001-PMM2.vcf.gz Sentieon-HG002-PMM2.vcf.gz
hfeng-pmm3 Sentieon-HG001-PMM3.vcf.gz Sentieon-HG002-PMM3.vcf.gz
jlack-gatk NA12878_CCBR.vcf.gz NA24385_CCBR.vcf.gz
astatham-gatk AStatham-Garvan-HG001.hc.vqsr.vcf.gz AStatham-Garvan-HG002.hc.vqsr.vcf.gz
qzeng-custom LabCorp_HG001.vcf.gz LabCorp_HG002.vcf.gz
anovak-vg vg_may.HG001.vcf.gz vg_may.HG002.vcf.gz
eyeh-varpipe HG001-NA12878_varpipe_1.vcf HG002-NA24385_varpipe_1.vcf
ciseli-custom SIB-HG001.vcf.gz SIB-HG002.vcf.gz
ccogle-snppet HG001_NA12878_CancerPOP.vcf HG002_NA24385_CancerPOP.vcf
asubramanian-gatk pfda-truth-bi.hg001-rerun2.vcf.gz pfda-truth-bi.hg002-rerun2.vcf.gz
rpoplin-dv42 HG001-NA12878-pFDA.vcf.gz HG002-NA24385-pFDA.vcf.gz
cchapple-custom Saphetor-hg001.vcf.gz Saphetor-hg002.vcf.gz
gduggal-bwafb NA12878-bwa-HG001-freebayes-keeplcr.vcf.gz NA24385-bwa-HG002-freebayes-keeplcr.vcf.gz
gduggal-bwaplat NA12878-bwa-HG001-platypus-keeplcr.vcf.gz NA24385-bwa-HG002-platypus-keeplcr.vcf.gz
gduggal-bwavard NA12878-bwa-HG001-vardict-keeplcr.vcf.gz NA24385-bwa-HG002-vardict-keeplcr.vcf.gz
gduggal-snapfb NA12878-snap-HG001-freebayes-keeplcr.vcf.gz NA24385-snap-HG002-freebayes-keeplcr.vcf.gz
gduggal-snapplat NA12878-snap-HG001-platypus-keeplcr.vcf.gz NA24385-snap-HG002-platypus-keeplcr.vcf.gz
gduggal-snapvard NA12878-snap-HG001-vardict-keeplcr.vcf.gz NA24385-snap-HG002-vardict-keeplcr.vcf.gz
ghariani-varprowl default_with_model_NA12878.vcf.gz default_with_model_NA24385.vcf.gz
jpowers-varprowl HG001-NA12878-50x.vcf.gz HG002-NA24385-50x.vcf.gz
dgrover-gatk HG001-NA12878-dgrover.vcf.gz HG002-NA24385-dgrover.vcf.gz
egarrison-hhga HG001d.wg.hhga_yeoq.model.98.vcf.gz HG002d.wg.hhga_yeoq.model.98.vcf.gz
jli-custom pfdaTC_Bina_HG001.vcf.gz pfdaTC_Bina_HG002.vcf.gz
ndellapenna-hhga HG001d.wg.hhga_refined_prague.model.18.vcf.gz HG002d.wg.refined_prague.model.18.vcf.gz
*ccogle-snppet: This entry was not considered because the entry did not call variants across the whole genome

For more information about how these results were calculated, consult the comparison section below.

We are handing out the following community challenge awards:

  • highest-snp-performance to the entry submitted by Ryan Poplin et al. from Verily Life Sciences, for achieving the highest SNP F-score.
  • highest-snp-recall to the entry submitted by Brendan Gallagher et al. from Sentieon (labeled bgallagher-sentieon), for achieving the highest SNP recall.
  • highest-snp-precision to the entry submitted by Aaron Statham et al. from Kinghorn Center for Clinical Genomics, for achieving the highest SNP precision.
  • highest-indel-performance to the entry submitted by Deepak Grover from Sanofi-Genzyme, for achieving the highest indel F-score.
  • highest-indel-recall to the entry submitted by Deepak Grover from Sanofi-Genzyme, for achieving the highest indel recall.
  • highest-indel-precision to the entry submitted by Hanying Feng et al. from Sentieon (labeled hfeng-pmm3), for achieving the highest indel precision.

We are also handing out the following recognitions:

  • high-snp-performance to all entries achieving a SNP F-score of 99.920% or higher.
  • high-snp-recall to all entries achieving a SNP recall of 99.910% or higher.
  • high-snp-precision to all entries achieving a SNP precision of 99.920% or higher.
  • high-indel-performance to all entries achieving an indel F-score of 99.310% or higher.
  • high-indel-recall to all entries achieving an indel recall of 99.150% or higher.
  • high-indel-precision to all entries achieving an indel precision of 99.430% or higher.
Truth Data

One way of learning about the performance of software pipelines for processing human genome sequencing data is by using well-characterized datasets such as NA12878/HG001. The previous challenge made use of the GiaB/NIST NA12878 v2.19 truth data as part of the evaluation of accuracy. But considering how widely HG001 has been used to train software pipelines, this challenge focuses on a new sample, HG002, whose truth data was not yet available at the time of the challenge.

In particular, this challenge provided sequencing reads datasets (FASTQ) for both HG001 and HG002. During the challenge, participants had access to the aforementioned HG001 v2.19 truth data, but not to the HG002 truth data. Upon the closing of the Truth Challenge, the Genome in a Bottle (GiaB) consortium, led by the National Institute of Standards (NIST), released the HG002 truth data, as well as an updated version of the HG001 truth data. Both were initially assigned version 3.2.

The truth data files have been published on precisionFDA by Dr. Justin Zook (scientist at the National Institute of Standards and Technology and co-leader of the Genome in a Bottle Consortium efforts), in the precisionFDA discussion topic titled Help improve the new v3.2 GIAB benchmark calls for HG001 and HG002.

For your convenience, we've included below a copy of the announcement. It discusses known issues of the truth data as well as other important caveats, and we encourage you to read it to understand how this new truth data may have affected the results of the challenge, or may interact with your own evaluations in the future.

NEW RESULTS

Based on feedback received about our v3.2 calls, we have now uploaded a new version 3.2.2 set of calls that fixes a significant number of erroneously low confidence variants on chr7 and chr9, and also now excludes regions homologous to the hs37d5 decoy sequence.

Changes in v3.2.2: We found a bug in one of the input call sets that was causing many of the variants on chr7 in HG001 and HG002 and on chr9 on HG002 to be reported as not high confidence. We have corrected this issue, and all other chromosomes are the same as in v3.2.1.

Changes in v3.2.1: The only change in v3.2.1 is to exclude regions from our high confidence bed file if they have homology to the decoy sequence hs37d5. These decoy-related regions were generated by Heng Li and are available as mm-2-merged.bed at https://github.com/ga4gh/benchmarking-tools/tree/master/resources/stratification-bed-files/SegmentalDuplications. Our high-confidence vcf generally does not include variants if they are homologous to decoy sequence. Although most of these likely should be excluded, some real variants may be missed, so we are excluding these regions from our high-confidence regions until they are better characterized. However, it is important to note that these regions may be enriched for false positives if the decoy is not used in mapping.

Description of v3.2: We have been developing a new, reproducible integration process to create high-confidence SNP, small indel, and homozygous reference calls for benchmarking small variant calls. We have used this new process to generate v3.2 of our high-confidence calls for HG001 (aka NA12878), as well as the first high-confidence callset for HG002 (aka Ashkenazim Jewish Son), which are attached to this discussion and under ftp://ftp-trace.ncbi.nlm.nih.gov/giab/ftp/release/. These calls were generated only from data generated from our NIST Reference Material batch of cells. Planned future work will include (1) generating similar callsets on the GIAB/PGP AJ parents and Chinese trio, (2) incorporating additional datasets from these genomes to characterize more difficult regions, (3) incorporating pedigree information for phasing and to improve the calls, (4) developing better methods for homopolymer characterization, (5) making calls on X and Y for males, and (6) characterizing larger indels and structural variation. We greatly appreciate the feedback many of you have given already, and we very much welcome additional feedback to help us improve future versions of these calls as comments in this discussion, and about specific challenging variants at http://goo.gl/forms/OCUnvDXMEt1NEX8m2.

We highly recommend reading the information below prior to using these calls to understand how best to use them and their limitations. There is also a draft google doc that describes the methods used to form these calls and some preliminary comparisons to other callsets at: https://docs.google.com/document/d/1NE-NGSMslFFndQpbqnvu4wfnwgmgOcQu6KIjSbCauYg/edit?usp=sharing

Best Practices for Using High-confidence Calls:

Benchmarking variant calls is a complex process, and best practices are still being developed by the Global Alliance for Genomics and Health (GA4GH) Benchmarking Team (https://github.com/ga4gh/benchmarking-tools/). Several things are important to consider when benchmarking variant call accuracy:

  1. Complex variants (e.g., nearby SNPs and indels or block substitutions) can often be represented correctly in multiple ways in the vcf format. Therefore, we recommend using sophisticated benchmarking tools like those developed by members of the GA4GH benchmarking team. Currently, the most developed tools are rtgtools vcfeval (http://realtimegenomics.com/products/rtg-tools/) or hap.py (https://github.com/Illumina/hap.py). See supplementary information in https://docs.google.com/document/d/1NE-NGSMslFFndQpbqnvu4wfnwgmgOcQu6KIjSbCauYg/edit?usp=sharing for a suggested process for doing comparisons.
  2. By their nature, high-confidence variant calls and regions tend to include a subset of variants and regions that are easier to characterize. Accuracy of variant calls outside the high-confidence regions is generally likely to be lower than inside the high-confidence regions, so benchmarking against high-confidence calls will usually overestimate accuracy for all variants in the genome. Similarly, it is possible that a variant calling method has higher accuracy statistics compared to other methods when compared to the high-confidence variant calls but has lower accuracy compared to other methods for all variants in the genome.
  3. Stratification of performance for different variant types and different genome contexts can be very useful when assessing performance, because performance often differs between variant types and genome contexts. In addition, stratification can elucidate variant types and genome contexts that fall primarily outside high-confidence regions. Standardized bed files for stratifying by genome context are available from GA4GH at https://github.com/ga4gh/benchmarking-tools/tree/master/resources/stratification-bed-files.
  4. Particularly for targeted sequencing, it is critical to calculate confidence intervals around statistics like sensitivity because there may be very few examples of variants of some types in the benchmark calls in the targeted regions. Manual curation of sequence data in a genome browser for a subset of false positives and false negatives is essential for an accurate understanding of statistics like sensitivity and precision. Curation can often help elucidate whether the benchmark callset is wrong, the test callset is wrong, both callsets are wrong, or the true answer is unclear from current technologies. If you find questionable or challenging sites, reporting them will help improve future callsets. You can report information about particular sites at http://goo.gl/forms/OCUnvDXMEt1NEX8m2.

Known issues with high-confidence calls:

There are a number of known issues in the current high-confidence callset, which we plan to address in future releases:

  1. Compound heterozygous sites are sometimes represented as a heterozygous deletion that overlaps with a heterozygous SNP. At these sites, the genotype “0/1” or “0|1” should not be interpreted as meaning that the “0” haplotype is reference. Hap.py and vcfeval (with the --ref-overlap flag) interpret these sites as intended.
  2. The callset currently only includes local phasing information from freebayes, which does not use the PS field, so that locally phased variants should not be assumed to be phased with variants that are phased at a distant location.
  3. The high-confidence calls have an increased FP, FN, and genotyping error rate for single base indels in homopolymers. Some of these sites have a low allele fraction variant in multiple technologies so that it is unclear whether the site is a systematic sequencing error or a true variant in a fraction of the cells. Some of these sites are listed in https://docs.google.com/spreadsheets/d/1kHgRLinYcnxX3-ulvijf2HrIdrQWz5R5PtxZS-_s6ZM/edit?usp=sharing

Differences between v2.19 and v3.2 integration methods:

The new v3.2 integration methods differ from the previous GIAB calls (v2.18 and v2.19) in several ways, both in the data used and the integration process and heuristics:

  1. Only 4 datasets were used, selected because they represent 4 sequencing technologies that were generated from the NIST RM 8398 batch of DNA. They are: 300x Illumina paired end WGS, 100x Complete Genomics WGS, 1000x Ion exome sequencing, and SOLID WGS.
  2. Mapping and variant calling algorithms designed specifically for each technology were used to generate sensitive variant callsets where possible: novoalign + GATK-haplotypecaller and freebayes for Illumina, the vcfBeta file from the standard Complete Genomics pipeline, tmap+TVC for Ion exome, and Lifescope+GATK-HC for SOLID. This is intended to minimize bias towards any particular bioinformatics toolchain.
  3. Instead of forcing GATK to call genotypes at candidate variants in the bam files from each technology, we generate sensitive variant call sets and a bed file that describes the regions that were callable from each dataset. For Illumina, we used GATK callable loci to find regions with at least 20 reads with MQ >= 20 and with coverage less than 2x the median. For Complete Genomics, we used the callable regions defined by the vcfBeta file and excluded +-50bp around any no-called or half-called variant. For Ion, we intersected the exome targeted regions with callableloci (requiring at least 20 reads with MQ >= 20). Due to the shorter reads and low coverage for SOLID, it was only used to confirm variants, so no regions were considered callable.
  4. A new file with putative structural variants was used to exclude potential errors around SVs. These were SVs derived from multiple PacBio callers and multiple integrated illumina callers using MetaSV. These make up a significantly smaller fraction of the calls and genome (~4.5%) than the previous bed, which was a union of all dbVar calls for NA12878 (~10%).
  5. Homopolymers >10bp in length, including those interrupted by one nucleotide different from the homopolymer, are now excluded from the high-confidence regions, because these were seen to have a high error rate for all technologies. This constitutes a high fraction of the putative indel calls, so more work is needed to resolve these.
  6. A bug that caused nearby variants to be missed in v2.19 is fixed in the new calls.
  7. The new vcf contains variants outside the high-confidence bed file. This enables more robust comparison of complex variants or nearby variants that are near the boundary of the bed file. It also allows the user to evaluate concordance outside the high-confidence regions, but these concordance metrics should be interpreted with great care.
  8. We now output local phasing information when it is provided in the Illumina freebayes calls.

Shortly after the initial publication of v3.2 of the truth data, there were two subsequent updates leading to v3.2.1 and v3.2.2. These updates introduced improvements to the truth data based on feedback from the community and from artifacts discovered in initial evaluations of the challenge entries. The final comparison results discussed on this web page are based on comparisons against the latest version of the truth data (v3.2.2).

It is important to note that the methodology for compiling the truth data has changed substantially since the release of v2.19 of the HG001 truth dataset, and therefore the same VCF file may rank differently in a comparison against HG001 v2.19 versus a comparison against HG001 v3.2.2. These changes are discussed in the section titled "Differences between v2.19 and v3.2 integration methods" in Dr. Zook's announcement.

We would like to thank the National Institute of Standards and Technology, and the Genome in a Bottle consortium, for the excellent collaboration in this precisionFDA challenge. As can be attested by them, creating a truth dataset is a complicated process. The published dataset, even with its latest v3.2.2 improvements, is only an approximation of the truth. We hope you will keep that in mind when interpreting the results of this challenge and that you will exercise caution, as the determination of truth is always an ongoing process.

Comparison methodology

The precisionFDA system includes an initial VCF comparison framework which was used in the first precisionFDA challenge. We asked participants to use that framework to compare their HG001 VCFs against the GiaB/NIST HG001 v2.19 truth data while the challenge was still ongoing, to ensure that their files were in good shape. However, for the final evaluation of the results of this challenge, we partnered with the Benchmarking Task team of the GA4GH Data Working Group, to employ an updated comparison framework.

The GA4GH benchmarking team has been discussing the difficulties of conducting proper VCF comparisons for several months now, and has devised metrics definitions, initial specifications, early software prototypes, and other benchmarking resources for implementing VCF comparison best practices. We approached the chair of the benchmarking group, Dr. Justin Zook, to consult with him on how to best use these resources in the context of this precisionFDA challenge.

The outcome of our collaboration is the utilization of a particular methodology for comparing VCFs and counting results, which was finalized and tailored for the needs of this challenge, and which we hope will form the basis of other comparisons in the future. It is available on precisionFDA as an app, titled Vcfeval + Hap.py Comparison. (Given the departure from the previous comparison technique, and until we have received enough feedback for this new method, the "Comparison" feature of precisionFDA has not yet been updated.)

Software used

This particular comparison framework consists of a specific version of Real Time Genomics' vcfeval, used for VCF comparison, and a specific version of Illumina's hap.py, used for quantification; together they form the prototype GA4GH benchmarking workflow, as illustrated here.

The vcfeval tool generates an intermediate VCF which is further quantified by hap.py. The "quantify" tool from hap.py counts and stratifies variants. It counts SNPs and INDELs separately and provides additional counts for subtypes of each variant (indels of various lengths, het / hom, stratification regions). In order to determine if a variant specifies SNPs, insertions or deletions, "quantify" performs an alignment of the corresponding REF and ALT alleles. This improves handling of MNPs/complex variants and provides results that are comparable between different methods.

Metrics calculated

The quantification process calculates the following metrics, in harmony with "Comparison Method 3" as defined in this upcoming GA4GH benchmarking spec update:

MetricDefinition
TRUTH.TPTrue positives, from the perspective of the truth data, i.e. the number of sites in the Truth Call Set for which there are paths through the Query Call Set that are consistent with all of the alleles at this site, and for which there is an accurate genotype call for the event.
QUERY.TPTrue positives, from the perspective of the query data, i.e. the number of sites in the Query Call Set for which there are paths through the Truth Call Set that are consistent with all of the alleles at this site, and for which there is an accurate genotype call for the event.
TRUTH.FNFalse negatives, i.e. the number of sites in the Truth Call Set for which there is no path through the Query Call Set that is consistent with all of the alleles at this site, or sites for which there is an inaccurate genotype call for the event. Sites with correct variant but incorrect genotype are counted here.
QUERY.FPFalse positives, i.e. the number of sites in the Query Call Set for which there is no path through the Truth Call Set that is consistent with this site. Sites with correct variant but incorrect genotype are counted here.
FP.gtThe number of false positives where the non-REF alleles in the Truth and Query Call Sets match (i.e. cases where the truth is 1/1 and the query is 0/1 or similar).
RecallTRUTH.TP / (TRUTH.TP + TRUTH.FN)
PrecisionQUERY.TP / (QUERY.TP + QUERY.FP)
F-scoreHarmonic mean of Recall and Precision

Note that recall uses TRUTH.TP whereas precision uses QUERY.TP. For recall, TRUTH.TP counts the number of truth variants reproduced in the truth set representation, which is the same for all callers and should also be consistent with the confident regions (especially important around confident region boundaries and where variants can move between categories depending on how a variant caller decides to represent them). Using QUERY.TP for recall would have introduced problems where different methods create different representations (e.g. systematically calling MNPs as insertions + deletions can skew indel numbers in the query). For calculating precision, the tool uses QUERY.TP (i.e. the precision measures the relative number of wrong calls for all query calls).

Granular reporting and region stratification

The quantification process reports results with additional granularity, according to a few different dimensions, as outlined in the four tabs of the following table:

ValueMeaning
SNPSNP or MNP variants
INDELIndels and complex variants
ValueMeaning
*Aggregate numbers of all subtypes
tiSNPs that constitute transitions
tvSNPs that constitute transversions
I1_5Insertions of length 1-5
I6_15Insertions of length 6-15
I16_PLUSInsertions of length 16 or more
D1_5Deletions of length 1-5
D6_15Deletions of length 6-15
D16_PLUSDeletions of length 16 or more
C1_5Complex variants of length 1-5
C6_15Complex variants of length 6-15
C16_PLUSComplex variants of length 16 or more
ValueMeaning
*Aggregate numbers of all genotypes
hetOnly heterozygous variant calls (0/1 or similar genotypes)
homaltOnly homozygous alternative variant calls (1/1 or similar genotypes)
hetOnly heterozygous alternative variant calls (1/2 or similar genotypes)
ValueMeaning
*Aggregate numbers not limited to any subset (but still within confident regions)
func_cdsCoding exons from RefSeq
map_l100_m2_e1Regions in which 100bp reads map to >1 location with up to 2 mismatches and up to 1 indel
map_l150_m0_e0Regions in which 150bp reads map to >1 location with up to 0 mismatches and up to 0 indels
map_l150_m2_e1Regions in which 150bp reads map to >1 location with up to 2 mismatches and up to 1 indel
map_l150_m2_e0Regions in which 150bp reads map to >1 location with up to 2 mismatches and up to 0 indels
map_l100_m1_e0Regions in which 100bp reads map to >1 location with up to 1 mismatch and up to 0 indels
map_l125_m1_e0Regions in which 125bp reads map to >1 location with up to 1 mismatch and up to 0 indels
map_l250_m2_e1Regions in which 250bp reads map to >1 location with up to 2 mismatches and up to 1 indel
map_l250_m0_e0Regions in which 250bp reads map to >1 location with up to 0 mismatches and up to 0 indels
map_l150_m1_e0Regions in which 150bp reads map to >1 location with up to 1 mismatch and up to 0 indels
map_l125_m2_e0Regions in which 125bp reads map to >1 location with up to 2 mismatches and up to 0 indels
map_l100_m0_e0Regions in which 100bp reads map to >1 location with up to 0 mismatches and up to 0 indels
map_l250_m1_e0Regions in which 250bp reads map to >1 location with up to 1 mismatch and up to 0 indels
map_l125_m2_e1Regions in which 125bp reads map to >1 location with up to 2 mismatches and up to 1 indel
map_l250_m2_e0Regions in which 250bp reads map to >1 location with up to 2 mismatches and up to 0 indels
map_l125_m0_e0Regions in which 125bp reads map to >1 location with up to 0 mismatches and up to 0 indels
map_l100_m2_e0Regions in which 100bp reads map to >1 location with up to 2 mismatches and up to 0 indels
map_sirenRegions considered difficult to map by amplab SiRen
tech_badpromoters1000 promoter regions with lowest relative coverage in Illumina, relatively GC-rich
lowcmp_SimpleRepeat_homopolymer_gt10Homopolymers >10bp in length
lowcmp_SimpleRepeat_triTR_11to50Exact 3bp tandem repeats 11-50bp in length
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRgt6_lt51bp_gt95identity_mergedTandem repeats with >6bp unit size and <51bp in length and >95% identity
lowcmp_Human_Full_Genome_TRDB_hg19_150331_all_mergedAll tandem repeats from TRDB with adjacent repeats merged
lowcmp_SimpleRepeat_quadTR_11to50Exact 4bp tandem repeats 11-50bp in length
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRlt7_gt200bp_gt95identity_mergedTandem repeats with 1-6bp unit size and >200bp in length and >95% identity
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRlt7_lt101bp_gt95identity_mergedTandem repeats with 1-6bp unit size and <101bp in length and >95% identity
lowcmp_SimpleRepeat_quadTR_gt200Exact 4bp tandem repeats >200bp in length
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRgt6_lt101bp_gt95identity_mergedTandem repeats with >6bp unit size and <101bp in length and >95% identity
lowcmp_AllRepeats_gt200bp_gt95identity_mergedAll perfect and imperfect tandem repeats >200bp in length
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRgt6_51to200bp_gt95identity_mergedTandem repeats with >6bp unit size and 51-200bp in length and >95% identity
lowcmp_Human_Full_Genome_TRDB_hg19_150331_all_gt95identity_mergedAll tandem repeats from TRDB with >95% identity
lowcmp_SimpleRepeat_triTR_gt200Exact 3bp tandem repeats >200bp in length
lowcmp_SimpleRepeat_triTR_51to200Exact 3bp tandem repeats 51-200bp in length
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRlt7_lt51bp_gt95identity_mergedTandem repeats with 1-6bp unit size and <51bp in length and >95% identity
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRlt7_51to200bp_gt95identity_mergedTandem repeats with 1-6bp unit size and 51-200bp in length and >95% identity
lowcmp_AllRepeats_51to200bp_gt95identity_mergedAll perfect and imperfect tandem repeats 51-200bp in length
lowcmp_Human_Full_Genome_TRDB_hg19_150331All tandem repeats from TRDB
lowcmp_SimpleRepeat_quadTR_51to200Exact 4bp tandem repeats 51-200bp in length
lowcmp_AllRepeats_lt51bp_gt95identity_mergedAll perfect and imperfect tandem repeats <51bp in length
lowcmp_SimpleRepeat_homopolymer_6to10All homopolymers 6-10bp in length
lowcmp_SimpleRepeat_diTR_gt200Exact 2bp tandem repeats >200bp in length
lowcmp_SimpleRepeat_diTR_11to50Exact 2bp tandem repeats 11-50bp in length
lowcmp_SimpleRepeat_diTR_51to200Exact 2bp tandem repeats 51-200bp in length
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRgt6_gt200bp_gt95identity_mergedTandem repeats with >6bp unit size and >200bp in length and >95% identity
segdupSegmental duplications in GRCh37 of all sizes, not including ALTs from GRCh37
segdupwithaltSegmental duplications in GRCh37 >10kb, including ALTs from GRCh37
decoyRegions in GRCh37 to which decoy sequences in hs37d5 map
HG001compoundhetCompound heterozygous regions in which there are 2 different variants when phased variants within 50bp are combined in HG001/NA12878
HG002compoundhetCompound heterozygous regions in which there are 2 different variants when phased variants within 50bp are combined in HG002/NA24385
HG001complexvarComplex variant regions in which there are 2 or more variants within 50bp in HG001/NA12878
HG002complexvarComplex variant regions in which there are 2 or more variants within 50bp in HG002/NA24385

These dimensions help slice and dice the results according to all the different ways in which someone may want to look at them. The special value of '*' (for Subtype, Genotype, Subset) corresponds to summary statistics (i.e. as calculated across the whole genome in the confident regions, without further stratification).

Compilation of results

We ran hap.py (and specifically the HAP-207 version, with the engine set to a GA4GH-specific version of vcfeval) to compare HG001 and HG002 submissions against the GiaB/NIST v3.2.2 HG001 and HG002 truth data respectively. Our executions excluded the entry labeled ccogle-snppet as it did not call variants in the whole genome. The entries labeled ghariani-varprowl, jpowers-varprowl and qzeng-custom contained few VCF lines which were deemed incompatible by this new comparison framework (which performs stricter checks). Upon closer inspection, this was due to incompatible REF columns (such as the strings "nan" or "AC-7GATAGAA") or non-diploid genotypes (such as 0/1/2, or even 0/1/2/3). These were a small fraction, so we decided to exclude these offending lines from this comparison.

We have made available on precisionFDA complete archives of all the input files (including both the original and adjusted files, for the entries which we had to remove offending VCF lines), and the results (including the annotated VCF as output by hap.py, and extended statistics in CSV format). These are recapitulated in a precisionFDA post.

We would like to thank the benchmarking group of the Global Alliance for Genomics and Health for their excellent collaboration throughout this precisionFDA effort.

Results discussion
The compiled results (as described in the previous section) are summarized in the table at the top of this web page. In addition to that, we have created an interactive explorer which can be used to query the full set of all CSV files that were output by the comparison, and to show the metrics for any combination of the "Type", "Subtype", "Genotype" and "Subset" dimensions. You can use that to answer questions such as "how does a particular entry perform across different genome subsets", or "which entry performs best in large deletions?", etc.

EntryTypeSubtypeSubsetGenotypeF-scoreRecallPrecisionFrac_NATruth TP Truth FNQuery TPQuery FPFP gt% FP ma
6501-6550 / 86044 show all
ndellapenna-hhgaSNPtvmap_l100_m0_e0*
98.9920
98.3490
99.6435
65.9868
10901183109013917
43.5897
gduggal-bwavardSNP*lowcmp_SimpleRepeat_homopolymer_6to10het
98.7554
98.3399
99.1744
57.4612
10900184108119055
61.1111
bgallagher-sentieonSNPtvlowcmp_SimpleRepeat_homopolymer_6to10*
99.9129
99.8992
99.9266
60.1913
10900111089685
62.5000
dgrover-gatkSNPtvlowcmp_SimpleRepeat_homopolymer_6to10*
99.9175
99.8900
99.9450
60.3037
10899121089565
83.3333
gduggal-snapvardSNPtimap_l125_m2_e0homalt
97.8264
95.9500
99.7776
68.4862
10898460107662419
79.1667
raldana-dualsentieonINDEL*lowcmp_SimpleRepeat_quadTR_11to50het
98.8701
98.1095
99.6426
59.5927
10898210108733920
51.2821
jlack-gatkSNPtvlowcmp_SimpleRepeat_homopolymer_6to10*
99.8488
99.8809
99.8167
61.3691
108981310894205
25.0000
rpoplin-dv42SNPtvlowcmp_SimpleRepeat_homopolymer_6to10*
99.9221
99.8717
99.9725
60.1200
10897141089332
66.6667
ltrigg-rtg1INDEL*lowcmp_SimpleRepeat_quadTR_11to50het
98.7118
98.0915
99.3400
56.1553
10896212108377212
16.6667
jli-customINDELD6_15lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRlt7_lt51bp_gt95identity_merged*
98.6957
98.3216
99.0726
47.6211
108961861089610298
96.0784
cchapple-customSNPtvlowcmp_SimpleRepeat_homopolymer_6to10*
99.8854
99.8625
99.9082
58.4392
108961510886106
60.0000
eyeh-varpipeSNPtvlowcmp_SimpleRepeat_homopolymer_6to10*
99.7641
99.8350
99.6932
57.1024
1089318107243314
42.4242
gduggal-snapfbSNPtimap_l125_m2_e0homalt
97.7959
95.9060
99.7619
75.2353
10893465108942614
53.8462
jpowers-varprowlSNPtimap_l125_m1_e0homalt
99.2166
98.6148
99.8259
68.6168
10892153108921915
78.9474
ghariani-varprowlSNPtimap_l125_m1_e0homalt
99.2029
98.5967
99.8167
66.5358
10890155108902015
75.0000
mlin-fermikitINDEL*lowcmp_AllRepeats_lt51bp_gt95identity_mergedhetalt
82.5569
70.6298
99.3305
60.9945
108894528109797474
100.0000
anovak-vgSNP*func_cdshet
98.2393
97.5540
98.9342
34.7247
108882731086111770
59.8291
jlack-gatkSNPtimap_l125_m1_e0homalt
99.2298
98.5695
99.8991
63.8084
1088715810887119
81.8182
ltrigg-rtg2INDELD6_15lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRlt7_lt51bp_gt95identity_merged*
98.9176
98.2314
99.6135
43.6336
10886196108264239
92.8571
ltrigg-rtg2SNPtvlowcmp_SimpleRepeat_homopolymer_6to10*
99.8349
99.7617
99.9082
57.3447
108852610888106
60.0000
ckim-isaacSNP*lowcmp_SimpleRepeat_quadTR_11to50het
97.1980
95.1981
99.2838
34.4662
10884549109517911
13.9241
ltrigg-rtg1SNPtvlowcmp_SimpleRepeat_homopolymer_6to10*
99.8395
99.7434
99.9357
57.3637
10883281088774
57.1429
jli-customSNPtvlowcmp_SimpleRepeat_homopolymer_6to10*
99.8303
99.7434
99.9174
59.7828
10883281088195
55.5556
eyeh-varpipeSNPtvmap_l150_m1_e0*
97.4561
99.7067
95.3048
77.8264
10880321081953314
2.6266
cchapple-customSNP*map_l150_m1_e0homalt
98.2123
96.4961
99.9908
66.3750
108783951087411
100.0000
hfeng-pmm3SNPtvlowcmp_SimpleRepeat_homopolymer_6to10*
99.8394
99.6976
99.9816
59.7625
10878331087521
50.0000
hfeng-pmm1SNPtvlowcmp_SimpleRepeat_homopolymer_6to10*
99.8348
99.6884
99.9816
59.7111
10877341087421
50.0000
eyeh-varpipeINDEL*lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRlt7_lt51bp_gt95identity_mergedhet
90.2830
86.8621
93.9845
38.9075
1087616451660810631029
96.8015
egarrison-hhgaSNPtvlowcmp_SimpleRepeat_homopolymer_6to10*
99.7844
99.6701
99.8990
60.1508
108753610875119
81.8182
gduggal-bwafbSNPtvlowcmp_SimpleRepeat_homopolymer_6to10*
99.6838
99.6701
99.6976
61.7359
1087536108783321
63.6364
raldana-dualsentieonSNPtvlowcmp_SimpleRepeat_homopolymer_6to10*
99.8164
99.6701
99.9632
59.9499
10875361087243
75.0000
jmaeng-gatkSNPtvlowcmp_SimpleRepeat_homopolymer_6to10*
99.8072
99.6609
99.9540
61.5860
10874371087052
40.0000
ckim-gatkSNPtvlowcmp_SimpleRepeat_homopolymer_6to10*
99.7843
99.6517
99.9173
61.4911
10873381086995
55.5556
hfeng-pmm2SNPtvlowcmp_SimpleRepeat_homopolymer_6to10*
99.8118
99.6334
99.9908
59.9447
10871401086810
0.0000
gduggal-snapfbSNPtvlowcmp_SimpleRepeat_homopolymer_6to10*
99.2467
99.6334
98.8630
62.3274
10871401086912524
19.2000
dgrover-gatkINDELD6_15lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRlt7_lt51bp_gt95identity_merged*
98.2421
98.0870
98.3978
52.9535
1087021210870177173
97.7401
ndellapenna-hhgaSNPtvlowcmp_SimpleRepeat_homopolymer_6to10*
99.7292
99.5876
99.8713
60.0265
1086645108661410
71.4286
asubramanian-gatkSNP*map_l100_m2_e1homalt
56.2073
39.0920
99.9816
79.6586
10866169301086620
0.0000
astatham-gatkINDELD6_15lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRlt7_lt51bp_gt95identity_merged*
98.1969
98.0419
98.3525
52.6692
1086521710865182178
97.8022
egarrison-hhgaINDEL*lowcmp_SimpleRepeat_quadTR_11to50het
95.1518
97.8124
92.6322
56.2191
1086524311001875754
86.1714
ghariani-varprowlSNPtvlowcmp_SimpleRepeat_homopolymer_6to10*
98.2300
99.5784
96.9176
63.0014
10865461087934649
14.1618
anovak-vgSNP*lowcmp_SimpleRepeat_homopolymer_6to10het
97.5583
98.0061
97.1145
57.2508
1086322111039328130
39.6341
ltrigg-rtg1SNPtvmap_l100_m0_e0*
98.8489
98.0061
99.7063
58.4119
1086322110862326
18.7500
ltrigg-rtg2INDELI1_5*hetalt
98.3546
97.0076
99.7395
72.4903
10860335111052929
100.0000
qzeng-customSNPtvlowcmp_SimpleRepeat_homopolymer_6to10*
99.6831
99.5234
99.8433
60.8736
108595210830177
41.1765
anovak-vgSNPtilowcmp_Human_Full_Genome_TRDB_hg19_150331homalt
95.1588
93.9924
96.3545
65.5291
1085869410969415307
73.9759
anovak-vgSNPtilowcmp_Human_Full_Genome_TRDB_hg19_150331_all_mergedhomalt
95.1588
93.9924
96.3545
65.5291
1085869410969415307
73.9759
ckim-gatkINDELD6_15lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRlt7_lt51bp_gt95identity_merged*
98.1731
97.9516
98.3956
52.7295
1085522710855177173
97.7401
mlin-fermikitSNP*lowcmp_SimpleRepeat_homopolymer_6to10het
98.9381
97.9340
99.9631
55.1737
108552291084840
0.0000
gduggal-bwafbINDELD1_5HG002compoundhet*
91.0728
88.7045
93.5709
64.7882
10853138211658801716
89.3883
Interactively explore the results dataset


The interactive explorer includes an additional column, "% FP MA", indicating the percentage of the false positives which have matching non-ref alleles (i.e. FP.gt/QUERY.FP). For SNPs, this fraction varies greatly between pipelines, but for the majority it is less than 30%. Different callers would have the fewest FPs if these matching allele false positives were excluded from the FP counts, making clear that how performance metrics are defined can have a significant effect. For indels, this fraction varies as well, but is generally higher than 50%.

Using the interactive explorer you can also filter for different subsets, such as coding regions. Nonsurprisingly, there are very few FPs and FNs in coding regions for the accurate methods.

You can also see the fraction of calls that fall outside the confidence regions (reported as FRAC_NA). For SNPs, the fraction of calls outside GiaB high-confidence regions is ~10-25%, and for indels the fraction is higher than 50% for all methods. This is an important caveat to consider, since accuracy might be lower outside GiaB high-confidence regions.

F-score, Recall and Precision

Notably, most entries did much better at SNP accuracy measures than at indel accuracy measures. The majority of entries have high SNP f-score, recall, and precision -- above 99%. The respective numbers are lower for indels, and in fact indel recall is usually lower than indel precision. It is important to note, as mentioned in the truth data announcement, that the high-confidence calls have an increased FP, FN, and genotyping error rate for single base indels in homopolymers. This may be a contributing factor to the overall lower indel performance numbers.

Just like in the precisionFDA Consistency Challenge, it appears that pipelines are overall tuned for precision. Compared to the previous challenge, however, the input datasets had higher coverage, which may be a contributing factor to higher recall.

Award determination

We've determined the winners by taking the highest value in each metric (F-score, Recall, Precision) per variant type (SNPs vs indels). Please note that the award for highest f-score has been called "highest performance", to reflect the nature of the score. For each combination of metric per variant type, we are also recognizing entries with high values of that metric, based on a cutoff at the first substantial drop in the percentages. These cutoffs are ultimately subjective, and not directly applicable to other regulatory contexts (which may require well-defined thresholds ahead of time).

These awards and recognitions are meant to encourage the community to participate in challenges and do not constitute an endorsement by the FDA.

Beyond these results

We strongly urge the community to be cautious when interpreting these results. By their nature, high-confidence variant calls and regions tend to include a subset of variants and regions that are easier to characterize. Therefore, such benchmarking against the high-confidence truth data may in fact overestimate accuracy. Manual curation of sequence data in a genome browser for a subset of false positives and false negatives is essential for an accurate understanding of statistics like sensitivity and precision.

When interpreting results within stratification sub-categories, it may be useful to calculate confidence intervals around statistics like sensitivity because there may be very few examples of variants of some types in the benchmark calls in the stratification sub-category.

These results are based on the HG002 truth data. Pipelines may perform differently when confronted with other samples. For a more complete performance assessment of software pipelines, additional samples and experiments would be required.

Heterozygosity rates in chromosome X

We wanted to present an example of other ways in which this dataset can be analyzed, as a way to stimulate the community to conduct further experimentation with the precisionFDA Truth Challenge entries. Since the HG002 sample is male (in contrast to HG001, which is female), we decided to look at the performance of variation calling pipelines in chromosome X. It should be noted that chromosome X is not included in the confident regions of the GiaB truth data for HG002, so the comparisons have not evaluated that chromosome.

In XY male individuals, locations in chromosome X outside of the pseudoautosomal regions are expected to behave as haploid, and variant calling algorithms would ideally report them as haploid calls or as diploid homozygous calls. To measure that, we used bcftools stats -f .,PASS -r X:1-60001,X:2699521-154931044,X:155260561-155270560, which measures the count of heterozygous and homozygous SNPs in chromosome X outside of the PAR. We subsequently calculated the fraction of SNPs that are heterozygous, and generated the following table.

LabelPctHet
dgrover-gatk0.00%
jli-custom0.00%
ltrigg-rtg10.00%
ltrigg-rtg20.00%
mlin-fermikit0.51%
ckim-isaac1.10%
raldana-dualsentieon1.66%
hfeng-pmm11.76%
hfeng-pmm31.97%
hfeng-pmm22.04%
ndellapenna-hhga2.21%
egarrison-hhga2.30%
bgallagher-sentieon2.34%
astatham-gatk2.36%
rpoplin-dv422.42%
ckim-dragen4.17%
cchapple-custom5.11%
ckim-gatk5.23%
ckim-vqsr5.23%
asubramanian-gatk6.00%
ciseli-custom6.10%
qzeng-custom6.23%
jlack-gatk6.32%
gduggal-bwafb7.89%
jpowers-varprowl8.52%
eyeh-varpipe10.62%
ghariani-varprowl10.94%
gduggal-snapfb11.32%
gduggal-bwavard13.09%
gduggal-bwaplat16.11%
gduggal-snapvard16.21%
gduggal-snapplat17.29%
anovak-vg31.01%

Certain algorithms (such as the ones used in the ltrigg-rtg1 and ltrigg-rtg2 entries) can be gender-aware, and will not generate heterozygous calls when configured to run with a male option. Four entries in total had a zero heterozygosity fraction.

Final remarks

We want to thank those of you who participated in this challenge! As with our first challenge, by participating and putting your results and your thoughts out in the public, you fulfilled the first and most important goal of this challenge – to engage and start sharing data.

This challenge created a rich and interesting dataset that we hope people will study further. This dataset includes not only the submitted VCF files but also the complete set of comparison results (annotated VCFs and calculated statistics). It represents the essence of this challenge, and can hopefully be useful to inform future approaches, not only in terms of regulatory science but also for reference materials and benchmarking methodologies. We hope the GiaB group will be able to make use of it to improve reference calls (particularly around indels), and the GA4GH benchmarking group will be able to use it to improve the comparison methodology. We are excited to see how else the community will use this data set in the future.

Quotes

In closing, we would like to leave you with several encouraging quotes from our collaborators and larger community, which will hopefully inspire all readers to actively engage with the precisionFDA community, or participate in one of our upcoming challenges:


“We hope that in the near future, thanks to more well-characterized truth sets, the genomics community can determine whether these more-refined pipelines are indeed beneficial for enabling Precision Data for Precision Medicine.”
“Thanks for your patience in being a guinea pig for our high confidence calls and benchmarking tools! This has been really useful for us already!”
“These challenges are potentially of as much value to those constructing truth sets as they are to the rest of the community, as they provide a means to look at a wide variety of call sets with respect to multiple truth sets and start to delve into potential biases in various truth sets and the methods used to construct them. This is an evolving process and it is great to see how GIAB and GA4GH are also moving things forward in this regard.”
“We applaud the effort by NIST and FDA to develop more truth sets that the accuracy of all algorithms can be benchmarked against. We look forward to more comprehensive truth sets and open challenges that can drive the development of better pipeline algorithms and enable precision data for precision medicine.”
“Thanks for all your help! Hopefully through this process we all learned and improved across multiple fronts, so it's already a huge success. I’m already thinking of ways some of these results could be used in publications about benchmarking.”