PrecisionFDA
Truth Challenge

Engage and improve DNA test results with our community challenges

The precisionFDA Truth Challenge ran from April 26, 2016 to May 26, 2016, giving participants the unique and exciting opportunity to test their NGS pipelines on an uncharacterized sample (HG002) and publish results on precisionFDA for subsequent evaluation against newly-revealed "truth" data.
This web page discusses the results and lists the awards and recognitions handed out by the precisionFDA team. Due to novelties related to both the truth data and the comparison methodology, these results are only a first cut at conducting an evaluation. In the sections below, we present some lessons learned during the process of setting up the challenge, determining the truth, and using a new comparison methodology; and we finally invite the community to further explore these results and provide insight for the future.
Community Challenge Awards
Highest
SNP Performance
in the precisionFDA Truth Challenge
Awarded to
Verily Life Sciences
  • Ryan Poplin
  • Mark DePristo
  • Verily Life Sciences Team
Highest
SNP Recall
in the precisionFDA Truth Challenge
Awarded to
Sentieon
  • Rafael Aldana
  • Hanying Feng
  • Brendan Gallagher
  • Jun Ye
Highest
SNP Precision
in the precisionFDA Truth Challenge
Awarded to
Kinghorn Center
for Clinical Genomics
  • Aaron Statham
  • Mark Cowley
  • Joseph Copty
  • Mark Pinese
Highest
INDEL Performance
in the precisionFDA Truth Challenge
Awarded to
Sanofi-Genzyme
  • Deepak Grover
Highest
INDEL Recall
in the precisionFDA Truth Challenge
Awarded to
Sanofi-Genzyme
  • Deepak Grover
Highest
INDEL Precision
in the precisionFDA Truth Challenge
Awarded to
Sentieon
  • Rafael Aldana
  • Hanying Feng
  • Brendan Gallagher
  • Jun Ye
Must-read Introductory Remarks

We are very excited to see growing numbers of challenge participants! For this challenge we received 35 entries, not only from many who participated in the previous challenge, but also from several first time participants. Some participants showcased their developing new methods, while others decided to use existing methods and see how they perform in this particular environment. We are extremely grateful to all participants for their willingness to share their results and to contribute to this effort.

This second precisionFDA challenge was conducted in collaboration with the Genome in a Bottle (GiaB) consortium, which provided the "truth" data, as well as with the Global Alliance for Genomics and Health (GA4GH), which provided best practices and software for conducting the comparison of participants' entries to the truth data. In fact, this effort represents the first time that this new truth data has been investigated (and it's only an approximation of the truth, hence sometimes we say "truth" instead of truth), and the first time that this new comparison methodology has been applied at scale across the vast number of submitted entries.

Our evaluation against the HG002 truth data has produced several metrics (such as f-score, recall and precision) across different variant types (such as SNP and indels), subtypes (such as insertions or deletions of specific size ranges) and genomic contexts (such as whole genome, coding regions, etc.), leading to thousands of computed numbers per challenge entry. We've chosen six of them (f-score, recall and precision, across SNPs and across indels, in the whole genome) to use as the basis for handing out awards and recognitions.

These results are by no means the final word. Given the originality of the truth data and the comparison methodology, we expect the community to further conduct analyses and contribute to improvements in the benchmarking methodology, the correctness of the truth data, and the definition of comparison metrics.

We would like to acknowledge and thank all of those who participated in the precisionFDA Truth Challenge. As with the previous challenge, we hope that everyone will feel like a winner.

Overview of results

The following table summarizes the challenge entries, and the results of the comparison against the HG002 truth data.

We've given each entry a unique label, comprised of the name of the submitting user as well as a short mnemonic keyword representing the pipeline. (As with the previous challenge, these keywords are merely indicative of each pipeline's main component, hence somewhat subjective; for a more faithful description of each pipeline, refer to the full text that accompanied each submission by following the label links). Each submitted entry consisted of two VCFs, corresponding to the variants called on the HG001/NA12878 and HG002/NA24385 datasets respectively. Links to these files can be found in the "Datasets" tab of the table.

The entries are sorted alphabetically. You can click on any column header to re-sort the table according to that column.

Label Submitter Organization SNP-Fscore SNP-recall SNP-precision INDEL-Fscore INDEL-recall INDEL-precision
raldana-dualsentieon Rafael Aldana   et al. Sentieon 99.9260 99.9131 99.9389 99.1095 98.7566 99.4648
bgallagher-sentieon Brendan Gallagher   et al. Sentieon 99.9296 99.9673 99.8919 99.2678 99.2143 99.3213
mlin-fermikit Mike Lin DNAnexus Science 98.8629 98.2311 99.5029 95.5997 94.8918 96.3183
jmaeng-gatk Ju Heon Maeng Yonsei University 99.6144 99.4608 99.7686 99.1098 99.0216 99.1981
ckim-dragen Changhoon Kim Macrogen 99.8268 99.9524 99.7015 99.1359 99.1574 99.1143
ckim-gatk Changhoon Kim Macrogen 99.6466 99.4788 99.8150 99.2271 99.1551 99.2992
ckim-isaac Changhoon Kim Macrogen 98.5357 97.1616 99.9494 95.8099 93.7006 98.0163
ckim-vqsr Changhoon Kim Macrogen 99.2866 98.6511 99.9303 99.2541 99.0614 99.4476
ltrigg-rtg2 Len Trigg RTG 99.8749 99.8935 99.8562 99.2539 98.8759 99.6347
ltrigg-rtg1 Len Trigg RTG 99.8754 99.8921 99.8587 99.0160 98.3355 99.7061
hfeng-pmm1 Hanying Feng   et al. Sentieon 99.9496 99.9227 99.9766 99.3397 99.0289 99.6526
hfeng-pmm2 Hanying Feng   et al. Sentieon 99.9416 99.9254 99.9579 99.3119 99.0152 99.6103
hfeng-pmm3 Hanying Feng   et al. Sentieon 99.9548 99.9339 99.9756 99.3628 99.0161 99.7120
jlack-gatk Justin Lack NIH 99.7200 99.9393 99.5016 98.6899 98.8138 98.5664
astatham-gatk Aaron Statham   et al. KCCG 99.5934 99.2091 99.9807 99.3424 99.2404 99.4446
qzeng-custom Qian Zeng LabCorp 99.4966 99.2413 99.7533 96.8316 96.8703 96.7929
anovak-vg Adam Novak   et al. vgteam 98.4545 98.3357 98.5736 70.4960 69.7491 71.2591
eyeh-varpipe ErhChan Yeh   et al. Academia Sinica 99.4670 99.9638 98.9751 92.5779 91.3854 93.8021
ciseli-custom Christian Iseli   et al. SIB 97.7648 98.8356 96.7169 83.5453 82.5314 84.5844
ccogle-snppet* Christopher Cogle   et al. CancerPOP
asubramanian-gatk Ayshwarya Subramanian   et al. Broad Institute 98.9379 97.9985 99.8954 98.8418 98.5404 99.1451
rpoplin-dv42 Ryan Poplin   et al. Verily Life Sciences 99.9587 99.9447 99.9728 98.9802 98.7882 99.1728
cchapple-custom Charles Chapple   et al. Saphetor 99.8448 99.8832 99.8063 99.1388 98.8448 99.4346
gduggal-bwafb Geet Duggal   et al. DNAnexus Science 99.7820 99.8619 99.7021 96.9474 95.5004 98.4390
gduggal-bwaplat Geet Duggal   et al. DNAnexus Science 98.8646 98.0471 99.6958 92.6621 87.0843 99.0034
gduggal-bwavard Geet Duggal   et al. DNAnexus Science 99.3249 99.0431 99.6083 87.3464 87.1769 87.5166
gduggal-snapfb Geet Duggal   et al. DNAnexus Science 99.2501 99.8026 98.7037 92.2602 90.5733 94.0112
gduggal-snapplat Geet Duggal   et al. DNAnexus Science 99.0030 98.6815 99.3266 76.4210 69.0418 85.5664
gduggal-snapvard Geet Duggal   et al. DNAnexus Science 99.0871 98.9341 99.2406 83.0264 83.4429 82.6139
ghariani-varprowl Gunjan Hariani   et al. Quintiles 99.3496 99.8685 98.8361 87.2025 87.3272 87.0781
jpowers-varprowl Jason Powers   et al. Q2 Solutions 99.5004 99.5447 99.4561 86.4885 85.2886 87.7226
dgrover-gatk Deepak Grover Sanofi-Genzyme 99.9456 99.9631 99.9282 99.4009 99.3458 99.4561
egarrison-hhga Erik Garrison   et al. - 99.8985 99.8365 99.9607 97.4253 97.1646 97.6874
jli-custom Jian Li   et al. Roche 99.9382 99.9603 99.9160 99.3675 99.0788 99.6580
ndellapenna-hhga Nicolas Della Penna ANU 99.8818 99.8118 99.9519 97.3838 97.0938 97.6756
Label HG001 VCF HG002 VCF
raldana-dualsentieon Challenge201605_Sentieon_dualmap_HG001_final.vc... Challenge201605_Sentieon_dualmap_HG002_final.vc...
bgallagher-sentieon Sentieon DNAseq Gallagher HG001.vcf Sentieon DNAseq Gallagher HG002.vcf
mlin-fermikit HG001.fermikit.raw.vcf.gz HG002.fermikit.raw.vcf.gz
jmaeng-gatk HG001-NA12878-50x_illumina_WGS_TruthChallenge.R... HG002-NA24385-50x_illumina_WGS_TruthChallenge.R...
ckim-dragen DRAGEN_HG001-NA12878-50x.hard-filtered.vcf.gz DRAGEN_HG002-NA24385-50x.vcf.gz
ckim-gatk GATK_HG001-NA12878.Filtered.Variants.vcf.gz GATK_HG002-NA24385.Filtered.Variants.vcf.gz
ckim-isaac ISAAC_HG001-NA12878_all_passed_variants.vcf.gz ISAAC_HG002-NA24385_all_passed_variants.vcf.gz
ckim-vqsr VQSR_HG001-NA12878.recalibrated_variants.vcf.gz VQSR_HG002-NA24385.recalibrated_variants.vcf.gz
ltrigg-rtg2 RTG-HG001-NA12878-50x-AVR.vcf.gz RTG-HG002-NA24385-50x-AVR.vcf.gz
ltrigg-rtg1 RTG-HG001-NA12878-50x-GQ20.vcf.gz RTG-HG002-NA24385-50x-GQ20.vcf.gz
hfeng-pmm1 Sentieon-HG001-PMM1.vcf.gz Sentieon-HG002-PMM1.vcf.gz
hfeng-pmm2 Sentieon-HG001-PMM2.vcf.gz Sentieon-HG002-PMM2.vcf.gz
hfeng-pmm3 Sentieon-HG001-PMM3.vcf.gz Sentieon-HG002-PMM3.vcf.gz
jlack-gatk NA12878_CCBR.vcf.gz NA24385_CCBR.vcf.gz
astatham-gatk AStatham-Garvan-HG001.hc.vqsr.vcf.gz AStatham-Garvan-HG002.hc.vqsr.vcf.gz
qzeng-custom LabCorp_HG001.vcf.gz LabCorp_HG002.vcf.gz
anovak-vg vg_may.HG001.vcf.gz vg_may.HG002.vcf.gz
eyeh-varpipe HG001-NA12878_varpipe_1.vcf HG002-NA24385_varpipe_1.vcf
ciseli-custom SIB-HG001.vcf.gz SIB-HG002.vcf.gz
ccogle-snppet HG001_NA12878_CancerPOP.vcf HG002_NA24385_CancerPOP.vcf
asubramanian-gatk pfda-truth-bi.hg001-rerun2.vcf.gz pfda-truth-bi.hg002-rerun2.vcf.gz
rpoplin-dv42 HG001-NA12878-pFDA.vcf.gz HG002-NA24385-pFDA.vcf.gz
cchapple-custom Saphetor-hg001.vcf.gz Saphetor-hg002.vcf.gz
gduggal-bwafb NA12878-bwa-HG001-freebayes-keeplcr.vcf.gz NA24385-bwa-HG002-freebayes-keeplcr.vcf.gz
gduggal-bwaplat NA12878-bwa-HG001-platypus-keeplcr.vcf.gz NA24385-bwa-HG002-platypus-keeplcr.vcf.gz
gduggal-bwavard NA12878-bwa-HG001-vardict-keeplcr.vcf.gz NA24385-bwa-HG002-vardict-keeplcr.vcf.gz
gduggal-snapfb NA12878-snap-HG001-freebayes-keeplcr.vcf.gz NA24385-snap-HG002-freebayes-keeplcr.vcf.gz
gduggal-snapplat NA12878-snap-HG001-platypus-keeplcr.vcf.gz NA24385-snap-HG002-platypus-keeplcr.vcf.gz
gduggal-snapvard NA12878-snap-HG001-vardict-keeplcr.vcf.gz NA24385-snap-HG002-vardict-keeplcr.vcf.gz
ghariani-varprowl default_with_model_NA12878.vcf.gz default_with_model_NA24385.vcf.gz
jpowers-varprowl HG001-NA12878-50x.vcf.gz HG002-NA24385-50x.vcf.gz
dgrover-gatk HG001-NA12878-dgrover.vcf.gz HG002-NA24385-dgrover.vcf.gz
egarrison-hhga HG001d.wg.hhga_yeoq.model.98.vcf.gz HG002d.wg.hhga_yeoq.model.98.vcf.gz
jli-custom pfdaTC_Bina_HG001.vcf.gz pfdaTC_Bina_HG002.vcf.gz
ndellapenna-hhga HG001d.wg.hhga_refined_prague.model.18.vcf.gz HG002d.wg.refined_prague.model.18.vcf.gz
*ccogle-snppet: This entry was not considered because the entry did not call variants across the whole genome

For more information about how these results were calculated, consult the comparison section below.

We are handing out the following community challenge awards:

  • highest-snp-performance to the entry submitted by Ryan Poplin et al. from Verily Life Sciences, for achieving the highest SNP F-score.
  • highest-snp-recall to the entry submitted by Brendan Gallagher et al. from Sentieon (labeled bgallagher-sentieon), for achieving the highest SNP recall.
  • highest-snp-precision to the entry submitted by Aaron Statham et al. from Kinghorn Center for Clinical Genomics, for achieving the highest SNP precision.
  • highest-indel-performance to the entry submitted by Deepak Grover from Sanofi-Genzyme, for achieving the highest indel F-score.
  • highest-indel-recall to the entry submitted by Deepak Grover from Sanofi-Genzyme, for achieving the highest indel recall.
  • highest-indel-precision to the entry submitted by Hanying Feng et al. from Sentieon (labeled hfeng-pmm3), for achieving the highest indel precision.

We are also handing out the following recognitions:

  • high-snp-performance to all entries achieving a SNP F-score of 99.920% or higher.
  • high-snp-recall to all entries achieving a SNP recall of 99.910% or higher.
  • high-snp-precision to all entries achieving a SNP precision of 99.920% or higher.
  • high-indel-performance to all entries achieving an indel F-score of 99.310% or higher.
  • high-indel-recall to all entries achieving an indel recall of 99.150% or higher.
  • high-indel-precision to all entries achieving an indel precision of 99.430% or higher.
Truth Data

One way of learning about the performance of software pipelines for processing human genome sequencing data is by using well-characterized datasets such as NA12878/HG001. The previous challenge made use of the GiaB/NIST NA12878 v2.19 truth data as part of the evaluation of accuracy. But considering how widely HG001 has been used to train software pipelines, this challenge focuses on a new sample, HG002, whose truth data was not yet available at the time of the challenge.

In particular, this challenge provided sequencing reads datasets (FASTQ) for both HG001 and HG002. During the challenge, participants had access to the aforementioned HG001 v2.19 truth data, but not to the HG002 truth data. Upon the closing of the Truth Challenge, the Genome in a Bottle (GiaB) consortium, led by the National Institute of Standards (NIST), released the HG002 truth data, as well as an updated version of the HG001 truth data. Both were initially assigned version 3.2.

The truth data files have been published on precisionFDA by Dr. Justin Zook (scientist at the National Institute of Standards and Technology and co-leader of the Genome in a Bottle Consortium efforts), in the precisionFDA discussion topic titled Help improve the new v3.2 GIAB benchmark calls for HG001 and HG002.

For your convenience, we've included below a copy of the announcement. It discusses known issues of the truth data as well as other important caveats, and we encourage you to read it to understand how this new truth data may have affected the results of the challenge, or may interact with your own evaluations in the future.

NEW RESULTS

Based on feedback received about our v3.2 calls, we have now uploaded a new version 3.2.2 set of calls that fixes a significant number of erroneously low confidence variants on chr7 and chr9, and also now excludes regions homologous to the hs37d5 decoy sequence.

Changes in v3.2.2: We found a bug in one of the input call sets that was causing many of the variants on chr7 in HG001 and HG002 and on chr9 on HG002 to be reported as not high confidence. We have corrected this issue, and all other chromosomes are the same as in v3.2.1.

Changes in v3.2.1: The only change in v3.2.1 is to exclude regions from our high confidence bed file if they have homology to the decoy sequence hs37d5. These decoy-related regions were generated by Heng Li and are available as mm-2-merged.bed at https://github.com/ga4gh/benchmarking-tools/tree/master/resources/stratification-bed-files/SegmentalDuplications. Our high-confidence vcf generally does not include variants if they are homologous to decoy sequence. Although most of these likely should be excluded, some real variants may be missed, so we are excluding these regions from our high-confidence regions until they are better characterized. However, it is important to note that these regions may be enriched for false positives if the decoy is not used in mapping.

Description of v3.2: We have been developing a new, reproducible integration process to create high-confidence SNP, small indel, and homozygous reference calls for benchmarking small variant calls. We have used this new process to generate v3.2 of our high-confidence calls for HG001 (aka NA12878), as well as the first high-confidence callset for HG002 (aka Ashkenazim Jewish Son), which are attached to this discussion and under ftp://ftp-trace.ncbi.nlm.nih.gov/giab/ftp/release/. These calls were generated only from data generated from our NIST Reference Material batch of cells. Planned future work will include (1) generating similar callsets on the GIAB/PGP AJ parents and Chinese trio, (2) incorporating additional datasets from these genomes to characterize more difficult regions, (3) incorporating pedigree information for phasing and to improve the calls, (4) developing better methods for homopolymer characterization, (5) making calls on X and Y for males, and (6) characterizing larger indels and structural variation. We greatly appreciate the feedback many of you have given already, and we very much welcome additional feedback to help us improve future versions of these calls as comments in this discussion, and about specific challenging variants at http://goo.gl/forms/OCUnvDXMEt1NEX8m2.

We highly recommend reading the information below prior to using these calls to understand how best to use them and their limitations. There is also a draft google doc that describes the methods used to form these calls and some preliminary comparisons to other callsets at: https://docs.google.com/document/d/1NE-NGSMslFFndQpbqnvu4wfnwgmgOcQu6KIjSbCauYg/edit?usp=sharing

Best Practices for Using High-confidence Calls:

Benchmarking variant calls is a complex process, and best practices are still being developed by the Global Alliance for Genomics and Health (GA4GH) Benchmarking Team (https://github.com/ga4gh/benchmarking-tools/). Several things are important to consider when benchmarking variant call accuracy:

  1. Complex variants (e.g., nearby SNPs and indels or block substitutions) can often be represented correctly in multiple ways in the vcf format. Therefore, we recommend using sophisticated benchmarking tools like those developed by members of the GA4GH benchmarking team. Currently, the most developed tools are rtgtools vcfeval (http://realtimegenomics.com/products/rtg-tools/) or hap.py (https://github.com/Illumina/hap.py). See supplementary information in https://docs.google.com/document/d/1NE-NGSMslFFndQpbqnvu4wfnwgmgOcQu6KIjSbCauYg/edit?usp=sharing for a suggested process for doing comparisons.
  2. By their nature, high-confidence variant calls and regions tend to include a subset of variants and regions that are easier to characterize. Accuracy of variant calls outside the high-confidence regions is generally likely to be lower than inside the high-confidence regions, so benchmarking against high-confidence calls will usually overestimate accuracy for all variants in the genome. Similarly, it is possible that a variant calling method has higher accuracy statistics compared to other methods when compared to the high-confidence variant calls but has lower accuracy compared to other methods for all variants in the genome.
  3. Stratification of performance for different variant types and different genome contexts can be very useful when assessing performance, because performance often differs between variant types and genome contexts. In addition, stratification can elucidate variant types and genome contexts that fall primarily outside high-confidence regions. Standardized bed files for stratifying by genome context are available from GA4GH at https://github.com/ga4gh/benchmarking-tools/tree/master/resources/stratification-bed-files.
  4. Particularly for targeted sequencing, it is critical to calculate confidence intervals around statistics like sensitivity because there may be very few examples of variants of some types in the benchmark calls in the targeted regions. Manual curation of sequence data in a genome browser for a subset of false positives and false negatives is essential for an accurate understanding of statistics like sensitivity and precision. Curation can often help elucidate whether the benchmark callset is wrong, the test callset is wrong, both callsets are wrong, or the true answer is unclear from current technologies. If you find questionable or challenging sites, reporting them will help improve future callsets. You can report information about particular sites at http://goo.gl/forms/OCUnvDXMEt1NEX8m2.

Known issues with high-confidence calls:

There are a number of known issues in the current high-confidence callset, which we plan to address in future releases:

  1. Compound heterozygous sites are sometimes represented as a heterozygous deletion that overlaps with a heterozygous SNP. At these sites, the genotype “0/1” or “0|1” should not be interpreted as meaning that the “0” haplotype is reference. Hap.py and vcfeval (with the --ref-overlap flag) interpret these sites as intended.
  2. The callset currently only includes local phasing information from freebayes, which does not use the PS field, so that locally phased variants should not be assumed to be phased with variants that are phased at a distant location.
  3. The high-confidence calls have an increased FP, FN, and genotyping error rate for single base indels in homopolymers. Some of these sites have a low allele fraction variant in multiple technologies so that it is unclear whether the site is a systematic sequencing error or a true variant in a fraction of the cells. Some of these sites are listed in https://docs.google.com/spreadsheets/d/1kHgRLinYcnxX3-ulvijf2HrIdrQWz5R5PtxZS-_s6ZM/edit?usp=sharing

Differences between v2.19 and v3.2 integration methods:

The new v3.2 integration methods differ from the previous GIAB calls (v2.18 and v2.19) in several ways, both in the data used and the integration process and heuristics:

  1. Only 4 datasets were used, selected because they represent 4 sequencing technologies that were generated from the NIST RM 8398 batch of DNA. They are: 300x Illumina paired end WGS, 100x Complete Genomics WGS, 1000x Ion exome sequencing, and SOLID WGS.
  2. Mapping and variant calling algorithms designed specifically for each technology were used to generate sensitive variant callsets where possible: novoalign + GATK-haplotypecaller and freebayes for Illumina, the vcfBeta file from the standard Complete Genomics pipeline, tmap+TVC for Ion exome, and Lifescope+GATK-HC for SOLID. This is intended to minimize bias towards any particular bioinformatics toolchain.
  3. Instead of forcing GATK to call genotypes at candidate variants in the bam files from each technology, we generate sensitive variant call sets and a bed file that describes the regions that were callable from each dataset. For Illumina, we used GATK callable loci to find regions with at least 20 reads with MQ >= 20 and with coverage less than 2x the median. For Complete Genomics, we used the callable regions defined by the vcfBeta file and excluded +-50bp around any no-called or half-called variant. For Ion, we intersected the exome targeted regions with callableloci (requiring at least 20 reads with MQ >= 20). Due to the shorter reads and low coverage for SOLID, it was only used to confirm variants, so no regions were considered callable.
  4. A new file with putative structural variants was used to exclude potential errors around SVs. These were SVs derived from multiple PacBio callers and multiple integrated illumina callers using MetaSV. These make up a significantly smaller fraction of the calls and genome (~4.5%) than the previous bed, which was a union of all dbVar calls for NA12878 (~10%).
  5. Homopolymers >10bp in length, including those interrupted by one nucleotide different from the homopolymer, are now excluded from the high-confidence regions, because these were seen to have a high error rate for all technologies. This constitutes a high fraction of the putative indel calls, so more work is needed to resolve these.
  6. A bug that caused nearby variants to be missed in v2.19 is fixed in the new calls.
  7. The new vcf contains variants outside the high-confidence bed file. This enables more robust comparison of complex variants or nearby variants that are near the boundary of the bed file. It also allows the user to evaluate concordance outside the high-confidence regions, but these concordance metrics should be interpreted with great care.
  8. We now output local phasing information when it is provided in the Illumina freebayes calls.

Shortly after the initial publication of v3.2 of the truth data, there were two subsequent updates leading to v3.2.1 and v3.2.2. These updates introduced improvements to the truth data based on feedback from the community and from artifacts discovered in initial evaluations of the challenge entries. The final comparison results discussed on this web page are based on comparisons against the latest version of the truth data (v3.2.2).

It is important to note that the methodology for compiling the truth data has changed substantially since the release of v2.19 of the HG001 truth dataset, and therefore the same VCF file may rank differently in a comparison against HG001 v2.19 versus a comparison against HG001 v3.2.2. These changes are discussed in the section titled "Differences between v2.19 and v3.2 integration methods" in Dr. Zook's announcement.

We would like to thank the National Institute of Standards and Technology, and the Genome in a Bottle consortium, for the excellent collaboration in this precisionFDA challenge. As can be attested by them, creating a truth dataset is a complicated process. The published dataset, even with its latest v3.2.2 improvements, is only an approximation of the truth. We hope you will keep that in mind when interpreting the results of this challenge and that you will exercise caution, as the determination of truth is always an ongoing process.

Comparison methodology

The precisionFDA system includes an initial VCF comparison framework which was used in the first precisionFDA challenge. We asked participants to use that framework to compare their HG001 VCFs against the GiaB/NIST HG001 v2.19 truth data while the challenge was still ongoing, to ensure that their files were in good shape. However, for the final evaluation of the results of this challenge, we partnered with the Benchmarking Task team of the GA4GH Data Working Group, to employ an updated comparison framework.

The GA4GH benchmarking team has been discussing the difficulties of conducting proper VCF comparisons for several months now, and has devised metrics definitions, initial specifications, early software prototypes, and other benchmarking resources for implementing VCF comparison best practices. We approached the chair of the benchmarking group, Dr. Justin Zook, to consult with him on how to best use these resources in the context of this precisionFDA challenge.

The outcome of our collaboration is the utilization of a particular methodology for comparing VCFs and counting results, which was finalized and tailored for the needs of this challenge, and which we hope will form the basis of other comparisons in the future. It is available on precisionFDA as an app, titled Vcfeval + Hap.py Comparison. (Given the departure from the previous comparison technique, and until we have received enough feedback for this new method, the "Comparison" feature of precisionFDA has not yet been updated.)

Software used

This particular comparison framework consists of a specific version of Real Time Genomics' vcfeval, used for VCF comparison, and a specific version of Illumina's hap.py, used for quantification; together they form the prototype GA4GH benchmarking workflow, as illustrated here.

The vcfeval tool generates an intermediate VCF which is further quantified by hap.py. The "quantify" tool from hap.py counts and stratifies variants. It counts SNPs and INDELs separately and provides additional counts for subtypes of each variant (indels of various lengths, het / hom, stratification regions). In order to determine if a variant specifies SNPs, insertions or deletions, "quantify" performs an alignment of the corresponding REF and ALT alleles. This improves handling of MNPs/complex variants and provides results that are comparable between different methods.

Metrics calculated

The quantification process calculates the following metrics, in harmony with "Comparison Method 3" as defined in this upcoming GA4GH benchmarking spec update:

MetricDefinition
TRUTH.TPTrue positives, from the perspective of the truth data, i.e. the number of sites in the Truth Call Set for which there are paths through the Query Call Set that are consistent with all of the alleles at this site, and for which there is an accurate genotype call for the event.
QUERY.TPTrue positives, from the perspective of the query data, i.e. the number of sites in the Query Call Set for which there are paths through the Truth Call Set that are consistent with all of the alleles at this site, and for which there is an accurate genotype call for the event.
TRUTH.FNFalse negatives, i.e. the number of sites in the Truth Call Set for which there is no path through the Query Call Set that is consistent with all of the alleles at this site, or sites for which there is an inaccurate genotype call for the event. Sites with correct variant but incorrect genotype are counted here.
QUERY.FPFalse positives, i.e. the number of sites in the Query Call Set for which there is no path through the Truth Call Set that is consistent with this site. Sites with correct variant but incorrect genotype are counted here.
FP.gtThe number of false positives where the non-REF alleles in the Truth and Query Call Sets match (i.e. cases where the truth is 1/1 and the query is 0/1 or similar).
RecallTRUTH.TP / (TRUTH.TP + TRUTH.FN)
PrecisionQUERY.TP / (QUERY.TP + QUERY.FP)
F-scoreHarmonic mean of Recall and Precision

Note that recall uses TRUTH.TP whereas precision uses QUERY.TP. For recall, TRUTH.TP counts the number of truth variants reproduced in the truth set representation, which is the same for all callers and should also be consistent with the confident regions (especially important around confident region boundaries and where variants can move between categories depending on how a variant caller decides to represent them). Using QUERY.TP for recall would have introduced problems where different methods create different representations (e.g. systematically calling MNPs as insertions + deletions can skew indel numbers in the query). For calculating precision, the tool uses QUERY.TP (i.e. the precision measures the relative number of wrong calls for all query calls).

Granular reporting and region stratification

The quantification process reports results with additional granularity, according to a few different dimensions, as outlined in the four tabs of the following table:

ValueMeaning
SNPSNP or MNP variants
INDELIndels and complex variants
ValueMeaning
*Aggregate numbers of all subtypes
tiSNPs that constitute transitions
tvSNPs that constitute transversions
I1_5Insertions of length 1-5
I6_15Insertions of length 6-15
I16_PLUSInsertions of length 16 or more
D1_5Deletions of length 1-5
D6_15Deletions of length 6-15
D16_PLUSDeletions of length 16 or more
C1_5Complex variants of length 1-5
C6_15Complex variants of length 6-15
C16_PLUSComplex variants of length 16 or more
ValueMeaning
*Aggregate numbers of all genotypes
hetOnly heterozygous variant calls (0/1 or similar genotypes)
homaltOnly homozygous alternative variant calls (1/1 or similar genotypes)
hetOnly heterozygous alternative variant calls (1/2 or similar genotypes)
ValueMeaning
*Aggregate numbers not limited to any subset (but still within confident regions)
func_cdsCoding exons from RefSeq
map_l100_m2_e1Regions in which 100bp reads map to >1 location with up to 2 mismatches and up to 1 indel
map_l150_m0_e0Regions in which 150bp reads map to >1 location with up to 0 mismatches and up to 0 indels
map_l150_m2_e1Regions in which 150bp reads map to >1 location with up to 2 mismatches and up to 1 indel
map_l150_m2_e0Regions in which 150bp reads map to >1 location with up to 2 mismatches and up to 0 indels
map_l100_m1_e0Regions in which 100bp reads map to >1 location with up to 1 mismatch and up to 0 indels
map_l125_m1_e0Regions in which 125bp reads map to >1 location with up to 1 mismatch and up to 0 indels
map_l250_m2_e1Regions in which 250bp reads map to >1 location with up to 2 mismatches and up to 1 indel
map_l250_m0_e0Regions in which 250bp reads map to >1 location with up to 0 mismatches and up to 0 indels
map_l150_m1_e0Regions in which 150bp reads map to >1 location with up to 1 mismatch and up to 0 indels
map_l125_m2_e0Regions in which 125bp reads map to >1 location with up to 2 mismatches and up to 0 indels
map_l100_m0_e0Regions in which 100bp reads map to >1 location with up to 0 mismatches and up to 0 indels
map_l250_m1_e0Regions in which 250bp reads map to >1 location with up to 1 mismatch and up to 0 indels
map_l125_m2_e1Regions in which 125bp reads map to >1 location with up to 2 mismatches and up to 1 indel
map_l250_m2_e0Regions in which 250bp reads map to >1 location with up to 2 mismatches and up to 0 indels
map_l125_m0_e0Regions in which 125bp reads map to >1 location with up to 0 mismatches and up to 0 indels
map_l100_m2_e0Regions in which 100bp reads map to >1 location with up to 2 mismatches and up to 0 indels
map_sirenRegions considered difficult to map by amplab SiRen
tech_badpromoters1000 promoter regions with lowest relative coverage in Illumina, relatively GC-rich
lowcmp_SimpleRepeat_homopolymer_gt10Homopolymers >10bp in length
lowcmp_SimpleRepeat_triTR_11to50Exact 3bp tandem repeats 11-50bp in length
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRgt6_lt51bp_gt95identity_mergedTandem repeats with >6bp unit size and <51bp in length and >95% identity
lowcmp_Human_Full_Genome_TRDB_hg19_150331_all_mergedAll tandem repeats from TRDB with adjacent repeats merged
lowcmp_SimpleRepeat_quadTR_11to50Exact 4bp tandem repeats 11-50bp in length
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRlt7_gt200bp_gt95identity_mergedTandem repeats with 1-6bp unit size and >200bp in length and >95% identity
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRlt7_lt101bp_gt95identity_mergedTandem repeats with 1-6bp unit size and <101bp in length and >95% identity
lowcmp_SimpleRepeat_quadTR_gt200Exact 4bp tandem repeats >200bp in length
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRgt6_lt101bp_gt95identity_mergedTandem repeats with >6bp unit size and <101bp in length and >95% identity
lowcmp_AllRepeats_gt200bp_gt95identity_mergedAll perfect and imperfect tandem repeats >200bp in length
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRgt6_51to200bp_gt95identity_mergedTandem repeats with >6bp unit size and 51-200bp in length and >95% identity
lowcmp_Human_Full_Genome_TRDB_hg19_150331_all_gt95identity_mergedAll tandem repeats from TRDB with >95% identity
lowcmp_SimpleRepeat_triTR_gt200Exact 3bp tandem repeats >200bp in length
lowcmp_SimpleRepeat_triTR_51to200Exact 3bp tandem repeats 51-200bp in length
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRlt7_lt51bp_gt95identity_mergedTandem repeats with 1-6bp unit size and <51bp in length and >95% identity
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRlt7_51to200bp_gt95identity_mergedTandem repeats with 1-6bp unit size and 51-200bp in length and >95% identity
lowcmp_AllRepeats_51to200bp_gt95identity_mergedAll perfect and imperfect tandem repeats 51-200bp in length
lowcmp_Human_Full_Genome_TRDB_hg19_150331All tandem repeats from TRDB
lowcmp_SimpleRepeat_quadTR_51to200Exact 4bp tandem repeats 51-200bp in length
lowcmp_AllRepeats_lt51bp_gt95identity_mergedAll perfect and imperfect tandem repeats <51bp in length
lowcmp_SimpleRepeat_homopolymer_6to10All homopolymers 6-10bp in length
lowcmp_SimpleRepeat_diTR_gt200Exact 2bp tandem repeats >200bp in length
lowcmp_SimpleRepeat_diTR_11to50Exact 2bp tandem repeats 11-50bp in length
lowcmp_SimpleRepeat_diTR_51to200Exact 2bp tandem repeats 51-200bp in length
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRgt6_gt200bp_gt95identity_mergedTandem repeats with >6bp unit size and >200bp in length and >95% identity
segdupSegmental duplications in GRCh37 of all sizes, not including ALTs from GRCh37
segdupwithaltSegmental duplications in GRCh37 >10kb, including ALTs from GRCh37
decoyRegions in GRCh37 to which decoy sequences in hs37d5 map
HG001compoundhetCompound heterozygous regions in which there are 2 different variants when phased variants within 50bp are combined in HG001/NA12878
HG002compoundhetCompound heterozygous regions in which there are 2 different variants when phased variants within 50bp are combined in HG002/NA24385
HG001complexvarComplex variant regions in which there are 2 or more variants within 50bp in HG001/NA12878
HG002complexvarComplex variant regions in which there are 2 or more variants within 50bp in HG002/NA24385

These dimensions help slice and dice the results according to all the different ways in which someone may want to look at them. The special value of '*' (for Subtype, Genotype, Subset) corresponds to summary statistics (i.e. as calculated across the whole genome in the confident regions, without further stratification).

Compilation of results

We ran hap.py (and specifically the HAP-207 version, with the engine set to a GA4GH-specific version of vcfeval) to compare HG001 and HG002 submissions against the GiaB/NIST v3.2.2 HG001 and HG002 truth data respectively. Our executions excluded the entry labeled ccogle-snppet as it did not call variants in the whole genome. The entries labeled ghariani-varprowl, jpowers-varprowl and qzeng-custom contained few VCF lines which were deemed incompatible by this new comparison framework (which performs stricter checks). Upon closer inspection, this was due to incompatible REF columns (such as the strings "nan" or "AC-7GATAGAA") or non-diploid genotypes (such as 0/1/2, or even 0/1/2/3). These were a small fraction, so we decided to exclude these offending lines from this comparison.

We have made available on precisionFDA complete archives of all the input files (including both the original and adjusted files, for the entries which we had to remove offending VCF lines), and the results (including the annotated VCF as output by hap.py, and extended statistics in CSV format). These are recapitulated in a precisionFDA post.

We would like to thank the benchmarking group of the Global Alliance for Genomics and Health for their excellent collaboration throughout this precisionFDA effort.

Results discussion
The compiled results (as described in the previous section) are summarized in the table at the top of this web page. In addition to that, we have created an interactive explorer which can be used to query the full set of all CSV files that were output by the comparison, and to show the metrics for any combination of the "Type", "Subtype", "Genotype" and "Subset" dimensions. You can use that to answer questions such as "how does a particular entry perform across different genome subsets", or "which entry performs best in large deletions?", etc.

EntryTypeSubtypeSubsetGenotypeF-scoreRecallPrecisionFrac_NATruth TP Truth FNQuery TPQuery FPFP gt% FP ma
7001-7050 / 86044 show all
hfeng-pmm2INDELD1_5lowcmp_SimpleRepeat_diTR_11to50het
98.4016
97.3287
99.4985
47.5257
10129278101185147
92.1569
gduggal-bwafbINDEL*lowcmp_SimpleRepeat_diTR_11to50homalt
94.5282
97.7986
91.4694
41.2757
1012922810122944934
98.9407
gduggal-bwaplatSNPtilowcmp_Human_Full_Genome_TRDB_hg19_150331homalt
93.3161
87.6212
99.8027
73.3610
101221430101182017
85.0000
gduggal-bwaplatSNPtilowcmp_Human_Full_Genome_TRDB_hg19_150331_all_mergedhomalt
93.3161
87.6212
99.8027
73.3610
101221430101182017
85.0000
ghariani-varprowlINDELD1_5HG002complexvarhomalt
95.9196
95.4992
96.3438
51.1776
1012147710066382245
64.1361
jpowers-varprowlINDELD1_5HG002complexvarhomalt
96.2485
95.4897
97.0194
51.0621
1012047810058309250
80.9061
jlack-gatkINDELI1_5*hetalt
94.8892
90.3796
99.8725
62.8887
101181077101801312
92.3077
mlin-fermikitINDEL*lowcmp_SimpleRepeat_diTR_11to50homalt
93.0148
97.6827
88.7726
47.1861
101172401009712771255
98.2772
hfeng-pmm3INDELD1_5lowcmp_SimpleRepeat_diTR_11to50het
98.3558
97.1558
99.5859
47.3310
10111296101014237
88.0952
hfeng-pmm1INDELD1_5lowcmp_SimpleRepeat_diTR_11to50het
98.2885
97.1365
99.4681
47.3280
10109298100985438
70.3704
cchapple-customINDELD1_5lowcmp_SimpleRepeat_diTR_11to50het
98.2782
97.1365
99.4471
33.4959
10109298176269890
91.8367
jlack-gatkINDELI1_5HG002compoundhethetalt
94.9037
90.3820
99.9017
57.2515
10102107510162109
90.0000
anovak-vgSNP*lowcmp_Human_Full_Genome_TRDB_hg19_150331_all_gt95identity_merged*
91.2721
93.5191
89.1304
80.4771
10101700105781290640
49.6124
raldana-dualsentieonSNPtilowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.9604
99.9703
99.9505
48.8948
1009931009955
100.0000
eyeh-varpipeSNPtvmap_l125_m1_e0het
96.7521
99.7334
93.9440
75.6347
1009927999064413
2.0186
hfeng-pmm1SNPtilowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.9703
99.9604
99.9802
50.4100
1009841009822
100.0000
anovak-vgINDELD1_5HG002complexvarhomalt
92.9611
95.2727
90.7590
57.9711
10097501102241041840
80.6916
bgallagher-sentieonSNPtilowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.9456
99.9406
99.9505
49.3379
1009661009655
100.0000
dgrover-gatkSNPtilowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.9554
99.9406
99.9703
49.4797
1009661009633
100.0000
hfeng-pmm2SNPtilowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.9505
99.9307
99.9703
51.1679
1009571009532
66.6667
raldana-dualsentieonINDELD1_5lowcmp_SimpleRepeat_diTR_11to50het
97.9616
97.0020
98.9403
47.2355
1009531210084108100
92.5926
jli-customSNPtilowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.9406
99.9307
99.9505
49.4545
1009571009553
60.0000
hfeng-pmm3SNPtilowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.9406
99.9208
99.9604
49.7187
1009481009443
75.0000
mlin-fermikitSNP*map_l125_m2_e0homalt
66.1120
58.0777
76.7260
57.2932
1009172841009130612900
94.7403
jpowers-varprowlSNPtvlowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
97.9694
99.9307
96.0837
66.5958
10090710108412277
67.2330
hfeng-pmm2SNPtvlowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.9455
99.9208
99.9703
62.5348
1008981008933
100.0000
jlack-gatkSNPtilowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.8960
99.8713
99.9208
48.6105
10089131008987
87.5000
ghariani-varprowlSNPtvlowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
97.8554
99.9208
95.8736
65.4859
10089810107435273
62.7586
raldana-dualsentieonSNPtvlowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.9306
99.8911
99.9703
61.3700
10086111008633
100.0000
astatham-gatkSNPtilowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.9009
99.8317
99.9703
49.2989
10085171008533
100.0000
jpowers-varprowlSNPtilowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
97.5573
99.8218
95.3932
57.7564
100841810105488257
52.6639
ckim-dragenSNPtilowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.8713
99.8218
99.9208
48.8376
10084181009388
100.0000
ckim-dragenSNPtvlowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.9158
99.8613
99.9703
61.4918
10083141008533
100.0000
rpoplin-dv42SNPtilowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.8762
99.8119
99.9405
50.3909
10083191008365
83.3333
rpoplin-dv42SNPtvlowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.8811
99.8613
99.9009
61.7921
100831410083102
20.0000
hfeng-pmm3SNPtvlowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.9108
99.8613
99.9603
61.0721
10083141008344
100.0000
ghariani-varprowlSNPtilowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
97.8165
99.8119
95.8994
55.8743
100831910103432262
60.6481
bgallagher-sentieonSNPtvlowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.9009
99.8514
99.9504
61.6260
10082151008255
100.0000
hfeng-pmm1SNPtvlowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.9059
99.8514
99.9603
61.6429
10082151008244
100.0000
jlack-gatkSNPtvlowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.8910
99.8415
99.9405
60.8743
10081161008166
100.0000
dgrover-gatkSNPtvlowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.9009
99.8415
99.9603
61.7282
10081161008144
100.0000
jli-customSNPtvlowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.8860
99.8316
99.9405
61.6312
10080171008064
66.6667
gduggal-snapplatSNP*HG002compoundhethomalt
94.8523
93.4799
96.2656
42.1750
1007970310002388273
70.3608
jpowers-varprowlSNPtvmap_l125_m2_e0het
96.3204
96.5141
96.1274
80.0949
100783641007840694
23.1527
jmaeng-gatkSNPtilowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.8563
99.7624
99.9504
49.5168
10078241007855
100.0000
ciseli-customSNP*HG002compoundhethomalt
82.3846
93.4613
73.6552
42.1506
10077705100373590971
27.0474
cchapple-customSNPtilowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.8364
99.7525
99.9204
39.9139
10077251003686
75.0000
ckim-gatkSNPtilowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.8514
99.7426
99.9603
49.3467
10076261007644
100.0000
qzeng-customINDELD1_5lowcmp_SimpleRepeat_diTR_11to50het
95.7239
96.8194
94.6529
42.8294
1007633116976959707
73.7226
qzeng-customSNPtvlowcmp_AllRepeats_lt51bp_gt95identity_mergedhomalt
99.7317
99.7821
99.6812
62.4851
1007522100073221
65.6250
Interactively explore the results dataset


The interactive explorer includes an additional column, "% FP MA", indicating the percentage of the false positives which have matching non-ref alleles (i.e. FP.gt/QUERY.FP). For SNPs, this fraction varies greatly between pipelines, but for the majority it is less than 30%. Different callers would have the fewest FPs if these matching allele false positives were excluded from the FP counts, making clear that how performance metrics are defined can have a significant effect. For indels, this fraction varies as well, but is generally higher than 50%.

Using the interactive explorer you can also filter for different subsets, such as coding regions. Nonsurprisingly, there are very few FPs and FNs in coding regions for the accurate methods.

You can also see the fraction of calls that fall outside the confidence regions (reported as FRAC_NA). For SNPs, the fraction of calls outside GiaB high-confidence regions is ~10-25%, and for indels the fraction is higher than 50% for all methods. This is an important caveat to consider, since accuracy might be lower outside GiaB high-confidence regions.

F-score, Recall and Precision

Notably, most entries did much better at SNP accuracy measures than at indel accuracy measures. The majority of entries have high SNP f-score, recall, and precision -- above 99%. The respective numbers are lower for indels, and in fact indel recall is usually lower than indel precision. It is important to note, as mentioned in the truth data announcement, that the high-confidence calls have an increased FP, FN, and genotyping error rate for single base indels in homopolymers. This may be a contributing factor to the overall lower indel performance numbers.

Just like in the precisionFDA Consistency Challenge, it appears that pipelines are overall tuned for precision. Compared to the previous challenge, however, the input datasets had higher coverage, which may be a contributing factor to higher recall.

Award determination

We've determined the winners by taking the highest value in each metric (F-score, Recall, Precision) per variant type (SNPs vs indels). Please note that the award for highest f-score has been called "highest performance", to reflect the nature of the score. For each combination of metric per variant type, we are also recognizing entries with high values of that metric, based on a cutoff at the first substantial drop in the percentages. These cutoffs are ultimately subjective, and not directly applicable to other regulatory contexts (which may require well-defined thresholds ahead of time).

These awards and recognitions are meant to encourage the community to participate in challenges and do not constitute an endorsement by the FDA.

Beyond these results

We strongly urge the community to be cautious when interpreting these results. By their nature, high-confidence variant calls and regions tend to include a subset of variants and regions that are easier to characterize. Therefore, such benchmarking against the high-confidence truth data may in fact overestimate accuracy. Manual curation of sequence data in a genome browser for a subset of false positives and false negatives is essential for an accurate understanding of statistics like sensitivity and precision.

When interpreting results within stratification sub-categories, it may be useful to calculate confidence intervals around statistics like sensitivity because there may be very few examples of variants of some types in the benchmark calls in the stratification sub-category.

These results are based on the HG002 truth data. Pipelines may perform differently when confronted with other samples. For a more complete performance assessment of software pipelines, additional samples and experiments would be required.

Heterozygosity rates in chromosome X

We wanted to present an example of other ways in which this dataset can be analyzed, as a way to stimulate the community to conduct further experimentation with the precisionFDA Truth Challenge entries. Since the HG002 sample is male (in contrast to HG001, which is female), we decided to look at the performance of variation calling pipelines in chromosome X. It should be noted that chromosome X is not included in the confident regions of the GiaB truth data for HG002, so the comparisons have not evaluated that chromosome.

In XY male individuals, locations in chromosome X outside of the pseudoautosomal regions are expected to behave as haploid, and variant calling algorithms would ideally report them as haploid calls or as diploid homozygous calls. To measure that, we used bcftools stats -f .,PASS -r X:1-60001,X:2699521-154931044,X:155260561-155270560, which measures the count of heterozygous and homozygous SNPs in chromosome X outside of the PAR. We subsequently calculated the fraction of SNPs that are heterozygous, and generated the following table.

LabelPctHet
dgrover-gatk0.00%
jli-custom0.00%
ltrigg-rtg10.00%
ltrigg-rtg20.00%
mlin-fermikit0.51%
ckim-isaac1.10%
raldana-dualsentieon1.66%
hfeng-pmm11.76%
hfeng-pmm31.97%
hfeng-pmm22.04%
ndellapenna-hhga2.21%
egarrison-hhga2.30%
bgallagher-sentieon2.34%
astatham-gatk2.36%
rpoplin-dv422.42%
ckim-dragen4.17%
cchapple-custom5.11%
ckim-gatk5.23%
ckim-vqsr5.23%
asubramanian-gatk6.00%
ciseli-custom6.10%
qzeng-custom6.23%
jlack-gatk6.32%
gduggal-bwafb7.89%
jpowers-varprowl8.52%
eyeh-varpipe10.62%
ghariani-varprowl10.94%
gduggal-snapfb11.32%
gduggal-bwavard13.09%
gduggal-bwaplat16.11%
gduggal-snapvard16.21%
gduggal-snapplat17.29%
anovak-vg31.01%

Certain algorithms (such as the ones used in the ltrigg-rtg1 and ltrigg-rtg2 entries) can be gender-aware, and will not generate heterozygous calls when configured to run with a male option. Four entries in total had a zero heterozygosity fraction.

Final remarks

We want to thank those of you who participated in this challenge! As with our first challenge, by participating and putting your results and your thoughts out in the public, you fulfilled the first and most important goal of this challenge – to engage and start sharing data.

This challenge created a rich and interesting dataset that we hope people will study further. This dataset includes not only the submitted VCF files but also the complete set of comparison results (annotated VCFs and calculated statistics). It represents the essence of this challenge, and can hopefully be useful to inform future approaches, not only in terms of regulatory science but also for reference materials and benchmarking methodologies. We hope the GiaB group will be able to make use of it to improve reference calls (particularly around indels), and the GA4GH benchmarking group will be able to use it to improve the comparison methodology. We are excited to see how else the community will use this data set in the future.

Quotes

In closing, we would like to leave you with several encouraging quotes from our collaborators and larger community, which will hopefully inspire all readers to actively engage with the precisionFDA community, or participate in one of our upcoming challenges:


“We hope that in the near future, thanks to more well-characterized truth sets, the genomics community can determine whether these more-refined pipelines are indeed beneficial for enabling Precision Data for Precision Medicine.”
“Thanks for your patience in being a guinea pig for our high confidence calls and benchmarking tools! This has been really useful for us already!”
“These challenges are potentially of as much value to those constructing truth sets as they are to the rest of the community, as they provide a means to look at a wide variety of call sets with respect to multiple truth sets and start to delve into potential biases in various truth sets and the methods used to construct them. This is an evolving process and it is great to see how GIAB and GA4GH are also moving things forward in this regard.”
“We applaud the effort by NIST and FDA to develop more truth sets that the accuracy of all algorithms can be benchmarked against. We look forward to more comprehensive truth sets and open challenges that can drive the development of better pipeline algorithms and enable precision data for precision medicine.”
“Thanks for all your help! Hopefully through this process we all learned and improved across multiple fronts, so it's already a huge success. I’m already thinking of ways some of these results could be used in publications about benchmarking.”