PrecisionFDA
Truth Challenge

Engage and improve DNA test results with our community challenges

The precisionFDA Truth Challenge ran from April 26, 2016 to May 26, 2016, giving participants the unique and exciting opportunity to test their NGS pipelines on an uncharacterized sample (HG002) and publish results on precisionFDA for subsequent evaluation against newly-revealed "truth" data.
This web page discusses the results and lists the awards and recognitions handed out by the precisionFDA team. Due to novelties related to both the truth data and the comparison methodology, these results are only a first cut at conducting an evaluation. In the sections below, we present some lessons learned during the process of setting up the challenge, determining the truth, and using a new comparison methodology; and we finally invite the community to further explore these results and provide insight for the future.
Community Challenge Awards
Highest
SNP Performance
in the precisionFDA Truth Challenge
Awarded to
Verily Life Sciences
  • Ryan Poplin
  • Mark DePristo
  • Verily Life Sciences Team
Highest
SNP Recall
in the precisionFDA Truth Challenge
Awarded to
Sentieon
  • Rafael Aldana
  • Hanying Feng
  • Brendan Gallagher
  • Jun Ye
Highest
SNP Precision
in the precisionFDA Truth Challenge
Awarded to
Kinghorn Center
for Clinical Genomics
  • Aaron Statham
  • Mark Cowley
  • Joseph Copty
  • Mark Pinese
Highest
INDEL Performance
in the precisionFDA Truth Challenge
Awarded to
Sanofi-Genzyme
  • Deepak Grover
Highest
INDEL Recall
in the precisionFDA Truth Challenge
Awarded to
Sanofi-Genzyme
  • Deepak Grover
Highest
INDEL Precision
in the precisionFDA Truth Challenge
Awarded to
Sentieon
  • Rafael Aldana
  • Hanying Feng
  • Brendan Gallagher
  • Jun Ye
Must-read Introductory Remarks

We are very excited to see growing numbers of challenge participants! For this challenge we received 35 entries, not only from many who participated in the previous challenge, but also from several first time participants. Some participants showcased their developing new methods, while others decided to use existing methods and see how they perform in this particular environment. We are extremely grateful to all participants for their willingness to share their results and to contribute to this effort.

This second precisionFDA challenge was conducted in collaboration with the Genome in a Bottle (GiaB) consortium, which provided the "truth" data, as well as with the Global Alliance for Genomics and Health (GA4GH), which provided best practices and software for conducting the comparison of participants' entries to the truth data. In fact, this effort represents the first time that this new truth data has been investigated (and it's only an approximation of the truth, hence sometimes we say "truth" instead of truth), and the first time that this new comparison methodology has been applied at scale across the vast number of submitted entries.

Our evaluation against the HG002 truth data has produced several metrics (such as f-score, recall and precision) across different variant types (such as SNP and indels), subtypes (such as insertions or deletions of specific size ranges) and genomic contexts (such as whole genome, coding regions, etc.), leading to thousands of computed numbers per challenge entry. We've chosen six of them (f-score, recall and precision, across SNPs and across indels, in the whole genome) to use as the basis for handing out awards and recognitions.

These results are by no means the final word. Given the originality of the truth data and the comparison methodology, we expect the community to further conduct analyses and contribute to improvements in the benchmarking methodology, the correctness of the truth data, and the definition of comparison metrics.

We would like to acknowledge and thank all of those who participated in the precisionFDA Truth Challenge. As with the previous challenge, we hope that everyone will feel like a winner.

Overview of results

The following table summarizes the challenge entries, and the results of the comparison against the HG002 truth data.

We've given each entry a unique label, comprised of the name of the submitting user as well as a short mnemonic keyword representing the pipeline. (As with the previous challenge, these keywords are merely indicative of each pipeline's main component, hence somewhat subjective; for a more faithful description of each pipeline, refer to the full text that accompanied each submission by following the label links). Each submitted entry consisted of two VCFs, corresponding to the variants called on the HG001/NA12878 and HG002/NA24385 datasets respectively. Links to these files can be found in the "Datasets" tab of the table.

The entries are sorted alphabetically. You can click on any column header to re-sort the table according to that column.

Label Submitter Organization SNP-Fscore SNP-recall SNP-precision INDEL-Fscore INDEL-recall INDEL-precision
raldana-dualsentieon Rafael Aldana   et al. Sentieon 99.9260 99.9131 99.9389 99.1095 98.7566 99.4648
bgallagher-sentieon Brendan Gallagher   et al. Sentieon 99.9296 99.9673 99.8919 99.2678 99.2143 99.3213
mlin-fermikit Mike Lin DNAnexus Science 98.8629 98.2311 99.5029 95.5997 94.8918 96.3183
jmaeng-gatk Ju Heon Maeng Yonsei University 99.6144 99.4608 99.7686 99.1098 99.0216 99.1981
ckim-dragen Changhoon Kim Macrogen 99.8268 99.9524 99.7015 99.1359 99.1574 99.1143
ckim-gatk Changhoon Kim Macrogen 99.6466 99.4788 99.8150 99.2271 99.1551 99.2992
ckim-isaac Changhoon Kim Macrogen 98.5357 97.1616 99.9494 95.8099 93.7006 98.0163
ckim-vqsr Changhoon Kim Macrogen 99.2866 98.6511 99.9303 99.2541 99.0614 99.4476
ltrigg-rtg2 Len Trigg RTG 99.8749 99.8935 99.8562 99.2539 98.8759 99.6347
ltrigg-rtg1 Len Trigg RTG 99.8754 99.8921 99.8587 99.0160 98.3355 99.7061
hfeng-pmm1 Hanying Feng   et al. Sentieon 99.9496 99.9227 99.9766 99.3397 99.0289 99.6526
hfeng-pmm2 Hanying Feng   et al. Sentieon 99.9416 99.9254 99.9579 99.3119 99.0152 99.6103
hfeng-pmm3 Hanying Feng   et al. Sentieon 99.9548 99.9339 99.9756 99.3628 99.0161 99.7120
jlack-gatk Justin Lack NIH 99.7200 99.9393 99.5016 98.6899 98.8138 98.5664
astatham-gatk Aaron Statham   et al. KCCG 99.5934 99.2091 99.9807 99.3424 99.2404 99.4446
qzeng-custom Qian Zeng LabCorp 99.4966 99.2413 99.7533 96.8316 96.8703 96.7929
anovak-vg Adam Novak   et al. vgteam 98.4545 98.3357 98.5736 70.4960 69.7491 71.2591
eyeh-varpipe ErhChan Yeh   et al. Academia Sinica 99.4670 99.9638 98.9751 92.5779 91.3854 93.8021
ciseli-custom Christian Iseli   et al. SIB 97.7648 98.8356 96.7169 83.5453 82.5314 84.5844
ccogle-snppet* Christopher Cogle   et al. CancerPOP
asubramanian-gatk Ayshwarya Subramanian   et al. Broad Institute 98.9379 97.9985 99.8954 98.8418 98.5404 99.1451
rpoplin-dv42 Ryan Poplin   et al. Verily Life Sciences 99.9587 99.9447 99.9728 98.9802 98.7882 99.1728
cchapple-custom Charles Chapple   et al. Saphetor 99.8448 99.8832 99.8063 99.1388 98.8448 99.4346
gduggal-bwafb Geet Duggal   et al. DNAnexus Science 99.7820 99.8619 99.7021 96.9474 95.5004 98.4390
gduggal-bwaplat Geet Duggal   et al. DNAnexus Science 98.8646 98.0471 99.6958 92.6621 87.0843 99.0034
gduggal-bwavard Geet Duggal   et al. DNAnexus Science 99.3249 99.0431 99.6083 87.3464 87.1769 87.5166
gduggal-snapfb Geet Duggal   et al. DNAnexus Science 99.2501 99.8026 98.7037 92.2602 90.5733 94.0112
gduggal-snapplat Geet Duggal   et al. DNAnexus Science 99.0030 98.6815 99.3266 76.4210 69.0418 85.5664
gduggal-snapvard Geet Duggal   et al. DNAnexus Science 99.0871 98.9341 99.2406 83.0264 83.4429 82.6139
ghariani-varprowl Gunjan Hariani   et al. Quintiles 99.3496 99.8685 98.8361 87.2025 87.3272 87.0781
jpowers-varprowl Jason Powers   et al. Q2 Solutions 99.5004 99.5447 99.4561 86.4885 85.2886 87.7226
dgrover-gatk Deepak Grover Sanofi-Genzyme 99.9456 99.9631 99.9282 99.4009 99.3458 99.4561
egarrison-hhga Erik Garrison   et al. - 99.8985 99.8365 99.9607 97.4253 97.1646 97.6874
jli-custom Jian Li   et al. Roche 99.9382 99.9603 99.9160 99.3675 99.0788 99.6580
ndellapenna-hhga Nicolas Della Penna ANU 99.8818 99.8118 99.9519 97.3838 97.0938 97.6756
Label HG001 VCF HG002 VCF
raldana-dualsentieon Challenge201605_Sentieon_dualmap_HG001_final.vc... Challenge201605_Sentieon_dualmap_HG002_final.vc...
bgallagher-sentieon Sentieon DNAseq Gallagher HG001.vcf Sentieon DNAseq Gallagher HG002.vcf
mlin-fermikit HG001.fermikit.raw.vcf.gz HG002.fermikit.raw.vcf.gz
jmaeng-gatk HG001-NA12878-50x_illumina_WGS_TruthChallenge.R... HG002-NA24385-50x_illumina_WGS_TruthChallenge.R...
ckim-dragen DRAGEN_HG001-NA12878-50x.hard-filtered.vcf.gz DRAGEN_HG002-NA24385-50x.vcf.gz
ckim-gatk GATK_HG001-NA12878.Filtered.Variants.vcf.gz GATK_HG002-NA24385.Filtered.Variants.vcf.gz
ckim-isaac ISAAC_HG001-NA12878_all_passed_variants.vcf.gz ISAAC_HG002-NA24385_all_passed_variants.vcf.gz
ckim-vqsr VQSR_HG001-NA12878.recalibrated_variants.vcf.gz VQSR_HG002-NA24385.recalibrated_variants.vcf.gz
ltrigg-rtg2 RTG-HG001-NA12878-50x-AVR.vcf.gz RTG-HG002-NA24385-50x-AVR.vcf.gz
ltrigg-rtg1 RTG-HG001-NA12878-50x-GQ20.vcf.gz RTG-HG002-NA24385-50x-GQ20.vcf.gz
hfeng-pmm1 Sentieon-HG001-PMM1.vcf.gz Sentieon-HG002-PMM1.vcf.gz
hfeng-pmm2 Sentieon-HG001-PMM2.vcf.gz Sentieon-HG002-PMM2.vcf.gz
hfeng-pmm3 Sentieon-HG001-PMM3.vcf.gz Sentieon-HG002-PMM3.vcf.gz
jlack-gatk NA12878_CCBR.vcf.gz NA24385_CCBR.vcf.gz
astatham-gatk AStatham-Garvan-HG001.hc.vqsr.vcf.gz AStatham-Garvan-HG002.hc.vqsr.vcf.gz
qzeng-custom LabCorp_HG001.vcf.gz LabCorp_HG002.vcf.gz
anovak-vg vg_may.HG001.vcf.gz vg_may.HG002.vcf.gz
eyeh-varpipe HG001-NA12878_varpipe_1.vcf HG002-NA24385_varpipe_1.vcf
ciseli-custom SIB-HG001.vcf.gz SIB-HG002.vcf.gz
ccogle-snppet HG001_NA12878_CancerPOP.vcf HG002_NA24385_CancerPOP.vcf
asubramanian-gatk pfda-truth-bi.hg001-rerun2.vcf.gz pfda-truth-bi.hg002-rerun2.vcf.gz
rpoplin-dv42 HG001-NA12878-pFDA.vcf.gz HG002-NA24385-pFDA.vcf.gz
cchapple-custom Saphetor-hg001.vcf.gz Saphetor-hg002.vcf.gz
gduggal-bwafb NA12878-bwa-HG001-freebayes-keeplcr.vcf.gz NA24385-bwa-HG002-freebayes-keeplcr.vcf.gz
gduggal-bwaplat NA12878-bwa-HG001-platypus-keeplcr.vcf.gz NA24385-bwa-HG002-platypus-keeplcr.vcf.gz
gduggal-bwavard NA12878-bwa-HG001-vardict-keeplcr.vcf.gz NA24385-bwa-HG002-vardict-keeplcr.vcf.gz
gduggal-snapfb NA12878-snap-HG001-freebayes-keeplcr.vcf.gz NA24385-snap-HG002-freebayes-keeplcr.vcf.gz
gduggal-snapplat NA12878-snap-HG001-platypus-keeplcr.vcf.gz NA24385-snap-HG002-platypus-keeplcr.vcf.gz
gduggal-snapvard NA12878-snap-HG001-vardict-keeplcr.vcf.gz NA24385-snap-HG002-vardict-keeplcr.vcf.gz
ghariani-varprowl default_with_model_NA12878.vcf.gz default_with_model_NA24385.vcf.gz
jpowers-varprowl HG001-NA12878-50x.vcf.gz HG002-NA24385-50x.vcf.gz
dgrover-gatk HG001-NA12878-dgrover.vcf.gz HG002-NA24385-dgrover.vcf.gz
egarrison-hhga HG001d.wg.hhga_yeoq.model.98.vcf.gz HG002d.wg.hhga_yeoq.model.98.vcf.gz
jli-custom pfdaTC_Bina_HG001.vcf.gz pfdaTC_Bina_HG002.vcf.gz
ndellapenna-hhga HG001d.wg.hhga_refined_prague.model.18.vcf.gz HG002d.wg.refined_prague.model.18.vcf.gz
*ccogle-snppet: This entry was not considered because the entry did not call variants across the whole genome

For more information about how these results were calculated, consult the comparison section below.

We are handing out the following community challenge awards:

  • highest-snp-performance to the entry submitted by Ryan Poplin et al. from Verily Life Sciences, for achieving the highest SNP F-score.
  • highest-snp-recall to the entry submitted by Brendan Gallagher et al. from Sentieon (labeled bgallagher-sentieon), for achieving the highest SNP recall.
  • highest-snp-precision to the entry submitted by Aaron Statham et al. from Kinghorn Center for Clinical Genomics, for achieving the highest SNP precision.
  • highest-indel-performance to the entry submitted by Deepak Grover from Sanofi-Genzyme, for achieving the highest indel F-score.
  • highest-indel-recall to the entry submitted by Deepak Grover from Sanofi-Genzyme, for achieving the highest indel recall.
  • highest-indel-precision to the entry submitted by Hanying Feng et al. from Sentieon (labeled hfeng-pmm3), for achieving the highest indel precision.

We are also handing out the following recognitions:

  • high-snp-performance to all entries achieving a SNP F-score of 99.920% or higher.
  • high-snp-recall to all entries achieving a SNP recall of 99.910% or higher.
  • high-snp-precision to all entries achieving a SNP precision of 99.920% or higher.
  • high-indel-performance to all entries achieving an indel F-score of 99.310% or higher.
  • high-indel-recall to all entries achieving an indel recall of 99.150% or higher.
  • high-indel-precision to all entries achieving an indel precision of 99.430% or higher.
Truth Data

One way of learning about the performance of software pipelines for processing human genome sequencing data is by using well-characterized datasets such as NA12878/HG001. The previous challenge made use of the GiaB/NIST NA12878 v2.19 truth data as part of the evaluation of accuracy. But considering how widely HG001 has been used to train software pipelines, this challenge focuses on a new sample, HG002, whose truth data was not yet available at the time of the challenge.

In particular, this challenge provided sequencing reads datasets (FASTQ) for both HG001 and HG002. During the challenge, participants had access to the aforementioned HG001 v2.19 truth data, but not to the HG002 truth data. Upon the closing of the Truth Challenge, the Genome in a Bottle (GiaB) consortium, led by the National Institute of Standards (NIST), released the HG002 truth data, as well as an updated version of the HG001 truth data. Both were initially assigned version 3.2.

The truth data files have been published on precisionFDA by Dr. Justin Zook (scientist at the National Institute of Standards and Technology and co-leader of the Genome in a Bottle Consortium efforts), in the precisionFDA discussion topic titled Help improve the new v3.2 GIAB benchmark calls for HG001 and HG002.

For your convenience, we've included below a copy of the announcement. It discusses known issues of the truth data as well as other important caveats, and we encourage you to read it to understand how this new truth data may have affected the results of the challenge, or may interact with your own evaluations in the future.

NEW RESULTS

Based on feedback received about our v3.2 calls, we have now uploaded a new version 3.2.2 set of calls that fixes a significant number of erroneously low confidence variants on chr7 and chr9, and also now excludes regions homologous to the hs37d5 decoy sequence.

Changes in v3.2.2: We found a bug in one of the input call sets that was causing many of the variants on chr7 in HG001 and HG002 and on chr9 on HG002 to be reported as not high confidence. We have corrected this issue, and all other chromosomes are the same as in v3.2.1.

Changes in v3.2.1: The only change in v3.2.1 is to exclude regions from our high confidence bed file if they have homology to the decoy sequence hs37d5. These decoy-related regions were generated by Heng Li and are available as mm-2-merged.bed at https://github.com/ga4gh/benchmarking-tools/tree/master/resources/stratification-bed-files/SegmentalDuplications. Our high-confidence vcf generally does not include variants if they are homologous to decoy sequence. Although most of these likely should be excluded, some real variants may be missed, so we are excluding these regions from our high-confidence regions until they are better characterized. However, it is important to note that these regions may be enriched for false positives if the decoy is not used in mapping.

Description of v3.2: We have been developing a new, reproducible integration process to create high-confidence SNP, small indel, and homozygous reference calls for benchmarking small variant calls. We have used this new process to generate v3.2 of our high-confidence calls for HG001 (aka NA12878), as well as the first high-confidence callset for HG002 (aka Ashkenazim Jewish Son), which are attached to this discussion and under ftp://ftp-trace.ncbi.nlm.nih.gov/giab/ftp/release/. These calls were generated only from data generated from our NIST Reference Material batch of cells. Planned future work will include (1) generating similar callsets on the GIAB/PGP AJ parents and Chinese trio, (2) incorporating additional datasets from these genomes to characterize more difficult regions, (3) incorporating pedigree information for phasing and to improve the calls, (4) developing better methods for homopolymer characterization, (5) making calls on X and Y for males, and (6) characterizing larger indels and structural variation. We greatly appreciate the feedback many of you have given already, and we very much welcome additional feedback to help us improve future versions of these calls as comments in this discussion, and about specific challenging variants at http://goo.gl/forms/OCUnvDXMEt1NEX8m2.

We highly recommend reading the information below prior to using these calls to understand how best to use them and their limitations. There is also a draft google doc that describes the methods used to form these calls and some preliminary comparisons to other callsets at: https://docs.google.com/document/d/1NE-NGSMslFFndQpbqnvu4wfnwgmgOcQu6KIjSbCauYg/edit?usp=sharing

Best Practices for Using High-confidence Calls:

Benchmarking variant calls is a complex process, and best practices are still being developed by the Global Alliance for Genomics and Health (GA4GH) Benchmarking Team (https://github.com/ga4gh/benchmarking-tools/). Several things are important to consider when benchmarking variant call accuracy:

  1. Complex variants (e.g., nearby SNPs and indels or block substitutions) can often be represented correctly in multiple ways in the vcf format. Therefore, we recommend using sophisticated benchmarking tools like those developed by members of the GA4GH benchmarking team. Currently, the most developed tools are rtgtools vcfeval (http://realtimegenomics.com/products/rtg-tools/) or hap.py (https://github.com/Illumina/hap.py). See supplementary information in https://docs.google.com/document/d/1NE-NGSMslFFndQpbqnvu4wfnwgmgOcQu6KIjSbCauYg/edit?usp=sharing for a suggested process for doing comparisons.
  2. By their nature, high-confidence variant calls and regions tend to include a subset of variants and regions that are easier to characterize. Accuracy of variant calls outside the high-confidence regions is generally likely to be lower than inside the high-confidence regions, so benchmarking against high-confidence calls will usually overestimate accuracy for all variants in the genome. Similarly, it is possible that a variant calling method has higher accuracy statistics compared to other methods when compared to the high-confidence variant calls but has lower accuracy compared to other methods for all variants in the genome.
  3. Stratification of performance for different variant types and different genome contexts can be very useful when assessing performance, because performance often differs between variant types and genome contexts. In addition, stratification can elucidate variant types and genome contexts that fall primarily outside high-confidence regions. Standardized bed files for stratifying by genome context are available from GA4GH at https://github.com/ga4gh/benchmarking-tools/tree/master/resources/stratification-bed-files.
  4. Particularly for targeted sequencing, it is critical to calculate confidence intervals around statistics like sensitivity because there may be very few examples of variants of some types in the benchmark calls in the targeted regions. Manual curation of sequence data in a genome browser for a subset of false positives and false negatives is essential for an accurate understanding of statistics like sensitivity and precision. Curation can often help elucidate whether the benchmark callset is wrong, the test callset is wrong, both callsets are wrong, or the true answer is unclear from current technologies. If you find questionable or challenging sites, reporting them will help improve future callsets. You can report information about particular sites at http://goo.gl/forms/OCUnvDXMEt1NEX8m2.

Known issues with high-confidence calls:

There are a number of known issues in the current high-confidence callset, which we plan to address in future releases:

  1. Compound heterozygous sites are sometimes represented as a heterozygous deletion that overlaps with a heterozygous SNP. At these sites, the genotype “0/1” or “0|1” should not be interpreted as meaning that the “0” haplotype is reference. Hap.py and vcfeval (with the --ref-overlap flag) interpret these sites as intended.
  2. The callset currently only includes local phasing information from freebayes, which does not use the PS field, so that locally phased variants should not be assumed to be phased with variants that are phased at a distant location.
  3. The high-confidence calls have an increased FP, FN, and genotyping error rate for single base indels in homopolymers. Some of these sites have a low allele fraction variant in multiple technologies so that it is unclear whether the site is a systematic sequencing error or a true variant in a fraction of the cells. Some of these sites are listed in https://docs.google.com/spreadsheets/d/1kHgRLinYcnxX3-ulvijf2HrIdrQWz5R5PtxZS-_s6ZM/edit?usp=sharing

Differences between v2.19 and v3.2 integration methods:

The new v3.2 integration methods differ from the previous GIAB calls (v2.18 and v2.19) in several ways, both in the data used and the integration process and heuristics:

  1. Only 4 datasets were used, selected because they represent 4 sequencing technologies that were generated from the NIST RM 8398 batch of DNA. They are: 300x Illumina paired end WGS, 100x Complete Genomics WGS, 1000x Ion exome sequencing, and SOLID WGS.
  2. Mapping and variant calling algorithms designed specifically for each technology were used to generate sensitive variant callsets where possible: novoalign + GATK-haplotypecaller and freebayes for Illumina, the vcfBeta file from the standard Complete Genomics pipeline, tmap+TVC for Ion exome, and Lifescope+GATK-HC for SOLID. This is intended to minimize bias towards any particular bioinformatics toolchain.
  3. Instead of forcing GATK to call genotypes at candidate variants in the bam files from each technology, we generate sensitive variant call sets and a bed file that describes the regions that were callable from each dataset. For Illumina, we used GATK callable loci to find regions with at least 20 reads with MQ >= 20 and with coverage less than 2x the median. For Complete Genomics, we used the callable regions defined by the vcfBeta file and excluded +-50bp around any no-called or half-called variant. For Ion, we intersected the exome targeted regions with callableloci (requiring at least 20 reads with MQ >= 20). Due to the shorter reads and low coverage for SOLID, it was only used to confirm variants, so no regions were considered callable.
  4. A new file with putative structural variants was used to exclude potential errors around SVs. These were SVs derived from multiple PacBio callers and multiple integrated illumina callers using MetaSV. These make up a significantly smaller fraction of the calls and genome (~4.5%) than the previous bed, which was a union of all dbVar calls for NA12878 (~10%).
  5. Homopolymers >10bp in length, including those interrupted by one nucleotide different from the homopolymer, are now excluded from the high-confidence regions, because these were seen to have a high error rate for all technologies. This constitutes a high fraction of the putative indel calls, so more work is needed to resolve these.
  6. A bug that caused nearby variants to be missed in v2.19 is fixed in the new calls.
  7. The new vcf contains variants outside the high-confidence bed file. This enables more robust comparison of complex variants or nearby variants that are near the boundary of the bed file. It also allows the user to evaluate concordance outside the high-confidence regions, but these concordance metrics should be interpreted with great care.
  8. We now output local phasing information when it is provided in the Illumina freebayes calls.

Shortly after the initial publication of v3.2 of the truth data, there were two subsequent updates leading to v3.2.1 and v3.2.2. These updates introduced improvements to the truth data based on feedback from the community and from artifacts discovered in initial evaluations of the challenge entries. The final comparison results discussed on this web page are based on comparisons against the latest version of the truth data (v3.2.2).

It is important to note that the methodology for compiling the truth data has changed substantially since the release of v2.19 of the HG001 truth dataset, and therefore the same VCF file may rank differently in a comparison against HG001 v2.19 versus a comparison against HG001 v3.2.2. These changes are discussed in the section titled "Differences between v2.19 and v3.2 integration methods" in Dr. Zook's announcement.

We would like to thank the National Institute of Standards and Technology, and the Genome in a Bottle consortium, for the excellent collaboration in this precisionFDA challenge. As can be attested by them, creating a truth dataset is a complicated process. The published dataset, even with its latest v3.2.2 improvements, is only an approximation of the truth. We hope you will keep that in mind when interpreting the results of this challenge and that you will exercise caution, as the determination of truth is always an ongoing process.

Comparison methodology

The precisionFDA system includes an initial VCF comparison framework which was used in the first precisionFDA challenge. We asked participants to use that framework to compare their HG001 VCFs against the GiaB/NIST HG001 v2.19 truth data while the challenge was still ongoing, to ensure that their files were in good shape. However, for the final evaluation of the results of this challenge, we partnered with the Benchmarking Task team of the GA4GH Data Working Group, to employ an updated comparison framework.

The GA4GH benchmarking team has been discussing the difficulties of conducting proper VCF comparisons for several months now, and has devised metrics definitions, initial specifications, early software prototypes, and other benchmarking resources for implementing VCF comparison best practices. We approached the chair of the benchmarking group, Dr. Justin Zook, to consult with him on how to best use these resources in the context of this precisionFDA challenge.

The outcome of our collaboration is the utilization of a particular methodology for comparing VCFs and counting results, which was finalized and tailored for the needs of this challenge, and which we hope will form the basis of other comparisons in the future. It is available on precisionFDA as an app, titled Vcfeval + Hap.py Comparison. (Given the departure from the previous comparison technique, and until we have received enough feedback for this new method, the "Comparison" feature of precisionFDA has not yet been updated.)

Software used

This particular comparison framework consists of a specific version of Real Time Genomics' vcfeval, used for VCF comparison, and a specific version of Illumina's hap.py, used for quantification; together they form the prototype GA4GH benchmarking workflow, as illustrated here.

The vcfeval tool generates an intermediate VCF which is further quantified by hap.py. The "quantify" tool from hap.py counts and stratifies variants. It counts SNPs and INDELs separately and provides additional counts for subtypes of each variant (indels of various lengths, het / hom, stratification regions). In order to determine if a variant specifies SNPs, insertions or deletions, "quantify" performs an alignment of the corresponding REF and ALT alleles. This improves handling of MNPs/complex variants and provides results that are comparable between different methods.

Metrics calculated

The quantification process calculates the following metrics, in harmony with "Comparison Method 3" as defined in this upcoming GA4GH benchmarking spec update:

MetricDefinition
TRUTH.TPTrue positives, from the perspective of the truth data, i.e. the number of sites in the Truth Call Set for which there are paths through the Query Call Set that are consistent with all of the alleles at this site, and for which there is an accurate genotype call for the event.
QUERY.TPTrue positives, from the perspective of the query data, i.e. the number of sites in the Query Call Set for which there are paths through the Truth Call Set that are consistent with all of the alleles at this site, and for which there is an accurate genotype call for the event.
TRUTH.FNFalse negatives, i.e. the number of sites in the Truth Call Set for which there is no path through the Query Call Set that is consistent with all of the alleles at this site, or sites for which there is an inaccurate genotype call for the event. Sites with correct variant but incorrect genotype are counted here.
QUERY.FPFalse positives, i.e. the number of sites in the Query Call Set for which there is no path through the Truth Call Set that is consistent with this site. Sites with correct variant but incorrect genotype are counted here.
FP.gtThe number of false positives where the non-REF alleles in the Truth and Query Call Sets match (i.e. cases where the truth is 1/1 and the query is 0/1 or similar).
RecallTRUTH.TP / (TRUTH.TP + TRUTH.FN)
PrecisionQUERY.TP / (QUERY.TP + QUERY.FP)
F-scoreHarmonic mean of Recall and Precision

Note that recall uses TRUTH.TP whereas precision uses QUERY.TP. For recall, TRUTH.TP counts the number of truth variants reproduced in the truth set representation, which is the same for all callers and should also be consistent with the confident regions (especially important around confident region boundaries and where variants can move between categories depending on how a variant caller decides to represent them). Using QUERY.TP for recall would have introduced problems where different methods create different representations (e.g. systematically calling MNPs as insertions + deletions can skew indel numbers in the query). For calculating precision, the tool uses QUERY.TP (i.e. the precision measures the relative number of wrong calls for all query calls).

Granular reporting and region stratification

The quantification process reports results with additional granularity, according to a few different dimensions, as outlined in the four tabs of the following table:

ValueMeaning
SNPSNP or MNP variants
INDELIndels and complex variants
ValueMeaning
*Aggregate numbers of all subtypes
tiSNPs that constitute transitions
tvSNPs that constitute transversions
I1_5Insertions of length 1-5
I6_15Insertions of length 6-15
I16_PLUSInsertions of length 16 or more
D1_5Deletions of length 1-5
D6_15Deletions of length 6-15
D16_PLUSDeletions of length 16 or more
C1_5Complex variants of length 1-5
C6_15Complex variants of length 6-15
C16_PLUSComplex variants of length 16 or more
ValueMeaning
*Aggregate numbers of all genotypes
hetOnly heterozygous variant calls (0/1 or similar genotypes)
homaltOnly homozygous alternative variant calls (1/1 or similar genotypes)
hetOnly heterozygous alternative variant calls (1/2 or similar genotypes)
ValueMeaning
*Aggregate numbers not limited to any subset (but still within confident regions)
func_cdsCoding exons from RefSeq
map_l100_m2_e1Regions in which 100bp reads map to >1 location with up to 2 mismatches and up to 1 indel
map_l150_m0_e0Regions in which 150bp reads map to >1 location with up to 0 mismatches and up to 0 indels
map_l150_m2_e1Regions in which 150bp reads map to >1 location with up to 2 mismatches and up to 1 indel
map_l150_m2_e0Regions in which 150bp reads map to >1 location with up to 2 mismatches and up to 0 indels
map_l100_m1_e0Regions in which 100bp reads map to >1 location with up to 1 mismatch and up to 0 indels
map_l125_m1_e0Regions in which 125bp reads map to >1 location with up to 1 mismatch and up to 0 indels
map_l250_m2_e1Regions in which 250bp reads map to >1 location with up to 2 mismatches and up to 1 indel
map_l250_m0_e0Regions in which 250bp reads map to >1 location with up to 0 mismatches and up to 0 indels
map_l150_m1_e0Regions in which 150bp reads map to >1 location with up to 1 mismatch and up to 0 indels
map_l125_m2_e0Regions in which 125bp reads map to >1 location with up to 2 mismatches and up to 0 indels
map_l100_m0_e0Regions in which 100bp reads map to >1 location with up to 0 mismatches and up to 0 indels
map_l250_m1_e0Regions in which 250bp reads map to >1 location with up to 1 mismatch and up to 0 indels
map_l125_m2_e1Regions in which 125bp reads map to >1 location with up to 2 mismatches and up to 1 indel
map_l250_m2_e0Regions in which 250bp reads map to >1 location with up to 2 mismatches and up to 0 indels
map_l125_m0_e0Regions in which 125bp reads map to >1 location with up to 0 mismatches and up to 0 indels
map_l100_m2_e0Regions in which 100bp reads map to >1 location with up to 2 mismatches and up to 0 indels
map_sirenRegions considered difficult to map by amplab SiRen
tech_badpromoters1000 promoter regions with lowest relative coverage in Illumina, relatively GC-rich
lowcmp_SimpleRepeat_homopolymer_gt10Homopolymers >10bp in length
lowcmp_SimpleRepeat_triTR_11to50Exact 3bp tandem repeats 11-50bp in length
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRgt6_lt51bp_gt95identity_mergedTandem repeats with >6bp unit size and <51bp in length and >95% identity
lowcmp_Human_Full_Genome_TRDB_hg19_150331_all_mergedAll tandem repeats from TRDB with adjacent repeats merged
lowcmp_SimpleRepeat_quadTR_11to50Exact 4bp tandem repeats 11-50bp in length
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRlt7_gt200bp_gt95identity_mergedTandem repeats with 1-6bp unit size and >200bp in length and >95% identity
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRlt7_lt101bp_gt95identity_mergedTandem repeats with 1-6bp unit size and <101bp in length and >95% identity
lowcmp_SimpleRepeat_quadTR_gt200Exact 4bp tandem repeats >200bp in length
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRgt6_lt101bp_gt95identity_mergedTandem repeats with >6bp unit size and <101bp in length and >95% identity
lowcmp_AllRepeats_gt200bp_gt95identity_mergedAll perfect and imperfect tandem repeats >200bp in length
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRgt6_51to200bp_gt95identity_mergedTandem repeats with >6bp unit size and 51-200bp in length and >95% identity
lowcmp_Human_Full_Genome_TRDB_hg19_150331_all_gt95identity_mergedAll tandem repeats from TRDB with >95% identity
lowcmp_SimpleRepeat_triTR_gt200Exact 3bp tandem repeats >200bp in length
lowcmp_SimpleRepeat_triTR_51to200Exact 3bp tandem repeats 51-200bp in length
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRlt7_lt51bp_gt95identity_mergedTandem repeats with 1-6bp unit size and <51bp in length and >95% identity
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRlt7_51to200bp_gt95identity_mergedTandem repeats with 1-6bp unit size and 51-200bp in length and >95% identity
lowcmp_AllRepeats_51to200bp_gt95identity_mergedAll perfect and imperfect tandem repeats 51-200bp in length
lowcmp_Human_Full_Genome_TRDB_hg19_150331All tandem repeats from TRDB
lowcmp_SimpleRepeat_quadTR_51to200Exact 4bp tandem repeats 51-200bp in length
lowcmp_AllRepeats_lt51bp_gt95identity_mergedAll perfect and imperfect tandem repeats <51bp in length
lowcmp_SimpleRepeat_homopolymer_6to10All homopolymers 6-10bp in length
lowcmp_SimpleRepeat_diTR_gt200Exact 2bp tandem repeats >200bp in length
lowcmp_SimpleRepeat_diTR_11to50Exact 2bp tandem repeats 11-50bp in length
lowcmp_SimpleRepeat_diTR_51to200Exact 2bp tandem repeats 51-200bp in length
lowcmp_Human_Full_Genome_TRDB_hg19_150331_TRgt6_gt200bp_gt95identity_mergedTandem repeats with >6bp unit size and >200bp in length and >95% identity
segdupSegmental duplications in GRCh37 of all sizes, not including ALTs from GRCh37
segdupwithaltSegmental duplications in GRCh37 >10kb, including ALTs from GRCh37
decoyRegions in GRCh37 to which decoy sequences in hs37d5 map
HG001compoundhetCompound heterozygous regions in which there are 2 different variants when phased variants within 50bp are combined in HG001/NA12878
HG002compoundhetCompound heterozygous regions in which there are 2 different variants when phased variants within 50bp are combined in HG002/NA24385
HG001complexvarComplex variant regions in which there are 2 or more variants within 50bp in HG001/NA12878
HG002complexvarComplex variant regions in which there are 2 or more variants within 50bp in HG002/NA24385

These dimensions help slice and dice the results according to all the different ways in which someone may want to look at them. The special value of '*' (for Subtype, Genotype, Subset) corresponds to summary statistics (i.e. as calculated across the whole genome in the confident regions, without further stratification).

Compilation of results

We ran hap.py (and specifically the HAP-207 version, with the engine set to a GA4GH-specific version of vcfeval) to compare HG001 and HG002 submissions against the GiaB/NIST v3.2.2 HG001 and HG002 truth data respectively. Our executions excluded the entry labeled ccogle-snppet as it did not call variants in the whole genome. The entries labeled ghariani-varprowl, jpowers-varprowl and qzeng-custom contained few VCF lines which were deemed incompatible by this new comparison framework (which performs stricter checks). Upon closer inspection, this was due to incompatible REF columns (such as the strings "nan" or "AC-7GATAGAA") or non-diploid genotypes (such as 0/1/2, or even 0/1/2/3). These were a small fraction, so we decided to exclude these offending lines from this comparison.

We have made available on precisionFDA complete archives of all the input files (including both the original and adjusted files, for the entries which we had to remove offending VCF lines), and the results (including the annotated VCF as output by hap.py, and extended statistics in CSV format). These are recapitulated in a precisionFDA post.

We would like to thank the benchmarking group of the Global Alliance for Genomics and Health for their excellent collaboration throughout this precisionFDA effort.

Results discussion
The compiled results (as described in the previous section) are summarized in the table at the top of this web page. In addition to that, we have created an interactive explorer which can be used to query the full set of all CSV files that were output by the comparison, and to show the metrics for any combination of the "Type", "Subtype", "Genotype" and "Subset" dimensions. You can use that to answer questions such as "how does a particular entry perform across different genome subsets", or "which entry performs best in large deletions?", etc.

EntryTypeSubtypeSubsetGenotypeF-scoreRecallPrecisionFrac_NATruth TPTruth FNQuery TPQuery FPFP gt% FP ma
201-250 / 86044 show all
rpoplin-dv42INDELD16_PLUSmap_l150_m1_e0*
96.5517
93.3333
100.0000
94.2623
1411400
rpoplin-dv42INDELD16_PLUSmap_l150_m1_e0het
96.2963
92.8571
100.0000
91.8750
1311300
rpoplin-dv42INDELD16_PLUSmap_l150_m1_e0hetalt
100.0000
100.0000
100.0000
93.7500
10100
rpoplin-dv42INDELD16_PLUSmap_l150_m1_e0homalt
0.0000
100.0000
00000
rpoplin-dv42INDELD16_PLUSmap_l150_m2_e0*
96.9697
94.1176
100.0000
94.1392
1611600
rpoplin-dv42INDELD16_PLUSmap_l150_m2_e0het
96.7742
93.7500
100.0000
91.7127
1511500
rpoplin-dv42INDELD16_PLUSmap_l150_m2_e0hetalt
100.0000
100.0000
100.0000
94.1176
10100
rpoplin-dv42INDELD16_PLUSmap_l150_m2_e0homalt
0.0000
100.0000
00000
rpoplin-dv42INDELD16_PLUSmap_l150_m2_e1*
94.1176
88.8889
100.0000
94.2238
1621600
rpoplin-dv42INDELD16_PLUSmap_l150_m2_e1het
96.7742
93.7500
100.0000
91.8919
1511500
rpoplin-dv42INDELD16_PLUSmap_l150_m2_e1hetalt
66.6667
50.0000
100.0000
94.1176
11100
rpoplin-dv42INDELD16_PLUSmap_l150_m2_e1homalt
0.0000
100.0000
00000
rpoplin-dv42INDELD16_PLUSmap_l250_m0_e0*
100.0000
100.0000
100.0000
98.5714
10100
rpoplin-dv42INDELD16_PLUSmap_l250_m0_e0het
100.0000
100.0000
100.0000
97.5610
10100
rpoplin-dv42INDELD16_PLUSmap_l250_m0_e0hetalt
0.0000
100.0000
00000
rpoplin-dv42INDELD16_PLUSmap_l250_m0_e0homalt
0.0000
100.0000
00000
rpoplin-dv42INDELD16_PLUSmap_l250_m1_e0*
100.0000
100.0000
100.0000
96.9697
40400
rpoplin-dv42INDELD16_PLUSmap_l250_m1_e0het
100.0000
100.0000
100.0000
96.0000
30300
rpoplin-dv42INDELD16_PLUSmap_l250_m1_e0hetalt
100.0000
100.0000
100.0000
92.3077
10100
rpoplin-dv42INDELD16_PLUSmap_l250_m1_e0homalt
0.0000
100.0000
00000
rpoplin-dv42INDELD16_PLUSmap_l250_m2_e0*
100.0000
100.0000
100.0000
96.5278
50500
rpoplin-dv42INDELD16_PLUSmap_l250_m2_e0het
100.0000
100.0000
100.0000
96.4286
30300
rpoplin-dv42INDELD16_PLUSmap_l250_m2_e0hetalt
100.0000
100.0000
100.0000
92.3077
10100
rpoplin-dv42INDELD16_PLUSmap_l250_m2_e0homalt
100.0000
100.0000
100.0000
97.8723
10100
rpoplin-dv42INDELD16_PLUSmap_l250_m2_e1*
100.0000
100.0000
100.0000
96.5986
50500
rpoplin-dv42INDELD16_PLUSmap_l250_m2_e1het
100.0000
100.0000
100.0000
96.5517
30300
rpoplin-dv42INDELD16_PLUSmap_l250_m2_e1hetalt
100.0000
100.0000
100.0000
92.3077
10100
rpoplin-dv42INDELD16_PLUSmap_l250_m2_e1homalt
100.0000
100.0000
100.0000
97.8723
10100
rpoplin-dv42INDELD16_PLUSmap_sirenhetalt
80.7692
67.7419
100.0000
73.8095
21102200
rpoplin-dv42INDELD16_PLUSsegduphetalt
80.0000
66.6667
100.0000
93.4783
63600
rpoplin-dv42INDELD16_PLUSsegduphomalt
100.0000
100.0000
100.0000
94.5946
1201200
rpoplin-dv42INDELD16_PLUSsegdupwithalt*
0.0000
100.0000
00000
rpoplin-dv42INDELD16_PLUSsegdupwithalthet
0.0000
100.0000
00000
rpoplin-dv42INDELD16_PLUSsegdupwithalthetalt
0.0000
100.0000
00000
rpoplin-dv42INDELD16_PLUSsegdupwithalthomalt
0.0000
100.0000
00000
rpoplin-dv42INDELD16_PLUStech_badpromoters*
100.0000
100.0000
100.0000
33.3333
40400
rpoplin-dv42INDELD16_PLUStech_badpromotershet
100.0000
100.0000
100.0000
0.0000
40400
rpoplin-dv42INDELD16_PLUStech_badpromotershetalt
0.0000
0.0000
0.0000
00000
rpoplin-dv42INDELD16_PLUStech_badpromotershomalt
0.0000
100.0000
00000
rpoplin-dv42INDELD1_5decoy*
100.0000
100.0000
100.0000
99.9439
40400
rpoplin-dv42INDELD1_5decoyhet
100.0000
100.0000
100.0000
99.9601
20200
rpoplin-dv42INDELD1_5decoyhetalt
100.0000
100.0000
100.0000
99.6700
10100
rpoplin-dv42INDELD1_5decoyhomalt
100.0000
100.0000
100.0000
99.9450
10100
rpoplin-dv42INDELD1_5func_cds*
100.0000
100.0000
100.0000
40.7407
159016000
rpoplin-dv42INDELD1_5func_cdshet
100.0000
100.0000
100.0000
45.9119
8508600
rpoplin-dv42INDELD1_5func_cdshetalt
0.0000
100.0000
00000
rpoplin-dv42INDELD1_5func_cdshomalt
100.0000
100.0000
100.0000
31.4815
7407400
rpoplin-dv42INDELD1_5lowcmp_AllRepeats_gt200bp_gt95identity_merged*
95.2381
90.9091
100.0000
99.2679
1011000
rpoplin-dv42INDELD1_5lowcmp_AllRepeats_gt200bp_gt95identity_mergedhet
93.3333
87.5000
100.0000
99.1453
71700
rpoplin-dv42INDELD1_5lowcmp_AllRepeats_gt200bp_gt95identity_mergedhetalt
100.0000
100.0000
100.0000
98.3333
10100
Interactively explore the results dataset


The interactive explorer includes an additional column, "% FP MA", indicating the percentage of the false positives which have matching non-ref alleles (i.e. FP.gt/QUERY.FP). For SNPs, this fraction varies greatly between pipelines, but for the majority it is less than 30%. Different callers would have the fewest FPs if these matching allele false positives were excluded from the FP counts, making clear that how performance metrics are defined can have a significant effect. For indels, this fraction varies as well, but is generally higher than 50%.

Using the interactive explorer you can also filter for different subsets, such as coding regions. Nonsurprisingly, there are very few FPs and FNs in coding regions for the accurate methods.

You can also see the fraction of calls that fall outside the confidence regions (reported as FRAC_NA). For SNPs, the fraction of calls outside GiaB high-confidence regions is ~10-25%, and for indels the fraction is higher than 50% for all methods. This is an important caveat to consider, since accuracy might be lower outside GiaB high-confidence regions.

F-score, Recall and Precision

Notably, most entries did much better at SNP accuracy measures than at indel accuracy measures. The majority of entries have high SNP f-score, recall, and precision -- above 99%. The respective numbers are lower for indels, and in fact indel recall is usually lower than indel precision. It is important to note, as mentioned in the truth data announcement, that the high-confidence calls have an increased FP, FN, and genotyping error rate for single base indels in homopolymers. This may be a contributing factor to the overall lower indel performance numbers.

Just like in the precisionFDA Consistency Challenge, it appears that pipelines are overall tuned for precision. Compared to the previous challenge, however, the input datasets had higher coverage, which may be a contributing factor to higher recall.

Award determination

We've determined the winners by taking the highest value in each metric (F-score, Recall, Precision) per variant type (SNPs vs indels). Please note that the award for highest f-score has been called "highest performance", to reflect the nature of the score. For each combination of metric per variant type, we are also recognizing entries with high values of that metric, based on a cutoff at the first substantial drop in the percentages. These cutoffs are ultimately subjective, and not directly applicable to other regulatory contexts (which may require well-defined thresholds ahead of time).

These awards and recognitions are meant to encourage the community to participate in challenges and do not constitute an endorsement by the FDA.

Beyond these results

We strongly urge the community to be cautious when interpreting these results. By their nature, high-confidence variant calls and regions tend to include a subset of variants and regions that are easier to characterize. Therefore, such benchmarking against the high-confidence truth data may in fact overestimate accuracy. Manual curation of sequence data in a genome browser for a subset of false positives and false negatives is essential for an accurate understanding of statistics like sensitivity and precision.

When interpreting results within stratification sub-categories, it may be useful to calculate confidence intervals around statistics like sensitivity because there may be very few examples of variants of some types in the benchmark calls in the stratification sub-category.

These results are based on the HG002 truth data. Pipelines may perform differently when confronted with other samples. For a more complete performance assessment of software pipelines, additional samples and experiments would be required.

Heterozygosity rates in chromosome X

We wanted to present an example of other ways in which this dataset can be analyzed, as a way to stimulate the community to conduct further experimentation with the precisionFDA Truth Challenge entries. Since the HG002 sample is male (in contrast to HG001, which is female), we decided to look at the performance of variation calling pipelines in chromosome X. It should be noted that chromosome X is not included in the confident regions of the GiaB truth data for HG002, so the comparisons have not evaluated that chromosome.

In XY male individuals, locations in chromosome X outside of the pseudoautosomal regions are expected to behave as haploid, and variant calling algorithms would ideally report them as haploid calls or as diploid homozygous calls. To measure that, we used bcftools stats -f .,PASS -r X:1-60001,X:2699521-154931044,X:155260561-155270560, which measures the count of heterozygous and homozygous SNPs in chromosome X outside of the PAR. We subsequently calculated the fraction of SNPs that are heterozygous, and generated the following table.

LabelPctHet
dgrover-gatk0.00%
jli-custom0.00%
ltrigg-rtg10.00%
ltrigg-rtg20.00%
mlin-fermikit0.51%
ckim-isaac1.10%
raldana-dualsentieon1.66%
hfeng-pmm11.76%
hfeng-pmm31.97%
hfeng-pmm22.04%
ndellapenna-hhga2.21%
egarrison-hhga2.30%
bgallagher-sentieon2.34%
astatham-gatk2.36%
rpoplin-dv422.42%
ckim-dragen4.17%
cchapple-custom5.11%
ckim-gatk5.23%
ckim-vqsr5.23%
asubramanian-gatk6.00%
ciseli-custom6.10%
qzeng-custom6.23%
jlack-gatk6.32%
gduggal-bwafb7.89%
jpowers-varprowl8.52%
eyeh-varpipe10.62%
ghariani-varprowl10.94%
gduggal-snapfb11.32%
gduggal-bwavard13.09%
gduggal-bwaplat16.11%
gduggal-snapvard16.21%
gduggal-snapplat17.29%
anovak-vg31.01%

Certain algorithms (such as the ones used in the ltrigg-rtg1 and ltrigg-rtg2 entries) can be gender-aware, and will not generate heterozygous calls when configured to run with a male option. Four entries in total had a zero heterozygosity fraction.

Final remarks

We want to thank those of you who participated in this challenge! As with our first challenge, by participating and putting your results and your thoughts out in the public, you fulfilled the first and most important goal of this challenge – to engage and start sharing data.

This challenge created a rich and interesting dataset that we hope people will study further. This dataset includes not only the submitted VCF files but also the complete set of comparison results (annotated VCFs and calculated statistics). It represents the essence of this challenge, and can hopefully be useful to inform future approaches, not only in terms of regulatory science but also for reference materials and benchmarking methodologies. We hope the GiaB group will be able to make use of it to improve reference calls (particularly around indels), and the GA4GH benchmarking group will be able to use it to improve the comparison methodology. We are excited to see how else the community will use this data set in the future.

Quotes

In closing, we would like to leave you with several encouraging quotes from our collaborators and larger community, which will hopefully inspire all readers to actively engage with the precisionFDA community, or participate in one of our upcoming challenges:


“We hope that in the near future, thanks to more well-characterized truth sets, the genomics community can determine whether these more-refined pipelines are indeed beneficial for enabling Precision Data for Precision Medicine.”
“Thanks for your patience in being a guinea pig for our high confidence calls and benchmarking tools! This has been really useful for us already!”
“These challenges are potentially of as much value to those constructing truth sets as they are to the rest of the community, as they provide a means to look at a wide variety of call sets with respect to multiple truth sets and start to delve into potential biases in various truth sets and the methods used to construct them. This is an evolving process and it is great to see how GIAB and GA4GH are also moving things forward in this regard.”
“We applaud the effort by NIST and FDA to develop more truth sets that the accuracy of all algorithms can be benchmarked against. We look forward to more comprehensive truth sets and open challenges that can drive the development of better pipeline algorithms and enable precision data for precision medicine.”
“Thanks for all your help! Hopefully through this process we all learned and improved across multiple fronts, so it's already a huge success. I’m already thinking of ways some of these results could be used in publications about benchmarking.”