Skip to content

Latest commit

 

History

History
130 lines (90 loc) · 4.63 KB

metrics.md

File metadata and controls

130 lines (90 loc) · 4.63 KB

Runtime and accuracy metrics for all release models

WGS (Illumina)

Runtime

Runtime is on HG003 (all chromosomes).

Stage Time (minutes)
make_examples ~105m
call_variants ~177m
postprocess_variants (with gVCF) ~78m
total ~360m = ~6 hours

Accuracy

hap.py results on HG003 (all chromosomes, using NIST v4.2.1 truth), which was held out while training.

Type TRUTH.TP TRUTH.FN QUERY.FP METRIC.Recall METRIC.Precision METRIC.F1_Score
INDEL 501537 2964 1201 0.994125 0.997708 0.995913
SNP 3306401 21095 4556 0.99366 0.998625 0.996136

WES (Illumina)

Runtime

Runtime is on HG003 (all chromosomes).

Stage Time (minutes)
make_examples ~7m
call_variants ~2m
postprocess_variants (with gVCF) ~1m
total ~10m

Accuracy

hap.py results on HG003 (all chromosomes, using NIST v4.2.1 truth), which was held out while training.

Type TRUTH.TP TRUTH.FN QUERY.FP METRIC.Recall METRIC.Precision METRIC.F1_Score
INDEL 1024 27 8 0.97431 0.99241 0.983277
SNP 24938 341 57 0.986511 0.99772 0.992083

PacBio (HiFi)

Runtime

Runtime is on HG003 (all chromosomes).

Stage Time (minutes)
make_examples ~114m
call_variants ~153m
postprocess_variants (with gVCF) ~68m
total ~335m = ~5.58 hours

Accuracy

hap.py results on HG003 (all chromosomes, using NIST v4.2.1 truth), which was held out while training.

(The input BAM is phased already and DeepVariant was run with --use_hp_information=true.)

Type TRUTH.TP TRUTH.FN QUERY.FP METRIC.Recall METRIC.Precision METRIC.F1_Score
INDEL 501870 2631 2684 0.994785 0.994892 0.994839
SNP 3323707 3788 1612 0.998862 0.999516 0.999189

Hybrid (Illumina + PacBio HiFi)

Runtime

Runtime is on HG003 (all chromosomes).

Stage Time (minutes)
make_examples ~161m
call_variants ~188m
postprocess_variants (with gVCF) ~55m
total ~404m = ~6.73 hours

Accuracy

Evaluating on HG003 (all chromosomes, using NIST v4.2.1 truth), which was held out while training the hybrid model.

Type TRUTH.TP TRUTH.FN QUERY.FP METRIC.Recall METRIC.Precision METRIC.F1_Score
INDEL 503240 1261 1984 0.997501 0.996261 0.99688
SNP 3323691 3804 1709 0.998857 0.999486 0.999171

How to reproduce the metrics on this page

For simplicity and consistency, we report runtime with a CPU instance with 64 CPUs This is NOT the fastest or cheapest configuration.

Use gcloud compute ssh to log in to the newly created instance.

Download and run any of the following case study scripts:

# Get the script.
curl -O https://raw.githubusercontent.com/google/deepvariant/r1.3/scripts/inference_deepvariant.sh

# WGS
bash inference_deepvariant.sh --model_preset WGS

# WES
bash inference_deepvariant.sh --model_preset WES

# PacBio
bash inference_deepvariant.sh --model_preset PACBIO

# Hybrid
bash inference_deepvariant.sh --model_preset HYBRID_PACBIO_ILLUMINA

Runtime metrics are taken from the resulting log after each stage of DeepVariant. The runtime numbers reported above are the average of 5 runs each. The accuracy metrics come from the hap.py summary.csv output file. The runs are deterministic so all 5 runs produced the same output.