A universal SNP and small-indel variant caller using deep neural networks

DeepVariant uses convolutional neural networks to improve the accuracy of variant calling. Despite rapid advances in sequencing technologies, accurately calling genetic variants present in an individual genome from billions of short, errorful sequence reads remains challenging. Here we show that a d...

Full description

Saved in:
Bibliographic Details
Published inNature biotechnology Vol. 36; no. 10; pp. 983 - 987
Main Authors Poplin, Ryan, Chang, Pi-Chuan, Alexander, David, Schwartz, Scott, Colthurst, Thomas, Ku, Alexander, Newburger, Dan, Dijamco, Jojo, Nguyen, Nam, Afshar, Pegah T, Gross, Sam S, Dorfman, Lizzie, McLean, Cory Y, DePristo, Mark A
Format Journal Article
LanguageEnglish
Published New York Springer New York 01.11.2018
Nature Publishing Group
Subjects
Online AccessGet full text
ISSN1087-0156
1546-1696
1546-1696
DOI10.1038/nbt.4235

Cover

More Information
Summary:DeepVariant uses convolutional neural networks to improve the accuracy of variant calling. Despite rapid advances in sequencing technologies, accurately calling genetic variants present in an individual genome from billions of short, errorful sequence reads remains challenging. Here we show that a deep convolutional neural network can call genetic variation in aligned next-generation sequencing read data by learning statistical relationships between images of read pileups around putative variant and true genotype calls. The approach, called DeepVariant, outperforms existing state-of-the-art tools. The learned model generalizes across genome builds and mammalian species, allowing nonhuman sequencing projects to benefit from the wealth of human ground-truth data. We further show that DeepVariant can learn to call variants in a variety of sequencing technologies and experimental designs, including deep whole genomes from 10X Genomics and Ion Ampliseq exomes, highlighting the benefits of using more automated and generalizable techniques for variant calling.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
content type line 23
ISSN:1087-0156
1546-1696
1546-1696
DOI:10.1038/nbt.4235