Methods And Systems For Identifying Disease-induced Mutations

  *US10053736B2*
  US010053736B2                                 
(12)United States Patent(10)Patent No.: US 10,053,736 B2
  et al. (45) Date of Patent:Aug.  21, 2018

(54)Methods and systems for identifying disease-induced mutations 
    
(75)Inventor: Seven Bridges Genomics Inc.,  Cambridge, MA (US) 
(73)Assignee:Seven Bridges Genomics Inc.,  Cambridge, MA (US), Type: US Company 
(*)Notice: Subject to any disclaimer, the term of this patent is extended or adjusted under 35 U.S.C. 154(b) by 630 days. 
(21)Appl. No.: 14/517,451 
(22)Filed: Oct.  17, 2014 
(65)Prior Publication Data 
 US 2015/0197815 A1 Jul.  16, 2015 
 Related U.S. Patent Documents 
(60)Provisional application No. 61/892,670, filed on Oct.  18, 2013.
 
Jan.  1, 2013 C 12 Q 1 6886 F I Aug.  21, 2018 US B H C Jan.  1, 2013 C 12 Q 1 6883 L I Aug.  21, 2018 US B H C Jan.  1, 2013 C 12 Q 2600 112 L A Aug.  21, 2018 US B H C Jan.  1, 2013 C 12 Q 2600 156 L A Aug.  21, 2018 US B H C Jan.  1, 2013 C 12 Q 2600 16 L A Aug.  21, 2018 US B H C
(51)Int. Cl. G06F 019/00 (20180101); C12Q 001/6886 (20180101); C12Q 001/6883 (20180101)

 
(56)References Cited
 
 U.S. PATENT DOCUMENTS
 4,683,195  A  7/1987    Kary et al.     
 4,683,202  A  7/1987    Mullis     
 4,988,617  A  1/1991    Landegren et al.     
 5,234,809  A  8/1993    Boom et al.     
 5,242,794  A  9/1993    Whiteley et al.     
 5,494,810  A  2/1996    Barany et al.     
 5,511,158  A  4/1996    Sims     
 5,583,024  A  12/1996    McElroy et al.     
 5,674,713  A  10/1997    McElroy et al.     
 5,700,673  A  12/1997    McElroy et al.     
 5,701,256  A  12/1997    Marr et al.     
 6,054,278  A  4/2000    Dodge et al.     
 6,210,891  B1  4/2001    Nyren et al.     
 6,223,128  B1  4/2001    Allex et al.     
 6,306,597  B1  10/2001    Macevicz     
 6,582,938  B1  6/2003    Su et al.     
 6,818,395  B1  11/2004    Quake et al.     
 6,828,100  B1  12/2004    Ronaghi     
 6,833,246  B2  12/2004    Balasubramanian     
 6,890,763  B2  5/2005    Jackowski et al.     
 6,911,345  B2  6/2005    Quake et al.     
 6,925,389  B2  8/2005    Hitt et al.     
 6,989,100  B2  1/2006    Norton     
 7,169,560  B2  1/2007    Lapidus et al.     
 7,232,656  B2  6/2007    Balasubramanian et al.     
 7,282,337  B1  10/2007    Harris     
 7,321,623  B2  1/2008    Dambrackas     
 7,483,585  B2  1/2009    Brakus     
 7,577,554  B2  8/2009    Lystad et al.     
 7,580,918  B2  8/2009    Chang et al.     
 7,598,035  B2  10/2009    Macevicz     
 7,620,800  B2  11/2009    Huppenthal et al.     
 7,776,616  B2  8/2010    Heath et al.     
 7,809,509  B2  10/2010    Milosavljevic     
 7,835,871  B2  11/2010    Kain et al.     
 7,885,840  B2  2/2011    Sadiq et al.     
 7,917,302  B2  3/2011    Rognes     
 7,957,913  B2  6/2011    Chinitz et al.     
 7,960,120  B2  6/2011    Rigatti et al.     
 8,146,099  B2  3/2012    Tkatch et al.     
 8,165,821  B2  4/2012    Zhang     
 8,209,130  B1  6/2012    Kennedy et al.     
 8,340,914  B2  12/2012    Gatewood et al.     
 8,370,079  B2  2/2013    Sorenson et al.     
 8,639,847  B2  1/2014    Biaszczak et al.     
 8,972,201  B2  3/2015    Mande et al.     
 9,063,914  B2  6/2015    Kural et al.     
 9,092,402  B2  7/2015    Kural et al.     
 9,116,866  B2  8/2015    Kural     
 9,390,226  B2  7/2016    Kural     
 9,817,944  B2  11/2017    Kural     
 2002//0164629  A1  11/2002    Quake et al.     
 2002//0190663  A1  12/2002    Rasmussen     
 2004//0023209  A1  2/2004    Jonasson     
 2005//0089906  A1  4/2005    Furuta et al.     
 2006//0024681  A1  2/2006    Smith et al.     
 2006//0195269  A1  8/2006    Yeatman et al.     
 2006//0292611  A1  12/2006    Berka et al.     
 2007//0114362  A1  5/2007    Feng et al.     
 2007//0166707  A1  7/2007    Schadt et al.     
 2008//0003571  A1  1/2008    McKernan et al.     
 2008//0077607  A1  3/2008    Gatawood et al.     
 2008//0251711  A1  10/2008    Reilly     
 2008//0281463  A1  11/2008    Suh et al.     
 2008//0294403  A1  11/2008    Zhu et al.     
 2009//0026082  A1  1/2009    Rothberg et al.     
 2009//0119313  A1  5/2009    Pearce     
 2009//0127589  A1  5/2009    Rothberg et al.     
 2009//0191565  A1  7/2009    Lapidus et al.     
 2009//0233809  A1  9/2009    Faham et al.     
 2009//0300781  A1  12/2009    Bancroft et al.     
 2009//0318310  A1  12/2009    Liu et al.     
 2009//0325145  A1  12/2009    Sablon et al.     
 2010//0010992  A1  1/2010    Morris     
 2010//0035252  A1  2/2010    Rothberg et al.     
 2010//0137143  A1  6/2010    Rothberg et al.     
 2010//0169026  A1  7/2010    Sorenson et al.     
 2010//0188073  A1  7/2010    Rothberg et al.     
 2010//0197507  A1  8/2010    Rothberg et al.     
 2010//0240046  A1  9/2010    Palmer et al.     
 2010//0282617  A1  11/2010    Rothberg et al.     
 2010//0285578  A1  11/2010    Selden et al.     
 2010//0300559  A1  12/2010    Schultz et al.     
 2010//0300895  A1  12/2010    Nobile et al.     
 2010//0301398  A1  12/2010    Rothberg et al.     
 2010//0304982  A1  12/2010    Hinz et al.     
 2011//0004413  A1  1/2011    Carnevali et al.     
 2011//0009278  A1  1/2011    Kain et al.     
 2011//0098193  A1  4/2011    Kingsmore et al.     
 2011//0207135  A1  8/2011    Faham et al.     
 2011//0257889  A1  10/2011    Klammer et al.     
 2012//0030566  A1  2/2012    Victor     
 2012//0040851  A1  2/2012    Lieberman et al.     
 2012//0041727  A1  2/2012    Mishra et al.     
 2012//0157322  A1  6/2012    Myllykangas et al.     
 2012//0239706  A1  9/2012    Steinfadt     
 2012//0330566  A1  12/2012    Chaisson     
 2013//0029879  A1  1/2013    Shetty et al.     
 2013//0035904  A1  2/2013    Kuhn     
 2013//0059740  A1  3/2013    Drmanac et al.     
 2013//0073214  A1  3/2013    Hyland et al.     
 2013//0124100  A1  5/2013    Drmanac et al.     
 2013//0232480  A1  9/2013    Winterfeldt et al.     
 2013//0311106  A1  11/2013    White et al.     
 2013//0332081  A1  12/2013    Reese et al.     
 2013//0345066  A1  12/2013    Brinza et al.     
 2014//0012866  A1  1/2014    Bowman et al.     
 2014//0025312  A1  1/2014    Chin et al.     
 2014//0051588  A9  2/2014    Drmanac et al.     
 2014//0129201  A1  5/2014    Kennedy et al.     
 2014//0136120  A1  5/2014    Colwell et al.     
 2014//0200147  A1  7/2014    Bartha et al.     
 2014//0278590  A1  9/2014    Abbassi et al.     
 2014//0280360  A1  9/2014    Webber et al.     
 2014//0281708  A1  9/2014    Adam et al.     
 2014//0323320  A1  10/2014    Jia et al.     
 2015//0020061  A1  1/2015    Ravi     
 2015//0056613  A1  2/2015    Kural     
 2015//0057946  A1  2/2015    Kural     
 2015//0066383  A1  3/2015    Wernicke     
 2015//0094212  A1  4/2015    Gottimukkala et al.     
 2015//0110754  A1  4/2015    Bai et al.     
 2015//0112602  A1  4/2015    Kural et al.     
 2015//0112658  A1  4/2015    Kural et al.     
 2015//0197815  A1  7/2015    Kural     
 2015//0199472  A1  7/2015    Kural     
 2015//0199473  A1  7/2015    Kural     
 2015//0199474  A1  7/2015    Kural     
 2015//0199475  A1  7/2015    Kural     
 2015//0227685  A1  8/2015    Kural     
 2015//0293994  A1  10/2015    Kelly     
 2015//0302145  A1  10/2015    Kural et al.     
 2015//0310167  A1  10/2015    Kural et al.     
 2015//0344970  A1  12/2015    Vogelstein et al.     
 2015//0347678  A1  12/2015    Kural     
 2015//0356147  A1  12/2015    Mishra et al.     
 2016//0259880  A1  9/2016    Semenyuk     
 2016//0306921  A1  10/2016    Kural     
 2016//0364523  A1  12/2016    Locke et al.     
 2017//0058320  A1  3/2017    Locke et al.     
 2017//0058341  A1  3/2017    Locke et al.     
 2017//0058365  A1  3/2017    Locke et al.     
 2017//0193351  A1  7/2017    Lee et al.     
 2017//0199959  A1  7/2017    Locke     
 2017//0199960  A1  7/2017    Ghose et al.     
 2017//0242958  A1  8/2017    Brown     

 
 FOREIGN PATENT DOCUMENTS 
 
       KR       101282798       B1                7/2013      
       WO       2007086935       A3                8/2007      
       WO       2010010992       A1                1/2010      
       WO       2011139797       A2                11/2011      
       WO       2012096579       A3                7/2012      
       WO       2012098515       A1                7/2012      
       WO       2012142531       A2                10/2012      
       WO       2013035904       A1                3/2013      
       WO       2013043909       A1                3/2013      
       WO       2013106737       A1                7/2013      
       WO       2013184643       A1                12/2013      
       WO       2015027050       A1                2/2015      
       WO       2015048753       A1                4/2015      
       WO       2015058093       A1                4/2015      
       WO       2015058095       A1                4/2015      
       WO       2015058097       A1                4/2015      
       WO       2015058120       A1                4/2015      
       WO       2015061099       A1                4/2015      
       WO       2015061103       A1                4/2015      
       WO       2015105963       A1                7/2015      
       WO       2015123269       A1                8/2015      
       WO       2016141294       A1                9/2016      
       WO       2016201215       A1                12/2016      
       WO       2017066753       A1                4/2017      
       WO       2017120128       A1                7/2017      
       WO       2017123864       A1                7/2017      
       WO       2017147124       A1                8/2017      

 OTHER PUBLICATIONS
  
  Chuang and Roth, 2001, Gene recognition based on DAG shortest paths, Bioinformatics 17(Suppl. 1):s56-s64.
  Florea and Salzberg, 2013, Genome-guided transcriptome assembly in the age of next-generation sequencing, IEEE/ACM Trans Comp Biol Bioinf 10(5):1234-1240.
  Gotoh, 1999, Multiple sequence alignment: algorithms and applications, Adv Biophys 36:159-206.
  International Search Report and Written Opinion dated Mar. 19, 2015, for International Application No. PCT/US2014/061162 with International Filing Date Oct. 17, 2014 (12 pages).
  Katoh, 2005, MAFFT version 5: improvement in accuracy of multiple sequence alignment, Nucl Acids Res 33(2):511-518.
  Kehr et al., 2014, Genome alignment with graph data structures: a comparison, BMC Bioinformatics 15:99.
  Kim et al., 2013, TopHat2: accurate alignment of transcriptomes in the presence of insertions, deletions and gene fusions, Genome Biology 14(4):R36.
  Lee et al., 2002, Multiple sequence alignment using partial order graphs, Bioinformatics 18(3):452-464.
  Lee, 2003, Generating consensus sequences from partial order multiple sequence alignment graphs, Bioinformatics 19(8):999-1008.
  Li & Homer, 2010, A survey of sequence alignment algorithms for next-generation sequencing, Briefings in Bioinformatics 11(5):473-483.
  Lucking et al., 2011, PICS-Ord: unlimited coding of ambiguous regions by pairwise identity and cost scores ordination, BMC Bioinf 12:10.
  Schneeberger et al., 2009, Simulataneous alignment of short reads against multiple genomes, Genome Biology 10: R98.
  Subramanian et al., 2008, DIALIGN-TX: greedy and progessive approaches for segment-based multiple sequence alignment, Alg Mol Biol 3(1):1-11.
  Wallace et al., 2005, Multiple sequence alignments, Curr Op Struct Biol 15(3):261-266.
  Abouelhoda, 2012, Tavaxy: integrating Taverna and Galaxy workflows with cloud computing support, BMC Bioinformatics 13:77.
  Agarwal, 2013, SINNET: Social Interaction Network Extractor from Text, Proc IJCNLP 33-36.
  Aguiar, 2012, HapCompass: A fast cycle basis algorithm for accurate haplotype assembly of sequence data, J Comp Biol 19(6):577-590.
  Aguiar, 2013, Haplotype assembly in polyploid genomes and identical by descent shared tracts, BioInformatics 29(13): i352-i360.
  Airoldi, 2008, Mixed membership stochastic blockmodels, JMLR 9:1981-2014.
  Albers, 2011, Dindel: Accurate indel calls from shod-read data, Genome Research 21:961-973.
  Altera, 2007, Implementation of the Smith-Waterman algorithm on reconfigurable supercomputing platform, White Paper ver 1.0 (18 pages).
  Altschul, 1986, Optimal Sequence Alignment Using Affine Gap Costs, Bull Math Biol 48(5/6):603-616.
  Bansal, 2008, An MCMC algorithm for haplotype assembly from whole-genome sequence data, Genome Res 18:1336-1346.
  Bao, 2013, BRANCH: boosting RNA-Seq assemblies with partial or related genomic sequences, Bioninformatics 29 (10):1250-1259.
  Berlin, 2014, Assembling large genomes with single-molecule sequencing and locality sensitive hashing, bioRxlv preprint (35 pages); retrieved from the internet on Jan. 29, 2015, at.
  Bertone, 2004, Global Identification of human transcribed sequences with genome tiling arrays; Science 306:2242-2246.
  Bertrand, 2009, Genetic map refinement using a comparative genomic approach, J Comp Biol 16(10):1475-1486.
  Black, 2005, A simple answer for a splicing conundrum, PNAS 102:4927-8.
  Boyer, 1977, A Fast String Searching Algorithm, Comm ACM 20(10):762-772.
  Buhler, 2001, Search algorithms for biosequences using random projection, dissertation, University of Washington (203 pages); retreived from the internet on Jun. 3, 2016, at.
  Carrington, 1985, Polypeptide ligation occurs during post-translational modification of concanavalin A, Nature 313:64-67.
  Chang, 2005, The application of alternative splicing graphs in quantitative analysis of alternative splicing form from EST database, Int J Comp Appl Tech 22(1):14.
  Chen, 2012, Transient hypermutability, chromothripsis and replication-based mechanisms in the generation of concurrent clustered mutations, Mutation Res 750(1):562-59.
  Chin, 2013, Nonhybrid finished microbial genome assemblies from long-read SMRT sequencing data, Nat Meth 10(6):563-569.
  Clark, 2014, Illumina announces landmark $1,000 human genome sequencing, Wired, Jan. 15, 2014.
  Cock, 2013, Galaxy tools and workflows for sequence analysis with applications in molecular plant pathology, Peer J 1: e167.
  Cohen-Boulakia, 2014, Distilling structure in Taverna scientific workflows: a refactoring approach, BMC Bioinformatics 15(Suppl 1):S12.
  Compeau, 2011, How to apply de Bruijn graphs to genome assembly, Nat Biotech 29(11):987-991.
  Costa, 2010, Uncovering the Complexity of Transcriptomes with RNA-Seq, J Biomed Biotech 853916.
  Danecek, 2011, The variant call format and VCFtools, Bioinformatics 27(15):2156-2158.
  Delcher, 1999, Alignment of whole genomes, Nucl. Acids Res 27(11):2369-76.
  DePristo, 2011, A framework for variation discovery and genotyping using next-generation DNA sequencing data, Nat Gen 43:491-498.
  Dinov, 2011, Applications of the pipeline environment for visual informatics and genomic computations, BMC Bioinformatics 12:304.
  Dudley, 2009, A quick guide for developing effective bioinformatics programming skills, PLoS Comput Biol 5(12): e1000589.
  Durbin, 2014, Efficient haplotype matching and storage using the positional Burrows-Wheeler transform (PBWT), Bioinformatics 30(9):1266-1272.
  Durham, 2005, EGene: a configurable pipeline system for automated sequence analysis, Bioinformatics 21(12):2812-2813.
  Enedelman, 2011, New algorithm improves fine structure of the barley consensus SNP map, BMC Genomics 12(1):407 (and whole document).
  Farrar, 2007, Striped Smith-Waterman speeds database searches six times over other SIMD implementations, Bioinformatics 23(2):156-161.
  Fiers, 2008, High-throughput Bioinformatics with the Cyrille2 Pipeline System, BMC Bioinformatics 9:96.
  Fitch, 1970, Distinguishing homologous from analogous proteins, Systematic Zoology 19:99-113.
  Flicek, 2009, Sense from sequence reads: methods for alignment and assembly, Nat Meth Suppl 6(11s)s6-s12.
  Florea, 2005, Gene and alternative splicing annotation with AIR, Genome Res 15:54-66.
  Garber, 2011, Computational methods for transcriptome annotation and quantification using RNA-Seq, Nat Meth 8(6):469-477.
  Glusman, 2014, Whole-genome haplotyping approaches and genomic medicine, Genome Med 6:73.
  Goto, 2010, BioRuby: bioinformatics software for the Ruby programming language, Bioinformatics 26(20):2617-2619.
  Gotoh, 1982, An improved Algorithm for Matching Biological Sequences, J Mol Biol 162:705-708.
  Grabherr, 2011, Full-length transcriptome assembly from RNA-Seq data without a reference genome, Nat Biotech 29(7):644-654.
  Grasso, 2004, Combining partial order alignment and progressive multiple sequence alignment increases alignment speed and scalability to very large alignment problems, Bioinformatics 20(10):1546-1556.
  Guttman 2010, Ab initio reconstruction of cell type-specific transcriptomes in mouse reveals the conserved multi-exonic structure of lincRNAs, Nat Biotech 28(5):503-510.
  Guttman; 2010, Ab initio reconstruction of transcriptomes of pluripotent and lineage committed cells reveals gene structures of thousands of lincRNAs, NIH-PA Author Manuscript.
  Haas, 2004, DAGchainer: a tool for mining segmental genome duplications and synteny, Bioinformatics 20(18):3643-3646.
  Harenberg, 2014, Community detection in large-scale networks: a survey and empirical evaluation, WIREs Comp Stat 6:426-439.
  Harrow, 2012, GENCODE: The reference human genome annotation for The ENCODE Project, Genome Res 22:1760-1774.
  He, 2010, Optimal algorithms for haplotype assembly from whole-genome sequence data, Bioinformatics 26:i183-i190.
  Heber, 2002, Splicing graphs and EST assembly problems, Bioinformatics 18 Suppl:181-188.
  Hein, 1989, A new method that simultaneously aligns and reconstructs ancestral sequences for any number of homologous sequences when the phylogeny is given, Mol Biol Evol 6(6):649-668.
  Hein, 1989, A tree reconstruction method that is economical in the number of pairwise comparisons used, Mol Biol Evol 6(6):649-668.
  Hokamp, 2003, Wrapping up BLAST and Other Applications for Use on Unix Clusters, Bioinformatics 19(3)441-42.
  Holland, 2008, BioJava: an open-source framework for bioinformatics, Bioinformatics 24(18):2096-2097.
  Homer, 2010, Improved variant discovery through local re-alignment of short-read next generation sequencing data using SRMA, Genome Biol 11(10):R99.
  Hoon, 2003, Biopipe: A flexible framework for protocol-based bioinformatics analysis, Genome Res 13(8):1904-1915.
  Horspool, 1980, Practical Fast Searching in Strings, Software—Practice & Experience 10:501-506.
  Huang, Chapter 3: Bio-Sequence Comparison and Alignment, ser. Curr Top Comp Mol Biol. Cambridge, Mass.: The MIT Press, 2002.
  Hull, 2006, Taverna: a tool for building and running workflows of services, Nucl Acids Res 34(Web Server Issue): W729-32.
  Hutchinson, 2014, Allele-specific methylation occurs at genetic variants associated with complex diseases, PLoS One 9(6):e98464.
  International HapMap Consortium, 2005, A haplotype map of the human genome. Nature 437:1299-1320.
  International Preliminary Report on Patentability issued in application No. PCT/US2014/052065 dated Feb. 23, 2016.
  International Search Report and Written Opinion dated Mar. 31, 2015 for International Application No. PCT/US2015/010604 filed Jan. 8, 2015 (13 pages).
  International Search Report and Written Opinion dated Dec. 30, 2014, for PCT/US14/58328, with International Filing Date Sep. 30, 2014 (15 pages).
  International Search Report and Written Opinion dated Feb. 17, 2015, for International Patent Application No. PCT/US2014/061156, filed Oct. 17, 2014 (19 pages).
  International Search Report and Written Opinion dated Jan. 5, 2016, for International Patent Application PCT/US2015/054461 with International Filing Date Oct. 7, 2015 (7 pages).
  International Search Report and Written Opinion dated May 11, 2015, for International Patent Application No. PCT/US2015/015375 with International Filing Date Feb. 11, 2015 (12 pages).
  International Search Report and Written Opinion dated May 5, 2016, for International Patent Application No. PCT/US2016/020899, wiht International Filing Date Mar. 4, 2016 (12 pages).
  International Search Report and Written Opinion dated Dec. 11, 2014, for International Patent Application No. PCT/US14/52065, filed Aug. 21, 2014, (18 pages).
  International Search Report and Written Opinion dated Dec. 30, 2014, for International Patent Application No. PCT/US14/58328, filed Sep. 30, 2014 (22 pages).
  International Search Report and Written Opinion dated Feb. 4, 2015, for Patent Application No. PCT/US2014/061158, filed Oct. 17, 2014, (11 pages).
  International Search Report and Written Opinion dated Jan. 27, 2015, for International Patent Application No. PCT/US2014/060680, filed Oct. 215, 2014, (11 pages).
  International Search Report and Written Opinion dated Sep. 2, 2016, for International Patent Application No. PCT/US2016/033201 with International Filing Date May 19, 2016 (14 pages).
  International Search Report and Written Opinion dated Sep. 7, 2016, for International Application No. PCT/US2016/036873 with International filing date Jun. 10, 2016 (8 pages).
  International Search Report and Written Opinion of the International Searching Authority dated Nov. 17, 2015 for International Application No. PCT/US2015/048891 (11 Pages).
  Kano, 2010, Text mining meets workflow: linking U-Compare with Taverna, Bioinformatics 26(19):2486-7.
  Kawas, 2006, BioMoby extensions to the Taverna workflow management and enactment software, BMC Bioinformatics 7:523.
  Kent, 2002, BLAT—The Blast-Like Alignment Tool, Genome Research 4:656-664.
  Kim, 2005, ECgene: Genome-based EST clustering and gene modeling for alternative splicing, Genome Res 15:566-576.
  Kim, 2008, A Scaffold Analysis Tool Using Mate-Pair Information in Genome Sequencing, Journal of Biomedicine and Biotechnology 8(3):195-197.
  Krabbenhoft, 2008, Integrating ARC grid middleware with Taverna workflows, Bioinformatics 24(9):1221-2.
  Kuhn, 2010, CDK-Taverna: an open workflow environment for cheminformatics, BMC Bioinformatics 11:159.
  Kumar, 2010, Comparing de novo assemblers for 454 transcriptome data, BMC Genomics 11:571.
  Kurtz, 2004, Versatile and open software for comparing large genomes, Genome Biol 5:R12.
  LaFramboise, 2009, Single nucleotide polymorphism arrays: a decade of biological, computational and technological advance, Nucleic Acids Res 37(13):4181-4193.
  Lam, 2008, Compressed indexing and local alignment of DNA, Bioinformatics 24(6):791-97.
  Langmead, 2009, Ultrafast and memory-efficient alignment of short DNA sequences to the human genome, Genome Biol 10:R25.
  Lanzen, 2008, The Taverna Interaction Service: enabling manual interaction in workflows, Bioinformatics 24(8):1118-20.
  Larkin, 2007, Clustal Wand Clustal X version 2.0, Bioinformatics 23(21):2947-2948.
  Layer, 2015, Efficient compression and analysis of large genetic variation datasets, Biorxiv preprint, available at http://biorxiv.org/content/early/2015/04/20/018259.
  Layer, 2015, Efficient genotype compression and analysis of large genetic-variation data sets, Nat Meth 13(1):63-65.
  Lee, 2005, Bioinformatics analysis of alternative splicing, Brief Bioinf 6(1):23-33.
  Lee, 2014, Accurate read mapping using a graph-based human pan-genome, ASHG 2014 Abstracts.
  Lee, 2014, MOSAIK: A hash-based algorithm for accurate next-generation sequencing short-read mapping, PLoS One 9(3):e90581.
  LeGault, 2010, Learning Probalistic Splice Graphs from RNA-Seq data, pages.cs.wisc.edu/˜legault/cs760_writeup.pdf; retrieved from the internet on Apr. 6, 2014.
  LeGault, 2013, Inference of alternative splicing from RNA-Seq data with probabilistic splice graphs, Bioinformatics 29(18):2300-2310.
  Leipzig, 2004, The alternative splicing gallery (ASG): Bridging the gap between genome and transcriptome, Nuc Acids Res 23(13):3977-3983.
  Li, 2008, Automated manipulation of systems biology models using libSBML within Taverna workflows, Bioinformatics 24(2):287-9.
  Li, 2008, Performing statistical analyses on quantitative data in Taverna workflows: an example using R and maxdBrowse to identify differentially-expressed genes from microarray data, BMC Bioinformatics 9:334.
  Li, 2008, SOAP: short oligonucleotide alignment program, Bioinformatics 24(5):713-14.
  Li, 2009, Fast and accurate short read alignment with Burrows-Wheeler Transform. Bioinformatics 25:1754-60.
  Li, 2009, SOAP2: an improved ultrafast tool for short read alignment, Bioinformatics 25(15):1966-67.
  Li, 2009, The Sequence Alignment/Map format and SAMtools, Bioinformatics 25(16):2078-9.
  Li, 2015, BGT: efficient and flexible genotype query across many samples, arXiv:1506.08452 [q-bio.GN].
  Li, 2015, Towards Better Understanding of Artificats in Variant Calling from High-Coverage Samples, arXiv:1404.0929 [q-bio.GN].
  Life Technologies, 2013, Rapid Exome Sequencing Using the Ion Proton System and Ion Ampllseq Technology, Application Note (5 Pages).
  Lindgreen, 2012, AdapterRemoval: easy cleaning of next-generation sequence reads, BMC Res Notes 5:337.
  Lipman, 1985, Rapid and sensitive protein similarity searches, Science 227(4693):1435-41.
  Ma, 2010, Multiple genome alignment based on longest path in directed acyclic graphs, Int J Bioinformatics 6(4):366-683.
  Machine translation of KR 10-1282798 B1 generated on Jan. 6, 2016, by the website of the European Patent Office (23 pages).
  Machine translation produced on Jun. 1, 2015, by Espacenet of WO 2010/010992 A1 (11 pages).
  Machine translation produced on Jun. 1, 2015, by WPIO website of WO 2013/035904 (10 pages).
  Mamoulis, 2004, Non-contiguous sequence pattern queries, in Advances in Database Technology—EDBT 2004: 9th International Conference on Extending Database Technology, Heraklion, Crete, Greece, Mar. 14-18, 2004, Proceedings (18 pages); retreived from the Internet on Jun. 3, 2016, at.
  Manolio, 2010, Genome wide association studies and assessment of the risk of disease, NEJM 363(2):166-76.
  Mardis, 2010, The $1,000 genome, the $100,000 analysis?, Genome Med 2:84-85.
  Margulies, 2005, Genome sequencing in micro-fabricated high-density picotiter reactors, Nature 437:376-380.
  Marth, 1999, A general approach to single-nucleotide polymorphism discovery, Nature Genetics 23:452-456.
  Mazrouee, 2014, FastHap: fast and accurate single individual haplotype reconstructions using fuzzy conflict graphs, Bioinformatics 30:i371-i378.
  McKenna, 2010, The Genome Analysis Toolkit: a MapReduce framework for analyzing next-generation DNA sequencing data, Genome Res 20(9):1297-1303.
  McSherry, 2001, Spectral partitioning of random graphs, Proc 42nd IEEE Symp Found Comp Sci 529-537.
  Miller, 2010, Assembly Algorithms for Next-Generation Sequencing Data, Genomics 95(6):315-327.
  Misra, 2011, Anatomy of a hash-based long read sequence mapping algorithm for next generation DNA sequencing, Bioinformatics 27(2):189-195.
  Missier, 2010, Taverna, reloaded, Proc. Scientific and Statistical Database Management, 22nd Int Conf, Heidelberg, Germany, Jun./Jul. 2010, Gertz & Ludascher, Eds., Springer.
  Moudrianakis, 1965, Base sequence determination in nucleic acids with electron microscope III: chemistry and microscopy of guanine-labelled DNA, PNAS 53:564-71.
  Mount, 2001, Multiple Sequence Alignment, Bioinformatics, 2001, Cold Spring Harbor Laboratory Press, Cold Spring Harbor, New York, pp. 139-204.
  Nagalakshmi, 2010, RNA-Seq: A Method for Comprehensive Transcriptome Analysis, Curr Proc Mol Biol 4.11.1.13.
  Nagarajan, 2013, Sequence assembly demystified, Nat Rev 14:157-167.
  Najafi, 2016, Fundamental limits of pooled-DNA sequencing, arXiv:1604.04735.
  Nakao, 2005, Large-scale analysis of human alternative protein isoforms: pattern classification and correlation with subcellular localization signals, Nucl Ac Res 33(8):2355-2363.
  Needleman, 1970, A general method applicable to the search for similarities in the amino acid sequence of two proteins, J Mol Biol 48(3):443-453.
  Nenadic, 2010, Nested Workflows, The Taverna Knowledge Blog, Dec. 13, 2010. Retrieved on Feb. 25, 2016 from http://taverna.knowledgeblog.org/2010/12/13/nested-workflows/.
  Newman, 2013, Community detection and graph portioning, Europhys Lett 103(2):28003, arXiv:1305.4974v1.
  Ning, 2001, SSAHA: a fast search method for large DNA databases, Genome Res 11(10):1725-9.
  Oinn, 2004, Taverna: a tool for the composition and enactment of bioinformatics workflows, Bioinformatics 20(17):3045-54.
  Oinn, 2006, Taverna: lessons in creating a workflow environment for the life sciences, Concurrency and Computation: Practice and Experience 18(10)1067-1100.
  O'Rawe, 2013, Low Concordance of Multiple Variant-Calling Pipelines: Practical Implications for Exome and Genome Sequencing, Genome Med 5:28.
  Oshlack, 2010, From RNA-seq reads to differential expression results. Genome Bio 11:220.
  Pabinger, 2013, A survey of tools for variant analysis of next-generation genome sequencing data, Brief Bioinf.
  Paterson, 2009, An XML transfer schema for exchange of genomic and genetic mapping data: Implementation as a web service in a Taverna workflow, BMC Bioinformatics 10:252.
  Pearson, 1988, Improved tools for biological sequence comparison, PNAS 85(8):2444-8.
  Pe'er, 2006, Evaluating and improving power in whole-genome association studies using fixed marker sets. Nat Genet 38:663-667.
  Peixoto, 2014, Efficient Monte Carlo and greedy heuristic for the inference of stochastic block models, Phys. Rev. E 89, 012804.
  Pope, 2014, ROVER Variant Caller: Read-Pair Overlap Considerate Variant-Calling Software Applied to PCR-Based Massively Parallel Sequencing Datasets, Source Code Bio Med 9:3.
  Popitsch, 2013, NGC: lossless and lossy compression of aligned high-throughput sequencing data, Nucl Acids Res, 41(1):e27.
  Posada, 1998, Model Test: testing the model of DNA substitution, Bioinformatics 14(9):817-8.
  Potter, 1994, ASC: An Associative-Computing Paradigm, Computer 27(11):19-25.
  Potter, 2004, The ensemble analysis pipeline, Genome Res 14:934-941.
  Quail, et al., 2012, A tale of three next generation sequencing platforms: comparison of Ion Torrent, Pacific Biosciences and Illumina MiSeq sequencers, BMC Genomics 13:341.
  Rajaram, 2013, Pearl millet [Pennisetum glaucum (L.) R. Br.] consensus linkage map constructed using four RIL mapping populations and newly developed EST-SSRs, BMC Genomics 14(1):159.
  Ramirez-Gonzalez, 2011, Gee Fu: a sequence version and web-services database tool for genomic assembly, genome feature and NGS data, Bioinformatics 27(19):2754-2755.
  Raphael, 2004; A novel method for multiple alignment of sequences with repeated and shuffled elements, Genome Res 14:2336-2346.
  Robertson, 2010, De novo assembly and analysis of RNA-seq data, Nat Meth 7(11):909.
  Rodelsperger, 2008, Syntenator: Multiple gene order alignments with a gene-specific scoring function, Alg Mol Biol 3:14.
  Rognes, 2000, Six-fold speed-up of Smith-Waterman sequence database searching using parallel processing on common microprocessors, Bioinformatics 16(8):699-706.
  Rognes, 2001, ParAlign: a parallel sequence alignment algorithm for rapid and sensitive database searches; Nucl Ac Res 29(7):1647-1652.
  Rognes, 2011, Faster Smith-Waterman database searches with inter-sequence SIMD parallelisation, Bioinformatics 12:221.
  Ronquist, 2012, MrBayes 3.2: efficient Bayesian phylogenetic inference and model choice across a large model space, Syst Biol 61(3):539-42.
  Rothberg, 2011, An integrated semiconductor device enabling non-optical genome sequencing, Nature 475:348-352.
  Saebo, 2005, PARALIGN: rapid and sensitive sequence similarity searches powered by parallel computing technology, Nucl Ac Res 33:W535-W539.
  Sato, 2008, Directed acyclic graph kernels for structural RNA analysis, BMC (BioMed Central) Bioinformatics 9(318).
  Schenk, 2013, A pipeline for comprehensive and automated processing of electron diffraction data in IPLT, J Struct Biol 182(2):173-185.
  Schwikowski, 2002, Weighted sequence graphs: boosting iterated dynamic programming using locally suboptimal solutions, Disc Appl Mat 127:95-117.
  Shao, 2006, Bioinformatic analysis of exon repetition, exon scrambling and trans-splicing in humans, Bioinformatics 22: 692-698.
  Sievers, 2011, Fast, scalable generation of high-quality protein multiple sequence alignments using Clustal Omeag, Mol Syst Biol 7:539.
  Slater, 2005, Automated generation of heuristics for biological sequence comparison, BMC Bioinformatics 6:31.
  Smith, 1981, identification of common molecular subsequences, J Mol Biol, 147(1):195-197.
  Smith, 2012, Multiple insert size paired-end sequencing for deconvolution of complex transcriptions, RNA Bio 9(5)596-609.
  Soni, 2007, Progress toward ultrafast DNA sequencing using solid-state nanopores, Clin Chem 53(11):1996-2001.
  Sosa, 2012, Next-Generation Sequencing of Human Mitochondrial Reference Genomes Uncovers High Heteroplasmy Frequency, PLoS One 8(10):e1002737.
  Sroka, 2006, XQTav: an XQuery processor for Taverna environment, Bioinformatics 22(10):1280-1.
  Sroka, 2010, A formal semantics for the Taverna 2 workflow model, J Comp Sys Sci 76(6):490-508.
  Sroka, 2011, CalcTav—integration of a spreadsheet and Taverna workbench, Bioinformatics 27(18):2618-9.
  Stephens, 2001, A new statistical method for haplotype reconstruction from population data, Am J Hum Genet 68:978-989.
  Stewart, 2011, A comprehensive map of mobile element insertion polymorphisms in humans, PLoS Genetics 7(8):1-19.
  Sudmant, 2015, An integrated map of structural variation in 2,504 human genomes, Nature 526:75-81.
  Sun, 2006, Pairwise Comparison Between Genomic Sequences and Optical maps, dissertation, New York University (131 pages); retreived from the internet on Jun. 3, 2016, at.
  Szaikowski, 2012, Fast and robust multiple sequence alignment with phylogeny-aware gap placement, BMC (BioMed Central) Bioinformatics 13(129).
  Szaikowski, 2013, Graph-based modeling of tandem repeats improves global multiple sequence alignment, Nucl Ac Res 41(17):e162.
  Tan, 2010, A Comparison of Using Taverna and BPEL in Building Scientific Workflows: the case of caGrid, Concurr Comput 22(9):1098-1117.
  Tan, 2010, CaGrid Workflow Toolkit: a Taverna based workflow tool for cancer grid, BMC Bioinformatics 11:542.
  Tarhio, 1993, Approximate Boyer-Moore String Matching, SIAM J Comput 22(2):243-260.
  Tewhey, 2011, The importance of phase information for human genomics, Nat Rev Gen 12:215-223.
  The 1000 Genomes Project, 2015, A global reference for human genetic variation, Nature 526:68-74.
  The Variant Call Format (VCF) Version 4.2 Specification (Jan. 26, 2015), available at https://samtools.github.io/hts-specs/VCFv4.2.pdf.
  Thomas, 2014, Community-wide effort aims to better represent variation in human reference genome, Genome Web (11 pages).
  Torri, 2012, Next generation sequence analysis and computational genomics using graphical pipeline workflows, Genes (Basel) 3(3):545-575.
  Trapnell, 2009, TopHat: discovering splice junctions with RNA-Seq, Bioinformatics 25:1105-1111.
  Trapnell, 2010, Transcript assembly and abundance estimation from RNA-Seq reveals thousands of new transcripts and switching among isoforms, Nat Biotech 28(5):511-515.
  Trapnell, 2010, Transcript assembly and quantification by RNA-Seq reveals unannotated trancripts and isoform switching during cell differentiation, Nat Biotech 28(5):511-515.
  Truszkowski, 2011, New developments on the cheminformatics open workflow environment CDK-Taverna, J Cheminform 3:54.
  Turi, 2007, Taverna Workflows: Syntax and Semantics, IEEE Int Conf on e-Science and Grid Computing 441-448.
  Wang, 2009, RNA-Seq: a revolutionary tool for transcriptomics, Nat Rev Genet 10(1):57-63.
  Wang, 2011, Next sequencing has lower sequence coverage and poorer SNP-detection capability in the regulatory regions, Scientific Reports 1:55.
  Wassink, 2009, Using R in Taverna: RShell v1.2. BMC Res Notes 2:138.
  Waterman, 1976, Some biological sequence metrics, Adv Math 20(3):367-387.
  Wellcome Trust Case Control Consortium, 2007, Genome-wide association study of 14,000 cases of seven common diseases and 3,000 shared controls, Nature 447:661-678.
  Wolstencroft, 2005, Panoply of Utilities in Taverna, Proc 2005 1st Int Conf e-Science and Grid Computing 156-162.
  Wolstencroft, 2013, The Taverna Workflow Suite: Designing and Executing Workflows of Web Services on the Desktop, Web or in the Cloud, Nucl Acids Res 41(W1):W556-W561.
  Wu, 2010, Fast and SNP-tolerant detection of complex variants and splicing in short reads, Bioinformatics, 26(7):873-881.
  Xing, 2006, An expectation-maximization algorithm for probabilistic reconstructions of full-length isoforms from splice graphs, Nucleic Acids Research, 34:3150-3160.
  Yang, 2013, Leveraging reads that span multiple single nucleotide polymorphisms for haplotype inference from sequencing data, Bioinformatics 29(18):2245-2252.
  Yang, 2014, Community detection in networks with node attributes, proc IEEE ICDM '13, arXiv:1401.7267.
  Yanovsky, 2008, Read mapping algorithms for single molecule sequencing data, Proc 8th Int Workshop Alg Bioinformatics 5251:38-49.
  Yildiz, 2014, BIFI: a Taverna plugin for a simplified and user-friendly workflow platform, BMC Res Notes 7:740.
  Yu, 2007, A tool for creating and parallelizing bioinformatics pipelines, DOD High Performance Computing Conf 417-420.
  Yu, 2010, The construction of a tetraploid cotton genome wide comprehensive reference map, Genomics 95:230-240.
  Zhang, 2013, Taverna Mobile: Taverna workflows on Android, EMBnet J 19(B):43-45.
  Zhao, 2012, Why Workflows Break-Understanding and Combating Decay in Taverna Workflows, eScience 2012, Chicago, Oct. 2012.
  Barbieri, 2013, Exome sequencing identifies recurrent SPOP, FOXA1 and MED12 mutations in prostate cancer, Nature Genetics 44:6 685-689.
  Beerenwinkel, 2007, Conjunctive Bayesian Networks, Bernoulli 13(4), 893-909.
  Browning et al, Haplotype phasing: existing methods and new developments, 2011, vol. 12, Nature Reviews Genetics.
  Caboche et al, Comparison of mapping algorithms used in high-throughput sequencing: application to Ion Torrent data, 2014, vol. 15, BMC Genomics.
  Cartwright, DNA assembly with gaps (DAWG): simulating sequence evolution, 2005, pp. iii31-iii38, vol. 21, Oxford University Press.
  Duan et al., Optimizing de novo common wheat transcriptome assembly using short-read RNA-Seq data. (2012) pp. 1-12, vol. 13, BMC Genomics.
  Gerlinger, 2012, Intratumor Heterogeneity and Branched Evolution Revealed by Multiregion Sequencing, 366:10 883-892.
  International Search Report and Written Opinion dated Aug. 31, 2017, for International Application No. PCT/US2017/018830 with International Filing Date Feb. 22, 2017, (11 pages).
  International Search Report and Written Opinion dated Apr. 19, 2017 for International Patent Application No. PCT/US2017/012015, (14 Pages).
  International Search Report and Written Opinion dated Jan. 10, 2017, for International Patent Application No. PCT/US16/57324 with International Filing Date Oct. 17, 2016, (7 pages).
  International Search Report and Written Opinion dated Apr. 7, 2017, for International Patent Application No. PCT/US17/13329, filed Jan. 13, 2017, (9 pages).
  International Search Report and Written Opinion dated Feb. 4, 2015, for International Patent Application No. PCT/US2014/061198, filed Oct. 17, 2014, (8 pages).
  International Search Report and Written Opinion dated Feb. 10, 2015, for International Patent Application No. PCT/US2014/060690, filed Oct. 15, 2014, PCT/US2014/060690 (11 pages).
  Lecca, 2015, Defining order and timing of mutations during cancer progression: the TO-DAG probabilistic graphical model, Frontiers in Genetics, vol. 6 Article 309 1-17.
  Lee et al. Accurate read mapping using a graph-based human pan-genome. (May 2015) American Society of Human Genetics 64th Annual Meeting Platform Abstracts; Abstract 41.
  Marth et al., 1999—A general approach to single-nucleotide polymorphism discovery, pp. 452-456, vol. 23, Nature Genetics.
  Mourad, 2012, A hierarchical Bayesian network approach for linkage disequilibrium modeling and data-dimensionality reduction prior to genome-wide association studies, BMC Bioinformatics 12:16 1-20.
  Myers, The Fragment Assembly String Graph, Bioinformatics, 2005, pp. ii79-ii85, vol. 21.
  Pop et al., 2004, Comparative genome assembly, Briefings in Bioinformatics vol. 5, pp. 237-248.
  Pruesse, 2012, SINA: Accurate high-throughput multiple sequence alignment of ribosomal RNA genes, Bioinformatics 28:14 1823-1829.
  Sturgeon, RCDA: a highly sensitive and specific alternatively spliced transcript assembly tool featuring upstream consecutive exon structures, Genomics, Dec. 2012, 100(6): 357-362.
  Uchiyama et al., CGAT: a comparative genome analysis tool for visualizing alignments in the analysis of complex evolutionary changes between closely related genomes, 2006, e-pp. 1-17, vol. 7:472; BMC Bioinformatics.
  Zeng, 2013, PyroHMMvar: a sensitive and accurate method to call short indels and SNPs for Ion Torrent and 454 data, Bioinformatics 29:22 2859-2868.
  Zhang et al., Construction of a high-density genetic map for sesame based on large scale marker development by specific length amplified fragment (SLAF) sequencing. (2013) pp. 1-12, vol. 13, BMC Plant Biology.
 
     Primary Examiner —Mary K Zeman
     Art Unit — 1631
     Exemplary claim number — 1
 
(74)Attorney, Agent, or Firm — Kip L. Bodi

(57)

Abstract

The invention includes methods and systems for identifying diseased-induced mutations by producing multi-dimensional reference sequence constructs that account for variations between individuals, different diseases, and different stages of those diseases. Once constructed, these reference sequence constructs can be used to align sequence reads corresponding to genetic samples from patients suspected of having a disease, or who have had the disease and are in suspected remission. The reference sequence constructs also provide insight to the genetic progression of the disease.
20 Claims, 12 Drawing Sheets, and 12 Figures


RELATED APPLICATIONS

[0001] This application claims priority to U.S. Patent Application No. 61/892,670, filed Oct. 18, 2013, which is incorporated by reference herein in its entirety.

FIELD OF THE INVENTION

[0002] The invention relates to methods and systems for identifying disease-induced mutations, such as caused by cancer. The invention additionally provides methods for identifying mutations that may be causative of advanced disease, such as metastatic cancers.

BACKGROUND

[0003] Many diseases result from inherited or random mutations in a patient's genetic sequence. Additionally, in diseases such as cancer, advanced stages of disease may manifest as new changes in the genetic sequence of diseased cells. Accordingly, there is increased interest in sequencing diseased cells, e.g., from a biopsy or freely circulating, to determine the type or stage of a disease. Thus, patients who have undergone treatment for a disease may have new biopsy samples sequenced to monitor disease recurrence and/or progression. Such monitoring allows for early intervention in the event of recurrence, and also avoids unnecessary treatment when changes are not detected.
[0004] While there are many diseases that can be typed and tracked with genetic screening, cancer mutation screening has received the most attention. In some instances, the type of cancer can be immediately identified because of one tell-tale mutation, such as BRCA1. In most instances, however, cancer typing involves discovering and analyzing several sequences from a patient. Because these samples originate from the same patient, the samples are not independent of each other, but rather, are interrelated both developmentally and structurally. Furthermore, in most instances, accurate typing of a tumor requires knowledge of three sequences: the subject's healthy sequence (as found in non-cancerous parts of the body), the sequence of a major cancerous clone, and the sequences of minor clones (which may often be metastatic).
[0005] The prospect of sequencing several samples to obtain a complete picture of a disease is less intimidating due to recent advances in genetic sequencing. Next-generation sequencing (e.g., whole-transcriptome shotgun sequencing, pyrosequencing, ion semiconductor sequencing, sequencing by synthesis) can generate millions of reads, covering an entire genome, in just a few days. To achieve this throughput, NGS sequencing uses massive parallelization on smaller nucleic acid sequences that together make up a larger body of genetic information, e.g., a chromosome or a genome. Starting from a genetic sample, the nucleic acids (e.g., DNA) are broken up, amplified, and read with extreme speed. Once the reads are produced, the reads are aligned to a reference genome, e.g., GRCh37, with a computer to produce longer, assembled sequences, known as contigs. Because the sequence data from next generation sequencers often comprises millions of shorter sequences that together represent the totality of the target sequence, aligning the reads is complex and computationally expensive. Additionally, in order to minimize sequence distortions caused by random sequencing errors (i.e., incorrect sequencing machine outputs), each portion of the probed sequence is sequenced multiple times (e.g., 2 to 100 times, or more) to minimize the influence of any random sequencing errors on the final alignments and output sequences generated.
[0006] Once all of the data, corresponding to all of the nucleic acid reads is collected, and the reads are aligned against the reference, the reads are assembled and compared to the reference, as well as to each other, to determine the relationship between the samples. The workflow for this analysis is shown pictorially in FIG. 1. Each assembled read is typically compared to the reference whereupon the variations between the assembled sequence and the reference are cataloged in a file, known as a variant file, which may be in one of several accepted formats. These variant files can then be compared to each other, in order to determine how different the genetic material varies between cells with varying stages of the disease. The variant files can also be the basis for later comparison with new samples from the patient to screen for recurrence or disease progression.
[0007] The workflow illustrated in FIG. 1 suffers from several drawbacks. Because there may be millions of genetic differences between the reference sequence and the sequences of the samples of the patient, it is often very difficult to pinpoint the key differences between the non-diseased and diseased tissues. In theory, this problem can be avoided by comparing the sequences of the diseased and non-diseased samples directly, however the use of the reference sequence for the original alignment “infects” the downstream analysis. Typically, certain portions of the patient's samples that did not align with the reference are treated as equivalent mutations in the variant files, even though they are not, in fact, equivalent. Furthermore, structural variations between the reference sequence and the patient's samples, and between the patient's samples themselves, result in variant files with differing indexes for the same (or similar) mutations. Especially in the instance of recurrence screening, the lack of a stable index makes it very difficult to identify new smaller mutations.
[0008] Typically a sequence alignment is constructed by aggregating pairwise alignments between two linear strings of sequence information, one of which is a standard reference. As an example of alignment, two strings, S1 (SEQ ID NO. 15: AGCTACGTACACTACC) and S2 (SEQ ID NO. 16: AGCTATCGTACTAGC) can be aligned against each other. S1 typically corresponds to a read and S2 correspond to a portion of the reference sequence. With respect to each other, S1 and S2 can consist of substitutions, deletions, and insertions. Typically, the terms are defined with regard to transforming string S1 into string S2: a substitution occurs when a letter or sequence in S2 is replaced by a different letter or sequence of the same length in S1, a deletion occurs when a letter or sequence in S2 is “skipped” in the corresponding section of S1, and an insertion occurs when a letter or sequence occurs in S1 between two positions that are adjacent in S2. For example, the two sequences S1 and S2 can be aligned as below. The alignment below represents thirteen matches, a deletion of length one, an insertion of length two, and one substitution:
[0009] 
[00001] [TABLE-US-00001]
    (S1)
    (SEQ ID NO. 15)
    AGCTA−CGTACACTACC
   
    (S2)
    (SEQ ID NO. 16)
    AGCTATCGTAC−−TAGC
[0010] One of skill in the art will appreciate that there are exact and approximate algorithms for sequence alignment. Exact algorithms will find the highest scoring alignment, but can be computationally expensive. The two most well-known exact algorithms are Needleman-Wunsch (J Mol Biol, 48(3):443-453, 1970) and Smith-Waterman (J Mol Biol, 147(1):195-197, 1981; Adv. in Math. 20(3), 367-387, 1976). A further improvement to Smith-Waterman by Gotoh (J Mol Biol, 162(3), 705-708, 1982) reduces the calculation time from O(m2n) to O(mn) where m and n are the sequence sizes being compared and is more amendable to parallel processing. In the field of bioinformatics, it is Gotoh's modified algorithm that is often referred to as the Smith-Waterman algorithm. Smith-Waterman approaches are being used to align larger sequence sets against larger reference sequences as parallel computing resources become more widely and cheaply available. See, e.g., Amazon.com's cloud computing resources available at http://aws.amazon.com. All of the above journal articles are incorporated herein by reference in their entireties.
[0011] The Smith-Waterman (SW) algorithm aligns linear sequences by rewarding overlap between bases in the sequences, and penalizing gaps between the sequences. Smith-Waterman also differs from Needleman-Wunsch, in that SW does not require the shorter sequence to span the string of letters describing the longer sequence. That is, SW does not assume that one sequence is a read of the entirety of the other sequence. Furthermore, because SW is not obligated to find an alignment that stretches across the entire length of the strings, a local alignment can begin and end anywhere within the two sequences.
[0012] The SW algorithm is easily expressed for an n×m matrix H, representing the two strings of length n and m, in terms of equation (1) below:
          Hk0=H0l=0(for 0≤k≤n and 0≤l≤m)
          Hij=max{Hi-1,j-1+s(ai,bj),Hi-1,j−Win,Hi,j-1−Wdel,0}(for 1≤i≤n and 1≤j≤m)  (1)
In the equations above, s(ai,bj) represents either a match bonus (when ai=bj) or a mismatch penalty (when ai≠bj), and insertions and deletions are given the penalties Win and Wdel, respectively. In most instance, the resulting matrix has many elements that are zero. This representation makes it easier to backtrace from high-to-low, right-to-left in the matrix, thus identifying the alignment.
[0013] Once the matrix has been fully populated with scores, the SW algorithm performs a backtrack to determine the alignment. Starting with the maximum value in the matrix, the algorithm will backtrack based on which of the three values (Hi-1,j-1, Hi-1,j, or Hi,j-1) was used to compute the final maximum value for each cell. The backtracking stops when a zero is reached. See, e.g., FIG. 4 part (B), which does not represent the prior art, but illustrates the concept of a backtrack, and the corresponding local alignment when the backtrack is read. Accordingly, the “best alignment,” as determined by the algorithm, may contain more than the minimum possible number of insertions and deletions, but will contain far less than the maximum possible number of substitutions.
[0014] When applied as SW or SW-Gotoh, the techniques use a dynamic programming algorithm to perform local sequence alignment of the two strings, S and A, of sizes m and n, respectively. This dynamic programming technique employs tables or matrices to preserve match scores and avoid recomputation for successive cells. Each element of the string can be indexed with respect to a letter of the sequence, that is, if S is the string ATCGAA, S[1]=A, S[4]=G, etc. Instead of representing the optimum alignment as Hi,j (above), the optimum alignment can be represented as B[j,k] in equation (2) below:
          B[j,k]=max(p[j,k],i[j,k],d[j,k],0)(for 0<j≤m,0<k≤n)  (2)
The arguments of the maximum function, B[j,k], are outlined in equations (3)-(5) below, wherein MISMATCH_PENALTY, MATCH_BONUS, INSERTION_PENALTY, DELETION_PENALTY, and OPENING_PENALTY are all constants, and all negative except for MATCH_BONUS. The match argument, p[j,k], is given by equation (3), below:
[0015]  [see pdf for image]
the insertion argument i[j,k], is given by equation (4), below:
          i[j,k]=max(p[j−1,k]+OPENING_PENALTY,i[j−1,k]d[j−1,k]+OPENING_PENALTY)+INSERTION_PENALTY  (4)
and the deletion argument d[j,k], is given by equation (5), below:
          d[j,k]=max(p[j,k−1]+OPENING_PENALTY,i[j,k−1]+OPENING_PENALTY,d[j,k−1])+DELETION_PENALTY  (5)
For all three arguments, the [0,0] element is set to zero to assure that the backtrack goes to completion, i.e., p[0,0]=i[0,0]=d[0,0]=0.
[0016] The scoring parameters are somewhat arbitrary, and can be adjusted to achieve the behavior of the computations. One example of the scoring parameter settings (Huang, Chapter 3: Bio-Sequence Comparison and Alignment, ser. Curr Top Comp Mol Biol. Cambridge, Mass.: The MIT Press, 2002) for DNA would be:
[0017] MATCH_BONUS: 10
[0018] MISMATCH_PENALTY: −20
[0019] INSERTION_PENALTY: −40
[0020] OPENING_PENALTY: −10
[0021] DELETION_PENALTY: −5
[0022] The relationship between the gap penalties (INSERTION_PENALTY, OPENING_PENALTY) above help limit the number of gap openings, i.e., favor grouping gaps together, by setting the gap insertion penalty higher than the gap opening cost. Of course, alternative relationships between MISMATCH_PENALTY, MATCH_BONUS, INSERTION_PENALTY, OPENING_PENALTY and DELETION_PENALTY are possible.
[0023] While the alignment methods, described above, have been useful for assembling reads produced with next-generation sequencing techniques, they are complex and time-consuming. Additionally, these techniques are ill-suited for identifying the important nuances between diseased cells of varying disease states because the uncertainty due to aligning the reads to a common reference often drowns out small changes in the genome.

SUMMARY

[0024] The invention provides improved methods and systems for identifying mutations that are induced by, or associated with, disease, especially cancer. The methods allow specific changes associated with advanced stages of the disease to be easily differentiated from lesser diseased cells, thus providing insight into the relationship between the size and location of the mutations and progression of the disease. This insight can be used to identify disease progression in other patients, and the relationship also provides for faster and more accurate typing of samples later collected from the same patient to monitor for disease progression or recurrence.
[0025] The invention employs multi-dimensional reference sequence constructs and alignment algorithms that allow new sequence samples to be simultaneously compared to multiple sequences relevant to the disease, thereby providing increased speed and accuracy in disease identification and typing. Furthermore, the reference sequence constructs of the invention accommodate structural variations, deletions, insertions, and polymorphisms between samples in a straightforward way, allowing a single construct to be assembled spanning an entire chromosome or the whole genome of a patient. Using a “look-back” type analysis, the described algorithms can also be used to align new reads in a multi-dimensional space including elements from sequences of various states of disease progression to provide more accurate alignment of sequence reads, while achieving lower error rates. Alternatively, constructs of the invention can be used to identify and/or study variations between individuals or cohorts having similar disease stages. In an embodiment, the invention is implemented by aligning sequence reads to a series of directed, acyclic sequences spanning branch points that account for sequence variation in the samples, including insertions, deletions, and substitutions. Such constructs, often represented as directed acyclic graphs (DAGs) can be assembled from existing variable sequence files or the constructs can be fabricated de novo using a standard reference as a starting point.
[0026] Once a sequence construct has been fabricated to account for sequence variation among differently diseased samples, those constructs can be used to identify disease risk in new samples from the same individual or, in some instances, from other individuals. In particular, because portions of the sequence construct can be tagged with secondary information, such as “metastatic,” the subsequent step of comparing a mutation, vis-à-vis the reference genome, to a table of known mutations can be eliminated. Thus, it is merely a matter of identifying a sample as being aligned to sequences in the construct that are indicative of the disease or stage. Alternatively, when a mutation is not known (i.e., not represented in the reference sequence construct), an alignment will be found, whereby the variant can be identified as a new mutation. Thus, using this iterative process, it is possible to compare and/or identify differences between major and minor cancer clones, or between pre-cancerous and cancerous samples.
[0027] The invention additionally includes systems for executing the methods of the invention. In one embodiment, a system comprises a distributed network of processors and storage capable of comparing a plurality of sequences (i.e., nucleic acid sequences, amino acid sequences) to a reference sequence construct (e.g., a DAG) representing observed variation in a genome or a region of a genome. The system is additionally capable of aligning the nucleic acid reads to produce a continuous sequence using an efficient alignment algorithm. Because the reference sequence construct compresses a great deal of redundant information, and because the alignment algorithm is so efficient, the reads can be tagged and assembled on an entire genome using commercially-available resources. The system comprises a plurality of processors that simultaneously execute a plurality of comparisons between a plurality of reads and the reference sequence construct. The comparison data may be accumulated and provided to a health care provider. Because the comparisons are computationally tractable, analyzing sequence reads will no longer represent a bottleneck between NGS sequencing and a meaningful discussion of a patient's genetic risks.

BRIEF DESCRIPTION OF THE FIGURES

[0028] FIG. 1 depicts the state-of-the-art methods for aligning sequence reads to a reference sequence, identifying variant files, and then comparing variant files to determine variations associated with disease types or disease progression;
[0029] FIGS. 2(A) and 2(B) depict the construction of a directed acyclic graph (DAG) representing genetic variation in a reference sequence. FIG. 2(A) shows the starting reference sequence and the addition of a deletion. FIG. 2(B) shows the addition of an insertion and a SNP, thus arriving at the Final DAG used for alignment;
[0030] FIG. 3 depicts three variant call format (VCF) entries represented as directed acyclic graphs;
[0031] FIG. 4 shows a pictorial representation of aligning a nucleic acid sequence read against a construct that accounts for an insertion event as well as the reference sequence. FIG. 4 also shows the matrices and the backtrack used to identify the proper location of the nucleic acid sequence read “ATCGAA”;
[0032] FIG. 5 depicts a workflow for producing a reference sequence construct that accounts for variations between the “normal” genome of a patient and an accepted reference in addition to variations between the “normal” genome of the patient and a genome of the patient corresponding to a diseased sample;
[0033] FIG. 6 illustrates a reference sequence construct based upon a reference sequence, a non-diseased sample that differs from the reference sequence by an insertion, and a diseased sample that differs from the non-diseased sample by a polymorphism associated with cancer;
[0034] FIG. 7 depicts a workflow for producing a reference sequence construct that accounts for variations between the “normal” genome of a patient, a major cancerous clone, and a minor cancerous clone. The minor cancerous clone may result in metastatic disease;
[0035] FIG. 8 depicts a workflow for producing a reference sequence construct that accounts for variations between the “normal” genome of a patient, a major cancerous clone, and several minor cancerous clones;
[0036] FIG. 9 depicts a workflow for producing a reference sequence construct that accounts for variations between the “normal” genome of a patient, a major cancerous clone, and several minor cancerous clones;
[0037] FIG. 10 depicts an associative computing model for parallel processing;
[0038] FIG. 11 depicts an architecture for parallel computation.

DETAILED DESCRIPTION

[0039] The invention includes methods and systems for identifying diseased-induced mutations by producing multi-dimensional reference sequence constructs that account for variations between individuals, different diseases, and different stages of those diseases. Once constructed, these reference sequence constructs can be used to align sequence reads corresponding to genetic samples from patients suspected of having a disease, or who have had the disease and are in suspected remission. The aligned sequences give immediate information about the nature of the samples, e.g., being of a metastatic nature. Thus, the reference sequence constructs can be used to monitor patients for recurrence or progression of a disease, such as cancer. The reference sequence constructs can also be used to study structural relationships between diseases and/or between disease states. The reference sequence constructs can be fabricated from previously determined variant files, or the reference sequence constructs can be created de novo, e.g., from samples originating from a patient.
[0040] In some embodiments, the reference sequence constructs are directed acyclic graphs (DAG), as described below, however the reference sequence can be any representation reflecting genetic variability in the sequences of different organisms within a species, provided the construct is formatted for alignment. The genetic variability represented in the construct may be between different tissues or cells within an individual. The genetic variability represented in the construct may be between different individuals or between different organisms. The genetic variability represented in the construct may be between similar tissues or cells that are at different stages of a disease.
[0041] In general, the reference sequence construct will comprise portions that are identical and portions that vary between sampled sequences. Accordingly, the constructs can be thought of as having positions (i.e., according to some canonical ordering) that comprise the same sequence(s) and some positions that comprise alternative sequences, reflecting genetic variability. The application additionally discloses methods for identifying a disease or a disease stage based upon alignment of a nucleic acid read to a location in the construct. The methods are broadly applicable to the fields of genetic sequencing and mutation screening.

Reference Sequence Constructs

[0042] Unlike prior art sequence alignment methods that use a single reference sequence to align and genotype nucleic acid reads, the invention uses a construct that can account for the variability in genetic sequences within a species, population, or even among different cells in a single organism. Representations of the genetic variation can be presented as directed acyclic graphs (DAGs) (discussed above) or row-column alignment matrices, and these constructs can be used with the alignment methods of the invention provided that the parameters of the alignment algorithms are set properly (discussed below).
[0043] In preferred embodiments of the invention, the construct is a directed acyclic graph (DAG), i.e., having a direction and having no cyclic paths. (That is, a sequence path cannot travel through a position on the reference construct more than once.) In the DAG, genetic variation in a sequence is represented as alternate nodes. The nodes can be a section of conserved sequence or a gene, or simply a nucleic acid. The different possible paths through the construct represent known genetic variation. A DAG may be constructed for an entire genome of an organism, or the DAG may be constructed only for a portion of the genome, e.g., a chromosome, or smaller segment of genetic information. In some embodiments, the DAG represents greater than 1000 nucleic acids, e.g., greater than 10,000 nucleic acids, e.g., greater than 100,000 nucleic acids, e.g., greater than 1,000,000 nucleic acids. A DAG may represent a species (e.g., homo sapiens) or a selected population (e.g., women having breast cancer), or even smaller subpopulations, such as genetic variation among different tumor cells in the same individual.
[0044] A simple example of DAG construction is shown in FIGS. 2(A) and 2(B). As shown in FIG. 2(A), the DAG begins with a reference sequence, shown in FIG. 2(A) as SEQ ID NO. 1: CATAGTACCTAGGTCTTGGAGCTAGTC. In practice, the reference sequence is often much longer, and may be an entire genome. In some embodiments, the sequence is a FASTA or FASTQ file. (FASTQ has become the default format for sequence data produced from next generation sequencers). In some embodiments, the reference sequence may be a standard reference, such as GRCh37. In some embodiment, the reference sequence is a sequence from non-diseased cells of a patient. As recognized by those of skill, each letter (or symbol) in the sequence actually corresponds to a nucleotide (e.g., a deoxyribonucleotide or a ribonucleotide) or an amino acid (e.g., histidine, leucine, lysine, etc.).
[0045] At the next step, a variant is added to the reference sequence, as shown in the bottom image of FIG. 1(A). As shown in FIG. 1(A) the variant is the deletion of the sequence “AG” from the reference between the lines in the figure, i.e., SEQ ID NO. 2. Graphically, this deletion is represented by breaking the reference sequence into nodes before and after the deletion, and connecting the nodes with an edge and also creating a path from one node to the “AG” and then to the other node. Thus, one path between the nodes represents the reference sequence, while the other path represents the deletion.
[0046] In practice, the variants are called to the DAG by applying the entries in a variant call format (VCF) file, such as can be found at the 1000 Genomes Project website. Because each VCF file is keyed to a specific reference genome, it is not difficult to identify where the strings should be located. In fact, each entry in a VCF file can be thought of as combining with the reference to create separate graph, as displayed in FIG. 3. Note the VCF entries in FIGS. 2(A) and 2(B) do not correspond to the VCF entries of FIG. 3. It is also possible to identify variants for inclusion into the DAG by comparing the sequences of non-diseased and diseased cells of an individual.
[0047] Moving to FIG. 2(B), a second VCF entry, corresponding to an insertion “GG” at a specific position is added to produce an expanded DAG, i.e., including SEQ ID NO. 3 and SEQ ID NO. 4. Next, a third VCF entry can be added to the expanded DAG to account for a SNP earlier in the reference sequence, i.e., including SEQ ID NOS. 5-8. Thus, in three steps, a DAG has been created against which nucleic acid reads can be aligned (as discussed below.)
[0048] In practice, the DAGs are represented in computer memory (hard disk, FLASH, cloud memory, etc.) as a set of nodes, S, wherein each node is defined by a string, a set of parent nodes, and a position. The string is the node's “content,” i.e., sequence; the parent nodes define the node's position with respect to the other nodes in the graph; and the position of the node is relative to some canonical ordering in the system, e.g., the reference genome. While it is not strictly necessary to define the graph with respect to a reference sequence, it does make manipulation of the output data simpler. Of course, a further constraint on S is that it cannot include loops.
[0049] In many embodiments, the nodes comprise a plurality of characters, as shown in FIGS. 2(A) and 2(B), however it is possible that a node may be a single character, e.g., representing a single base, as shown in FIG. 3. In instances where a node represents a string of characters, all of the characters in the node can be aligned with a single comparison step, rather than character-by-character calculations, as is done with conventional Smith-Waterman techniques. As a result, the computational burden is greatly reduced as compared to state-of-the-are methods. The reduced computational burden allows the alignment to be completed quicker, and with fewer resources. When used in next generation sequencing, where millions of small reads need to be aligned and assembled, this reduction in computational burden has tangible benefits in terms of reducing the cost of the alignment, while making meaningful information, i.e., genotype, available more quickly. In instances where a treatment will be tailored to a patient's genotype, the increased speed may allow a patient to begin treatment days earlier than using state-of-the-art methods.
[0050] Extrapolating this DAG method to larger structures, it is possible to construct DAGs that incorporate thousands of VCF entries representing the known variation in genetic sequences for a given region of a reference. Nonetheless, as a DAG becomes bulkier, the computations do take longer, and for many applications a smaller DAG is used that may only represent a portion of the sequence, e.g., a chromosome. In other embodiments, a DAG may be made smaller by reducing the size of the population that is covered by the DAG, for instance going from a DAG representing variation in breast cancer to a DAG representing variation in triple negative breast cancer. Alternatively, longer DAGs can be used that are customized based upon easily identified genetic markers that will typically result in a large portion of the DAG being consistent between samples. For example, aligning a set of nucleic acid reads from an African-ancestry female will be quicker against a DAG created with VCF entries from women of African ancestry as compared to a DAG accounting for all variations known in humans over the same sequence. It is to be recognized that the DAGs of the invention are dynamic constructs in that they can be modified over time to incorporate newly identified mutations. Additionally, algorithms in which the alignment results are recursively added to the DAG are also possible.
[0051] In the instance of string-to-DAG alignment, the gap penalties can be adjusted to make gap insertions even more costly, thus favoring an alignment to a sequence rather than opening a new gap in the overall sequence. Of course, with improvements in the DAG (discussed above) the incidence of gaps should decrease even further because mutations are accounted for in the DAG.

Alignment Algorithm

[0052] In one embodiment, an algorithm is used to align sequence reads against a directed acyclic graph (DAG). In contrast to the algorithm expressed in the Background, the alignment algorithm identifies the maximum value for Ci,j by identifying the maximum score with respect to each sequence contained at a position on the DAG (e.g., the reference sequence construct). In fact, by looking “backwards” at the preceding positions, it is possible to identify the optimum alignment across a plurality of possible paths.
[0053] The algorithm of the invention is carried out on a read (a.k.a. “string”) and a directed acyclic graph (DAG), discussed above. For the purpose of defining the algorithm, let S be the string being aligned, and let D be the directed acyclic graph to which S is being aligned. The elements of the string, S, are bracketed with indices beginning at 1. Thus, if S is the string ATCGAA, S[1]=A, S[4]=G, etc.
[0054] For the DAG, each letter of the sequence of a node will be represented as a separate element, d. A predecessor of d is defined as:
[0055] (i) If d is not the first letter of the sequence of its node, the letter preceding d in its node is its (only) predecessor;
[0056] (ii) If d is the first letter of the sequence of its node, the last letter of the sequence of any node that is a parent of d's node is a predecessor of d.
[0057] The set of all predecessors is, in turn, represented as P[d].
[0058] In order to find the “best” alignment, the algorithm seeks the value of M[j,d], the score of the optimal alignment of the first j elements of S with the portion of the DAG preceding (and including) d. This step is similar to finding Hi,j in equation 1 in the Background section. Specifically, determining M[j,d] involves finding the maximum of a, i, e, and 0, as defined below:
          M[j,d]=max{a,i,e,0}  (6)
[0059] where
[0060] e=max{M[j,p*]+DELETE_PENALTY} for p* in P[d]
[0061] i=M[j−1,d]+INSERT_PENALTY
[0062] a=max{M[j−1,p*]+MATCH_SCORE} for p* in P[d], if S[j]=d; max{M[j−1,p*]+MISMATCH_PENALTY} for p* in P[d], if S[j]≠d
[0063] As described above, e is the highest of the alignments of the first j characters of S with the portions of the DAG up to, but not including, d, plus an additional DELETE_PENALTY. Accordingly, if d is not the first letter of the sequence of the node, then there is only one predecessor, p, and the alignment score of the first j characters of S with the DAG (up-to-and-including p) is equivalent to M[j,p]+DELETE_PENALTY. In the instance where d is the first letter of the sequence of its node, there can be multiple possible predecessors, and because the DELETE_PENALTY is constant, maximizing [M[j,p*]+DELETE_PENALTY] is the same as choosing the predecessor with the highest alignment score with the first j characters of S.
[0064] In equation (6), i is the alignment of the first j−1 characters of the string S with the DAG up-to-and-including d, plus an INSERT_PENALTY, which is similar to the definition of the insertion argument in SW (see equation 1).
[0065] Additionally, a is the highest of the alignments of the first j characters of S with the portions of the DAG up to, but not including d, plus either a MATCH_SCORE (if the jth character of S is the same as the character d) or a MISMATCH_PENALTY (if the jth character of S is not the same as the character d). As with e, this means that if d is not the first letter of the sequence of its node, then there is only one predecessor, i.e., p. That means a is the alignment score of the first j−1 characters of S with the DAG (up-to-and-including p), i.e., M[j−1,p], with either a MISMATCH_PENALTY or MATCH_SCORE added, depending upon whether d and the jth character of S match. In the instance where d is the first letter of the sequence of its node, there can be multiple possible predecessors. In this case, maximizing {M[j,p*]+MISMATCH_PENALTY or MATCH_SCORE} is the same as choosing the predecessor with the highest alignment score with the first j−1 characters of S (i.e., the highest of the candidate M[j−1,p*] arguments) and adding either a MISMATCH_PENALTY or a MATCH_SCORE depending on whether d and the jth character of S match.
[0066] Again, as in the SW algorithm discussed in the Background, the penalties, e.g., DELETE_PENALTY, INSERT_PENALTY, MATCH_SCORE and MISMATCH_PENALTY, can be adjusted to encourage alignment with fewer gaps, etc.
[0067] As described in the equations above, the algorithm finds the maximum value for each read by calculating not only the insertion, deletion, and match scores for that element, but looking backward (against the direction of the DAG) to any prior nodes on the DAG to find a maximum score. Thus, the algorithm is able to traverse the different paths through the DAG, which contain the known mutations. Because the graphs are directed, the backtracks, which move against the direction of the graph, follow the preferred variant sequence toward the origin of the graph, and the maximum alignment score identifies the most likely alignment within a high degree of certainty. While the equations above are represented as “maximum” values, “maximum” is intended to cover any form of optimization, including, for example, switching the signs on all of the equations and solving for a minimum value.
[0068] Implementation of the disclosed algorithm is exemplified in FIG. 4, where a sequence “ATCGAA” is aligned against a DAG that represents a reference sequence SEQ ID NO. 10: TTGGATATGGG and a known insertion event SEQ ID NO. 11: TTGGATCGAATTATGGG, where the insertion is underlined. FIG. 4 part (A) shows a pictorial representation of the read being compared to the DAG while FIG. 4 part (B) shows the actual matrices that correspond to the comparison. Like the Smith-Waterman technique discussed in the Background, the algorithm of the invention identifies the highest score and performs a backtrack to identify the proper location of the read. FIG. 4 also highlights that the invention produces an actual match for the string against the construct, whereas the known methods (e.g., SW) would have been more likely to align the string to the wrong part of the reference, or reject the string as not generating a sufficiently-high alignment score to be included in the alignment. In the instances where the sequence reads include variants that were not included in the DAG, the aligned sequence will be reported out with a gap, insertion, etc.

Fabricating Constructs to Accommodate Disease Variability

[0069] As mentioned above, the reference sequence construct can be prepared from existing variant files, or the construct can be prepared de novo, by comparing certain sampled sequences to a reference sequence. An example of such de novo construction is shown in FIG. 5. Starting with a reference genome, e.g., GRCh37, a non-cancerous sample is sequenced and compared to the reference to produce a file of variants. This file of variants is incorporated into a reference sequence construct, e.g., a DAG, as described above. The variants may include insertions, deletions, polymorphisms, structural variants, etc. The resulting construct can then be used to align reads from a diseased sample from the individual. This alignment step will provide immediate information about the location of “new” mutations that are likely correlated with a diseased state (i.e., the major clone), immediately differentiating them from mutations that already existed in the non-cancerous sample (because the latter are already included in the construct). Moreover, because the major clone sample is not aligned directly to the reference, but rather to a reference sequence construct, most of the major clone sample should align completely to the construct, and any unaligned reads will give immediate clues about the nature of the disease.
[0070] Taking the example to the next level, the variants between the non-diseased and major clone sample can be incorporated into a new reference construct, “Reference with major clone” as shown in FIG. 5. Such a reference construct is shown in greater detail in FIG. 6. FIG. 6 uses three arbitrary sequences (SEQ ID NOS. 12-14) to illustrate the construction of the reference sequence and the incorporation of a reference, a non-diseased sequence, and a major clone. SEQ ID NO. 12 represents a portion of the reference sequence, which has been determined to be a good starting point. The individual in question, however, has a 15 bp insert that is not present in the reference sequence (SEQ ID NO. 13). Additionally, the individual, who has been identified as having cancer, has a sample of a tumor sequenced. Upon sequencing the tumor sample is found to have a polymorphism within the insert (SEQ ID NO. 14). As shown in the reference sequence construct at the bottom of FIG. 6, it is possible to account for all three sequences in this construct.
[0071] This construct in FIG. 6 is useful for at least two purposes. First, without any additional analysis, the reference sequence construct shows the relationship between the polymorphism and the major clone, separately from the insert, which arguably is not linked to the cancer. It should also be noted that the presence of a polymorphism within the insert would likely have been missed if SEQ ID NO. 14 was directly compared to SEQ ID NO. 12, as is typically done with contemporary cancer sequencing. Second, the reference sequence construct provides a new alignment tool against which new samples can be compared. For example a read (not shown) that aligns partially to the insert and includes the polymorphism is likely cancerous, whereas a read (not shown) that aligns partially to the insert, but contains “AC” instead of “GG” in the region of the polymorphism is likely not cancerous. Additionally, a new read that contained the polymorphism in addition to other mutations that did not align to the reference sequence construct suggests that the cancer may have progressed further and that greater testing, e.g., whole body MRI, may be in order.
[0072] The reference sequence construct shown in FIG. 6, may be improved iteratively, with the addition of more sequences, e.g., corresponding to identified minor clones, as shown in FIG. 7. In some cases the minor clones may represent metastatic cells or other progressed forms of the disease. Using the alignment algorithms described above, sequence reads corresponding to the minor clones can be aligned to the “Reference with the major clone” making clear where the mutations are located. Like the incorporation of the major clone mutation in FIG. 6, the minor clone mutations can also be incorporated into the reference sequence construct to create yet another new reference sequence construct “Reference with minor clone.” Like “Reference with the major clone,” “Reference with minor clone” can then provide information about the evolution of disease in the individual and be used to type new samples from the individual (or other individuals, as appropriate).
[0073] The process of aligning new samples to the reference sequence construct and then subsequently adding the newly-identified variants to the construct to create a new construct can be repeated indefinitely. In reality, because the reference sequence constructs are multivariate constructions stored on a non-transitory computer readable medium, the addition of new structures is trivial. Furthermore, the alignment of new reads to these highly complex reference constructs is computationally feasible and far less taxing than comparing new reads to each previous sequence individually. The process of aligning and improving on the constructs can be done in parallel, as shown in FIG. 8, or in series, as shown in FIG. 9. Whether the method is completed in parallel or in series, the resultant reference sequence construct should be identical. Nonetheless differing information about the evolution of the disease can be gleaned from the process depending upon the order in which new elements are added.

Opportunities for Parallelization

[0074] The sequential version of the Smith-Waterman-Gotoh algorithm has been adapted and significantly modified for massive parallelization. For example, an ASC model, called Smith-Waterman using Associative Massive Parallelism (SWAMP) is described in U.S. Patent Publication No. 2012/0239706, incorporated herein by reference in its entirety. Part of the parallelization for SWAMP (and other parallel processing systems) stems from the fact that the values along any anti-diagonal are independent of each other. Thus, all of the cells along a given anti-diagonal can be done in parallel to distribute the computational resources. The data dependencies shown in the above recursive equations limit the level of achievable parallelism but using a wavefront approach will still speed up this useful algorithm. A wavefront approach implemented by Wozniak (Comput Appl in the Biosciences (CABIOS), 13(2):145-150, 1997) on the Sun Ultra SPARC uses specialized SIMD-like video instructions. Wozniak used the SIMD registers to store the values parallel to the minor diagonal, reporting a two-fold speedup over a traditional implementation on the same machine. Following Wozniak's example, a similar way to parallelize code is to use the Streaming SIMD Extension (SSE) set for the x86 architecture. Designed by Intel, the vector-like operations complete a single operation/instruction on a small number of values (usually four, eight or sixteen) at a time. Many AMD and Intel chips support the various versions of SSE, and Intel has continued developing this technology with the Advanced Vector Extensions (AVX) for their modern chipsets.
[0075] In other implementations, Rognes and Seeberg (Bioinformatics (Oxford, England), 16(8):699-706, 2000) use the Intel Pentium processor with SSE's predecessor, MMX SIMD instructions for their implementation. The approach that developed out of the work of Rognes and Seeberg (Bioinformatics, 16(8):699-706, 2000) for ParAlign does not use the wavefront approach (Rognes, Nuc Acids Res, 29(7):1647-52, 2001; Saebo et al., Nuc Acids Res, 33(suppl 2):W535-W539, 2005). Instead, they align the SIMD registers parallel to the query sequence, computing eight values at a time, using a pre-computed query-specific score matrix. Additional details of this method can be found in U.S. Pat. No. 7,917,302, incorporated by reference herein. The way Rognes and Seeberg layout the SIMD registers, the north neighbor dependency could remove up to one third of the potential speedup gained from the SSE parallel “vector” calculations. To overcome this, they incorporate SWAT-like optimizations. With large affine gap penalties, the northern neighbor will be zero most of the time. If this is true, the program can skip computing the value of the north neighbor, referred to as the “lazy F evaluation” by Farrar (Bioinformatics, 23(2):156-161, 2007). Rognes and Seeberg are able to reduce the number of calculations of Equation 1 to speed up their algorithm by skipping it when it is below a certain threshold. A six-fold speedup was reported in (Rognes and Seeberg, Bioinformatics, 16(8):699-706, 2000) using 8-way vectors via the MMX/SSE instructions and the SWAT-like extensions.
[0076] In the SSE work done by Farrar (Bioinformatics, 23(2):156-161, 2007), a striped or strided pattern of access is used to line up the SIMD registers parallel to the query registers. Doing so avoids any overlapping dependencies. Again incorporating the SWAT-like optimizations (Farrar, Bioinformatics 23(2):156-161, 2007) achieves a 2-8 time speedup over Wozniak (CABIOS 13(2):145-150, 1997) and Rognes and Seeberg (Bioinformatics (Oxford, England), 16(8):699-706, 2000) SIMD implementations. The block substitution matrices and efficient and clever inner loop with the northern (F) conditional moved outside of that inner loop are important optimizations. The strided memory pattern access of the sixteen, 8-bit elements for processing improves the memory access time as well, contributing to the overall speedup.
[0077] Farrar (Sequence Analysis, 2008) extended his work for a Cell Processor manufactured by Sony, Toshiba and IBM. This Cell Processor has one main core and eight minor cores. The Cell Broadband Engine was the development platform for several more Smith-Waterman implementations including SWPS3 by Szalkowski, et. al (BMC Res Notes 1(107), 2008) and CBESW by Wirawan, et. al (BMC Bioinformatics 9 (377) 2008) both using Farrar's striping approach. Rudnicki, et. al. (Fund Inform. 96, 181-194, 2009) used the PS3 to develop a method that used parallelization over multiple databases sequences.
[0078] Rognes (BMC Bioinformatics 12 (221), 2011) also developed a multi-threaded approach called SWIPE that processes multiple database sequences in parallel. The focus was to use a SIMD approach on “ordinary CPUs.” This investigation using coarse-grained parallelism split the work using multiple database sequences in parallel is similar to the graphics processor units (GPU)-based tools described in the CUDASW by Liu, et al. (BMC Res Notes 2(73), 2009) and Ligowski and Rudnicki (Eight Annual International Workshop on High Performance Computational Biology, Rome, 2009). There have been other implementations of GPU work with CUDASW++2.0 by Liu, et. al. (BMC Res Notes 3(93), 2010) and Ligowski, et. al (GPU Computing Gems, Emerald Edition, Morgan Kaufmann, 155-157, 2011).
[0079] In other variations, small-scale vector parallelization (8, 16 or 32-way parallelism) can be used to make the calculations accessible via GPU implementations that align multiple sequences in parallel. The theoretical peak speedup for the calculations is a factor of m, which is optimal. A 96-fold speedup for the ClearSpeed implementation using 96 processing elements, confirming the theoretical speedup.

Parallel Computing Models

[0080] The main parallel model used to develop and extend Smith-Waterman sequence alignment is the ASsociative Computing (ASC) (Potter et al., Computer, 27(11):19-25, 1994). Efficient parallel versions of the Smith-Waterman algorithm are described herein. This model and one other model are described in detail in this section.
[0081] Some relevant vocabulary is defined here. Two terms of interest from Flynn's Taxonomy of computer architectures are MIMD and SIMD, two different models of parallel computing. A cluster of computers, classified as a multiple-instruction, multiple-data (MIMD) model is used as a proof-of-concept to overcome memory limitations in extremely large-scale alignments. Section 8 describes usage of the MIMD model. An extended data-parallel, single-instruction multiple-data (SIMD) model known as ASC is also described.
[0082] Multiple Instruction, Multiple Data (MIMD)
[0083] The multiple-data, multiple-instruction model or MIMD model describes the majority of parallel systems currently available, and include the currently popular cluster of computers. The MIMD processors have a full-fledged central processing unit (CPU), each with its own local memory (Quinn, Parallel Computing: Theory and Practice, 2nd ed., New York: McGraw-Hill, 1994). In contrast to the SIMD model, each of the MIMD processors stores and executes its own program asynchronously. The MIMD processors are connected via a network that allows them to communicate but the network used can vary widely, ranging from an Ethernet, Myrinet, and InfiniBand connection between machines (cluster nodes). The communications tend to employ a much looser communications structure than SIMDs, going outside of a single unit. The data is moved along the network asynchronously by individual processors under the control of their individual program they are executing. Typically, communication is handled by one of several different parallel languages that support message-passing. A very common library for this is known as the Message Passing Interface (MPI). Communication in a “SIMD-like” fashion is possible, but the data movements will be asynchronous. Parallel computations by MIMDs usually require extensive communication and frequent synchronizations unless the various tasks being executed by the processors are highly independent (i.e. the so-called “embarrassingly parallel” or “pleasingly parallel” problems). The work presented in Section 8 uses an AMD Opteron cluster connected via InfiniBand.
[0084] Unlike SIMDs, the worst-case time required for the message-passing is difficult or impossible to predict. Typically, the message-passing execution time for MIMD software is determined using the average case estimates, which are often determined by trial, rather than by a worst case theoretical evaluation, which is typical for SIMDs. Since the worst case for MIMD software is often very bad and rarely occurs, average case estimates are much more useful. As a result, the communication time required for a MIMD on a particular problem can be and is usually significantly higher than for a SIMD. This leads to the important goal in MIMD programming (especially when message-passing is used) to minimize the number of inter-processor communications required and to maximize the amount of time between processor communications. This is true even at a single card acceleration level, such as using graphics processors or GPUs.
[0085] Data-parallel programming is also an important technique for MIMD programming, but here all the tasks perform the same operation on different data and are only synchronized at various critical points. The majority of algorithms for MIMD systems are written in the Single-Program, Multiple-Data (SPMD) programming paradigm. Each processor has its own copy of the same program, executing the sections of the code specific to that processor or core on its local data. The popularity of the SPMD paradigm stems from the fact that it is quite difficult to write a large number of different programs that will be executed concurrently across different processors and still be able to cooperate on solving a single problem. Another approach used for memory-intensive but not compute-intensive problems is to create a virtual memory server, as is done with JumboMem, using the work presented in Section 8. This uses MPI in its underlying implementation.
[0086] Single Instruction, Multiple Data (SIMD)
[0087] The SIMD model consists of multiple, simple arithmetic processing elements called PEs. Each PE has its own local memory that it can fetch and store from, but it does not have the ability to compile or execute a program. As used herein, the term “parallel memory” refers to the local memories, collectively, in a computing system. For example, a parallel memory can be the collective of local memories in a SIMD computer system (e.g., the local memories of PEs), the collective of local memories of the processors in a MIMD computer system (e.g., the local memories of the central processing units) and the like. The compilation and execution of programs are handled by a processor called a control unit (or front end) (Quinn, Parallel Computing: Theory and Practice, 2nd ed., New York: McGraw-Hill, 1994). The control unit is connected to all PEs, usually by a bus.
[0088] All active PEs execute the program instructions received from the control unit synchronously in lockstep. “In any time unit, a single operation is in the same state of execution on multiple processing units, each manipulating different data” (Quinn, Parallel Computing: Theory and Practice, 2nd ed., New York: McGraw-Hill, 1994), at page 79. While the same instruction is executed at the same time in parallel by all active PEs, some PEs may be allowed to skip any particular instruction (Baker, SIMD and MASC: Course notes from CS 6/73301: Parallel and Distributed Computing—power point slides, (2004)2004). This is usually accomplished using an “if-else” branch structure where some of the PEs execute the if instructions and the remaining PEs execute the else part. This model is ideal for problems that are “data-parallel” in nature that have at most a small number of if-else branching structures that can occur simultaneously, such as image processing and matrix operations.
[0089] Data can be broadcast to all active PEs by the control unit and the control unit can also obtain data values from a particular PE using the connection (usually a bus) between the control unit and the PEs. Additionally, the set of PE are connected by an interconnection network, such as a linear array, 2-D mesh, or hypercube that provides parallel data movement between the PEs. Data is moved through this network in synchronous parallel fashion by the PEs, which execute the instructions including data movement, in lockstep. It is the control unit that broadcasts the instructions to the PEs. In particular, the SIMD network does not use the message-passing paradigm used by most parallel computers today. An important advantage of this is that SIMD network communication is extremely efficient and the maximum time required for the communication can be determined by the worst-case time of the algorithm controlling that particular communication.
[0090] The remainder of this section is devoted to describing the extended SIMD ASC model. ASC is at the center of the algorithm design and development for this discussion.
[0091] Associative Computing Model
[0092] The ASsocative Computing (ASC) model is an extended SIMD based on the STARAN associative SIMD computer, designed by Dr. Kenneth Batcher at Goodyear Aerospace and its heavily Navy-utilized successor, the ASPRO.
[0093] Developed within the Department of Computer Science at Kent State University, ASC is an algorithmic model for associative computing (Potter et al., Computer, 27(11):19-25, 1994) (Potter, Associative Computing: A Programming Paradigm for Massively Parallel Computers, Plenum Publishing, 1992). The ASC model grew out of work on the STARAN and MPP, associative processors built by Goodyear Aerospace. Although it is not currently supported in hardware, current research efforts are being made to both efficiently simulate and design a computer for this model.
[0094] As an extended SIMD model, ASC uses synchronous data-parallel programming, avoiding both multi-tasking and asynchronous point-to-point communication routing. Multi-tasking is unnecessary since only one task is executed at any time, with multiple instances of this task executed in lockstep on all active processing elements (PEs). ASC, like SIMD programmers, avoid problems involving load balancing, synchronization, and dynamic task scheduling, issues that must be explicitly handled in MPI and other MIMD cluster paradigms.
[0095] FIG. 10 shows a conceptual model of an ASC computer. There is a single control unit, also known as an instruction stream (IS), and multiple processing elements (PEs), each with its own local memory. The control unit and PE array are connected through a broadcast/reduction network and the PEs are connected together through a PE data interconnection network.
[0096] As seen in FIG. 10, a PE has access to data located in its own local memory. The data remains in place and responding (active) PEs process their local data in parallel. The reference to the word associative is related to the use of searching to locate data by content rather than memory addresses. The ASC model does not employ associative memory, instead it is an associative processor where the general cycle is to search-process-retrieve. An overview of the model is available in (Potter et al., Computer, 27(11):19-25, 1994).
[0097] The tabular nature of the algorithm lends itself to computation using ASC due to the natural tabular structure of ASC data structures. Highly efficient communication across the PE interconnection network for the lockstep shifting of data of the north and northwest neighbors, and the fast constant time associative functions for searching and for maximums across the parallel computations are well utilized by SWAMP
[0098] The associative operations are executed in constant time (Jin et al., 15th International Parallel and Distributed Processing Symposium (IPDPS'01) Workshops, San Francisco, p. 193, 2001), due to additional hardware required by the ASC model. These operations can be performed efficiently (but less rapidly) by any SIMD-like machine, and has been successfully adapted to run efficiently on several SIMD hardware platforms (Yuan et al., Parallel and Distributed Computing Systems (PDCS), Cambridge, Mass., 2009; Trahan et al., J. of Parallel and Distributed Computing (JPDC), 2009). SWAMP and other ASC algorithms can therefore be efficiently implemented on other systems that are closely related to SIMDs including vector machines, which is why the model is used as a paradigm.
[0099] The control unit fetches and decodes program instructions and broadcasts control signals to the PEs. The PEs, under the direction of the control unit, execute these instructions using their own local data. All PEs execute instructions in a lockstep manner, with an implicit synchronization between instructions. ASC has several relevant high-speed global operations: associative search, maximum/minimum search, and responder selection/detection. These are described in the following section.
[0100] Associative Functions
[0101] The functions relevant to the SWAMP algorithms are discussed below. Associative Search
[0102] The basic operation in an ASC algorithm is the associative search. An associative search simultaneously locates the PEs whose local data matches a given search key. Those PEs that have matching data are called responders and those with non-matching data are called non-responders. After performing a search, the algorithm can then restrict further processing to only affect the responders by disabling the non-responders (or vice versa). Performing additional searches may further refine the set of responders. Associative search is heavily utilized by SWAMP+ in selecting which PEs are active within a parallel act within a diagonal.
[0103] Maximum/Minimum Search
[0104] In addition to simple searches, where each PE compares its local data against a search key using a standard comparison operator (equal, less than, etc.), an associative computer can also perform global searches, where data from the entire PE array is combined together to determine the set of responders. The most common type of global search is the maximum/minimum search, where the responders are those PEs whose data is the maximum or minimum value across the entire PE array. The maximum value is used by SWAMP+ in every diagonal it processes to track the highest value calculated so far. Use of the maximum search occurs frequently, once in a logical parallel act, m+n times per alignment.
[0105] Responder Selection/Detection
[0106] An associative search can result in multiple responders and an associative algorithm can process those responders in one of three different modes: parallel, sequential, or single selection. Parallel responder processing performs the same set of operations on each responder simultaneously. Sequential responder processing selects each responder individually, allowing a different set of operations for each responder. Single responder selection (also known as pickOne) selects one, arbitrarily chosen, responder to undergo processing. In addition to multiple responders, it is also possible for an associative search to result in no responders. To handle this case, the ASC model can detect whether there were any responders to a search and perform a separate set of actions in that case (known as anyResponders). In SWAMP, multiple responders that contain characters to be aligned are selected and processed in parallel, based on the associative searches mentioned above. Single responder selection occurs if and when there are multiple values that have the exact same maximum value when using the maximum/minimum search.
[0107] PE Interconnection Network
[0108] Most associative processors include some type of PE interconnection network to allow parallel data movement within the array. The ASC model itself does not specify any particular interconnection network and, in fact, many useful associative algorithms do not require one. Typically associative processors implement simple networks such as 1D linear arrays or 2D meshes. These networks are simple to implement and allow data to be transferred quickly in a synchronous manner. The 1D linear array is sufficient for the explicit communication between PEs in the SWAMP algorithms, for example.

Parallel Computing Systems

[0109] A generalized parallel processing architecture is shown in FIG. 11. While each component is shown as having a direct connection, it is to be understood that the various elements may be geographically separated but connected via a network, e.g., the internet. While hybrid configurations are possible, the main memory in a parallel computer is typically either shared between all processing elements in a single address space, or distributed, i.e., each processing element has its own local address space. (Distributed memory refers to the fact that the memory is logically distributed, but often implies that it is physically distributed as well.) Distributed shared memory and memory virtualization combine the two approaches, where the processing element has its own local memory and access to the memory on non-local processors. Accesses to local memory are typically faster than accesses to non-local memory.
[0110] Computer architectures in which each element of main memory can be accessed with equal latency and bandwidth are known as Uniform Memory Access (UMA) systems. Typically, that can be achieved only by a shared memory system, in which the memory is not physically distributed. A system that does not have this property is known as a Non-Uniform Memory Access (NUMA) architecture. Distributed memory systems have non-uniform memory access.
[0111] Processor—processor and processor—memory communication can be implemented in hardware in several ways, including via shared (either multiported or multiplexed) memory, a crossbar switch, a shared bus or an interconnect network of a myriad of topologies including star, ring, tree, hypercube, fat hypercube (a hypercube with more than one processor at a node), or n-dimensional mesh.
[0112] Parallel computers based on interconnected networks must incorporate routing to enable the passing of messages between nodes that are not directly connected. The medium used for communication between the processors is likely to be hierarchical in large multiprocessor machines. Such resources are commercially available for purchase for dedicated use, or these resources can be accessed via “the cloud,” e.g., Amazon Cloud Computing.
[0113] A computer generally includes a processor coupled to a memory via a bus. Memory can include RAM or ROM and preferably includes at least one tangible, non-transitory medium storing instructions executable to cause the system to perform functions described herein. As one skilled in the art would recognize as necessary or best-suited for performance of the methods of the invention, systems of the invention include one or more processors (e.g., a central processing unit (CPU), a graphics processing unit (GPU), etc.), computer-readable storage devices (e.g., main memory, static memory, etc.), or combinations thereof which communicate with each other via a bus.
[0114] A processor may be any suitable processor known in the art, such as the processor sold under the trademark XEON E7 by Intel (Santa Clara, Calif.) or the processor sold under the trademark OPTERON 6200 by AMD (Sunnyvale, Calif.).
[0115] Memory may refer to a computer-readable storage device and can include any machine-readable medium on which is stored one or more sets of instructions (e.g., software embodying any methodology or function found herein), data (e.g., embodying any tangible physical objects such as the genetic sequences found in a patient's chromosomes), or both. While the computer-readable storage device can in an exemplary embodiment be a single medium, the term “computer-readable storage device” should be taken to include a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) that store the one or more sets of instructions or data. The term “computer-readable storage device” shall accordingly be taken to include, without limit, solid-state memories (e.g., subscriber identity module (SIM) card, secure digital card (SD card), micro SD card, or solid-state drive (SSD)), optical and magnetic media, and any other tangible storage media. Preferably, a computer-readable storage device includes a tangible, non-transitory medium. Such non-transitory media excludes, for example, transitory waves and signals. “Non-transitory memory” should be interpreted to exclude computer readable transmission media, such as signals, per se.
[0116] Input/output devices according to the invention may include a video display unit (e.g., a liquid crystal display (LCD) or a cathode ray tube (CRT) monitor), an alphanumeric input device (e.g., a keyboard), a cursor control device (e.g., a mouse or trackpad), a disk drive unit, a signal generation device (e.g., a speaker), a touchscreen, an accelerometer, a microphone, a cellular radio frequency antenna, and a network interface device, which can be, for example, a network interface card (NIC), Wi-Fi card, or cellular modem.

Sample Acquisition and Preparation

[0117] The invention includes methods for producing sequences (e.g., nucleic acid sequences, amino acid sequences) corresponding to nucleic acids recovered from biological samples. In some embodiments the resulting information can be used to identify mutations present in nucleic acid material obtained from a subject. In some embodiments, a sample, i.e., nucleic acids (e.g. DNA or RNA) are obtained from a subject, the nucleic acids are processed (lysed, amplified, and/or purified) and the nucleic acids are sequenced using a method described below. In many embodiments, the result of the sequencing is not a linear nucleic acid sequence, but a collection of thousands or millions of individual short nucleic acid reads that must be re-assembled into a sequence for the subject. Once the reads are aligned to produce a sequence, the aligned sequence can be compared to reference sequences to identify mutations that may be indicative of disease, for example. In other embodiments, the subject may be identified with particular mutations based upon the alignment of the reads against a reference sequence construct, i.e., a directed acyclic graph (“DAG”) as described above.
[0118] For any of the above purposes, methods may be applied to biological samples. The biological samples may, for example, comprise samples of blood, whole blood, blood plasma, tears, nipple aspirate, serum, stool, urine, saliva, circulating cells, tissue, biopsy samples, hair follicle or other samples containing biological material of the patient. One issue in conducting tests based on such samples is that, in most cases only a tiny amount of DNA or RNA containing a mutation of interest may be present in a sample. This is especially true in non-invasive samples, such as a buccal swab or a blood sample, where the mutant nucleic acids are present in very small amounts. In some embodiments, the nucleic acid fragments may be naturally short, that is, random shearing of relevant nucleic acids in the sample can generate short fragments. In other embodiments, the nucleic acids are purposely fragmented for ease of processing or because the sequencing techniques can only sequence reads of less than 1000 bases, e.g., less than 500 bases, e.g., less than 200 bases, e.g., less than 100 bases, e.g., less than 50 bases. While the methods described herein can be used to align sequences of varying length, in some embodiments, the majority of the plurality of nucleic acid reads will follow from the sequencing method and comprise less than 1000 bases, e.g., less than 500 bases, e.g., less than 200 bases, e.g., less than 100 bases, e.g., less than 50 bases.
[0119] Nucleic acids may be obtained by methods known in the art. Generally, nucleic acids can be extracted from a biological sample by a variety of techniques such as those described by Maniatis, et al., Molecular Cloning: A Laboratory Manual, Cold Spring Harbor, N.Y., pp. 280-281, (1982), the contents of which is incorporated by reference herein in its entirety.
[0120] It may be necessary to first prepare an extract of the sample and then perform further steps—i.e., differential precipitation, column chromatography, extraction with organic solvents and the like—in order to obtain a sufficiently pure preparation of nucleic acid. Extracts may be prepared using standard techniques in the art, for example, by chemical or mechanical lysis of the cell. Extracts then may be further treated, for example, by filtration and/or centrifugation and/or with chaotropic salts such as guanidinium isothiocyanate or urea or with organic solvents such as phenol and/or HCCl3 to denature any contaminating and potentially interfering proteins. In some embodiments, the sample may comprise RNA, e.g., mRNA, collected from a subject sample, e.g., a blood sample. General methods for RNA extraction are well known in the art and are disclosed in standard textbooks of molecular biology, including Ausubel et al., Current Protocols of Molecular Biology, John Wiley and Sons (1997). Methods for RNA extraction from paraffin embedded tissues are disclosed, for example, in Rupp and Locker, Lab Invest. 56:A67 (1987), and De Andres et al., BioTechniques 18:42044 (1995). The contents of each of these references is incorporated by reference herein in their entirety. In particular, RNA isolation can be performed using a purification kit, buffer set and protease from commercial manufacturers, such as Qiagen, according to the manufacturer's instructions. For example, total RNA from cells in culture can be isolated using Qiagen RNeasy mini-columns. Other commercially available RNA isolation kits include MASTERPURE Complete DNA and RNA Purification Kit (EPICENTRE, Madison, Wis.), and Paraffin Block RNA Isolation Kit (Ambion, Inc.). Total RNA from tissue samples can be isolated using RNA Stat-60 (Tel-Test). RNA prepared from tumor can be isolated, for example, by cesium chloride density gradient centrifugation.

Analytical Sequencing

[0121] Sequencing may be by any method known in the art. DNA sequencing techniques include classic dideoxy sequencing reactions (Sanger method) using labeled terminators or primers and gel separation in slab or capillary, sequencing by synthesis using reversibly terminated labeled nucleotides, pyrosequencing, 454 sequencing, allele specific hybridization to a library of labeled oligonucleotide probes, sequencing by synthesis using allele specific hybridization to a library of labeled clones that is followed by ligation, real time monitoring of the incorporation of labeled nucleotides during a polymerization step, polony sequencing, and SOLiD sequencing. Sequencing of separated molecules has more recently been demonstrated by sequential or single extension reactions using polymerases or ligases as well as by single or sequential differential hybridizations with libraries of probes. Prior to sequencing it may be additionally beneficial to amplify some or all of the nucleic acids in the sample. In some embodiments, the nucleic acids are amplified using polymerase chain reactions (PCR) techniques known in the art.
[0122] One example of a sequencing technology that can be used in the methods of the provided invention is Illumina sequencing (e.g., the MiSeq™ platform), which is a polymerase-based sequence-by-synthesis that may be utilized to amplify DNA or RNA. Illumina sequencing for DNA is based on the amplification of DNA on a solid surface using fold-back PCR and anchored primers. Genomic DNA is fragmented, and adapters are added to the 5′ and 3′ ends of the fragments. DNA fragments that are attached to the surface of flow cell channels are extended and bridge amplified. The fragments become double stranded, and the double stranded molecules are denatured. Multiple cycles of the solid-phase amplification followed by denaturation can create several million clusters of approximately 1,000 copies of single-stranded DNA molecules of the same template in each channel of the flow cell. Primers, DNA polymerase and four fluorophore-labeled, reversibly terminating nucleotides are used to perform sequential sequencing. After nucleotide incorporation, a laser is used to excite the fluorophores, and an image is captured and the identity of the first base is recorded. The 3′ terminators and fluorophores from each incorporated base are removed and the incorporation, detection and identification steps are repeated. When using Illumina sequencing to detect RNA the same method applies except RNA fragments are being isolated and amplified in order to determine the RNA expression of the sample. After the sequences are interrogated with the sequencer, they may be output in a data file, such as a FASTQ file, which is a text-based format for storing biological sequence and quality scores (see discussion above).
[0123] Another example of a DNA sequencing technique that may be used in the methods of the provided invention is Ion Torrent™ sequencing, offered by Life Technologies. See U.S. patent application numbers 2009/0026082, 2009/0127589, 2010/0035252, 2010/0137143, 2010/0188073, 2010/0197507, 2010/0282617, 2010/0300559, 2010/0300895, 2010/0301398, and 2010/0304982, the content of each of which is incorporated by reference herein in its entirety. In Ion Torrent™ sequencing, DNA is sheared into fragments of approximately 300-800 base pairs, and the fragments are blunt ended. Oligonucleotide adaptors are then ligated to the ends of the fragments. The adaptors serve as primers for amplification and sequencing of the fragments. The fragments can be attached to a surface and is attached at a resolution such that the fragments are individually resolvable. Addition of one or more nucleotides releases a proton (H+), which signal detected and recorded in a sequencing instrument. The signal strength is proportional to the number of nucleotides incorporated. Ion Torrent data may also be output as a FASTQ file.
[0124] Another example of a DNA and RNA sequencing technique that can be used in the methods of the provided invention is 454™ sequencing (Roche) (Margulies, M et al. 2005, Nature, 437, 376-380). 454™ sequencing is a sequencing-by-synthesis technology that utilizes also utilizes pyrosequencing. 454™ sequencing of DNA involves two steps. In the first step, DNA is sheared into fragments of approximately 300-800 base pairs, and the fragments are blunt ended. Oligonucleotide adaptors are then ligated to the ends of the fragments. The adaptors serve as primers for amplification and sequencing of the fragments. The fragments can be attached to DNA capture beads, e.g., streptavidin-coated beads using, e.g., Adaptor B, which contains 5′-biotin tag. The fragments attached to the beads are PCR amplified within droplets of an oil-water emulsion. The result is multiple copies of clonally amplified DNA fragments on each bead. In the second step, the beads are captured in wells (pico-liter sized). Pyrosequencing is performed on each DNA fragment in parallel. Addition of one or more nucleotides generates a light signal that is recorded by a CCD camera in a sequencing instrument. The signal strength is proportional to the number of nucleotides incorporated. Pyrosequencing makes use of pyrophosphate (PPi) which is released upon nucleotide addition. PPi is converted to ATP by ATP sulfurylase in the presence of adenosine 5′ phosphosulfate. Luciferase uses ATP to convert luciferin to oxyluciferin, and this reaction generates light that is detected and analyzed. In another embodiment, pyrosequencing is used to measure gene expression. Pyrosequecing of RNA applies similar to pyrosequencing of DNA, and is accomplished by attaching applications of partial rRNA gene sequencings to microscopic beads and then placing the attachments into individual wells. The attached partial rRNA sequence are then amplified in order to determine the gene expression profile. Sharon Marsh, Pyrosequencing® Protocols in Methods in Molecular Biology, Vol. 373, 15-23 (2007).
[0125] Another example of a DNA and RNA detection techniques that may be used in the methods of the provided invention is SOLiD™ technology (Applied Biosystems). SOLiD™ technology systems is a ligation based sequencing technology that may utilized to run massively parallel next generation sequencing of both DNA and RNA. In DNA SOLiD™ sequencing, genomic DNA is sheared into fragments, and adaptors are attached to the 5′ and 3′ ends of the fragments to generate a fragment library. Alternatively, internal adaptors can be introduced by ligating adaptors to the 5′ and 3′ ends of the fragments, circularizing the fragments, digesting the circularized fragment to generate an internal adaptor, and attaching adaptors to the 5′ and 3′ ends of the resulting fragments to generate a mate-paired library. Next, clonal bead populations are prepared in microreactors containing beads, primers, template, and PCR components. Following PCR, the templates are denatured and beads are enriched to separate the beads with extended templates. Templates on the selected beads are subjected to a 3′ modification that permits bonding to a glass slide. The sequence can be determined by sequential hybridization and ligation of partially random oligonucleotides with a central determined base (or pair of bases) that is identified by a specific fluorophore. After a color is recorded, the ligated oligonucleotide is cleaved and removed and the process is then repeated.
[0126] In other embodiments, SOLiD™ Serial Analysis of Gene Expression (SAGE) is used to measure gene expression. Serial analysis of gene expression (SAGE) is a method that allows the simultaneous and quantitative analysis of a large number of gene transcripts, without the need of providing an individual hybridization probe for each transcript. First, a short sequence tag (about 10-14 bp) is generated that contains sufficient information to uniquely identify a transcript, provided that the tag is obtained from a unique position within each transcript. Then, many transcripts are linked together to form long serial molecules, that can be sequenced, revealing the identity of the multiple tags simultaneously. The expression pattern of any population of transcripts can be quantitatively evaluated by determining the abundance of individual tags, and identifying the gene corresponding to each tag. For more details see, e.g. Velculescu et al., Science 270:484 487 (1995); and Velculescu et al., Cell 88:243 51 (1997, the contents of each of which are incorporated by reference herein in their entirety).
[0127] Another sequencing technique that can be used in the methods of the provided invention includes, for example, Helicos True Single Molecule Sequencing (tSMS) (Harris T. D. et al. (2008) Science 320:106-109). In the tSMS technique, a DNA sample is cleaved into strands of approximately 100 to 200 nucleotides, and a polyA sequence is added to the 3′ end of each DNA strand. Each strand is labeled by the addition of a fluorescently labeled adenosine nucleotide. The DNA strands are then hybridized to a flow cell, which contains millions of oligo-T capture sites that are immobilized to the flow cell surface. The templates can be at a density of about 100 million templates/cm2. The flow cell is then loaded into an instrument, e.g., HeliScope™ sequencer, and a laser illuminates the surface of the flow cell, revealing the position of each template. A CCD camera can map the position of the templates on the flow cell surface. The template fluorescent label is then cleaved and washed away. The sequencing reaction begins by introducing a DNA polymerase and a fluorescently labeled nucleotide. The oligo-T nucleic acid serves as a primer. The polymerase incorporates the labeled nucleotides to the primer in a template directed manner. The polymerase and unincorporated nucleotides are removed. The templates that have directed incorporation of the fluorescently labeled nucleotide are detected by imaging the flow cell surface. After imaging, a cleavage step removes the fluorescent label, and the process is repeated with other fluorescently labeled nucleotides until the desired read length is achieved. Sequence information is collected with each nucleotide addition step. Further description of tSMS is shown for example in Lapidus et al. (U.S. Pat. No. 7,169,560), Lapidus et al. (U.S. patent application number 2009/0191565), Quake et al. (U.S. Pat. No. 6,818,395), Harris (U.S. Pat. No. 7,282,337), Quake et al. (U.S. patent application number 2002/0164629), and Braslaysky, et al., PNAS (USA), 100: 3960-3964 (2003), the contents of each of these references is incorporated by reference herein in its entirety.
[0128] Another example of a sequencing technology that may be used in the methods of the provided invention includes the single molecule, real-time (SMRT) technology of Pacific Biosciences to sequence both DNA and RNA. In SMRT, each of the four DNA bases is attached to one of four different fluorescent dyes. These dyes are phospholinked. A single DNA polymerase is immobilized with a single molecule of template single stranded DNA at the bottom of a zero-mode waveguide (ZMW). A ZMW is a confinement structure which enables observation of incorporation of a single nucleotide by DNA polymerase against the background of fluorescent nucleotides that rapidly diffuse in an out of the ZMW (in microseconds). It takes several milliseconds to incorporate a nucleotide into a growing strand. During this time, the fluorescent label is excited and produces a fluorescent signal, and the fluorescent tag is cleaved off. Detection of the corresponding fluorescence of the dye indicates which base was incorporated. The process is repeated. In order to sequence RNA, the DNA polymerase is replaced with a with a reverse transcriptase in the ZMW, and the process is followed accordingly.
[0129] Another example of a sequencing technique that can be used in the methods of the provided invention is nanopore sequencing (Soni G V and Meller, A Clin Chem 53: 1996-2001) (2007). A nanopore is a small hole, of the order of 1 nanometer in diameter. Immersion of a nanopore in a conducting fluid and application of a potential across it results in a slight electrical current due to conduction of ions through the nanopore. The amount of current which flows is sensitive to the size of the nanopore. As a DNA molecule passes through a nanopore, each nucleotide on the DNA molecule obstructs the nanopore to a different degree. Thus, the change in the current passing through the nanopore as the DNA molecule passes through the nanopore represents a reading of the DNA sequence.
[0130] Another example of a sequencing technique that can be used in the methods of the provided invention involves using a chemical-sensitive field effect transistor (chemFET) array to sequence DNA (for example, as described in US Patent Application Publication No. 20090026082). In one example of the technique, DNA molecules can be placed into reaction chambers, and the template molecules can be hybridized to a sequencing primer bound to a polymerase. Incorporation of one or more triphosphates into a new nucleic acid strand at the 3′ end of the sequencing primer can be detected by a change in current by a chemFET. An array can have multiple chemFET sensors. In another example, single nucleic acids can be attached to beads, and the nucleic acids can be amplified on the bead, and the individual beads can be transferred to individual reaction chambers on a chemFET array, with each chamber having a chemFET sensor, and the nucleic acids can be sequenced.
[0131] Another example of a sequencing technique that can be used in the methods of the provided invention involves using an electron microscope (Moudrianakis E. N. and Beer M. Proc Natl Acad Sci USA. 1965 March; 53:564-71). In one example of the technique, individual DNA molecules are labeled using metallic labels that are distinguishable using an electron microscope. These molecules are then stretched on a flat surface and imaged using an electron microscope to measure sequences.
[0132] Additional detection methods can utilize binding to microarrays for subsequent fluorescent or non-fluorescent detection, barcode mass detection using a mass spectrometric methods, detection of emitted radiowaves, detection of scattered light from aligned barcodes, fluorescence detection using quantitative PCR or digital PCR methods. A comparative nucleic acid hybridization array is a technique for detecting copy number variations within the patient's sample DNA. The sample DNA and a reference DNA are differently labeled using distinct fluorophores, for example, and then hybridized to numerous probes. The fluorescent intensity of the sample and reference is then measured, and the fluorescent intensity ratio is then used to calculate copy number variations. Methods of comparative genomic hybridization array are discussed in more detail in Shinawi M, Cheung S W The array CGH and its clinical applications, Drug Discovery Today 13 (17-18): 760-70. Microarray detection may not produce a FASTQ file directly, however programs are available to convert the data produced by the microarray sequencers to a FASTQ, or similar, format.
[0133] Another method of detecting DNA molecules, RNA molecules, and copy number is fluorescent in situ hybridization (FISH). In Situ Hybridization Protocols (Ian Darby ed., 2000). FISH is a molecular cytogenetic technique that detects specific chromosomal rearrangements such as mutations in a DNA sequence and copy number variances. A DNA molecule is chemically denatured and separated into two strands. A single stranded probe is then incubated with a denatured strand of the DNA. The signals stranded probe is selected depending target sequence portion and has a high affinity to the complementary sequence portion. Probes may include a repetitive sequence probe, a whole chromosome probe, and locus-specific probes. While incubating, the combined probe and DNA strand are hybridized. The results are then visualized and quantified under a microscope in order to assess any variations.
[0134] In another embodiment, a MassARRAY-based gene expression profiling method is used to measure gene expression. In the MassARRAY™-based gene expression profiling method, developed by Sequenom, Inc. (San Diego, Calif.) following the isolation of RNA and reverse transcription, the obtained cDNA is spiked with a synthetic DNA molecule (competitor), which matches the targeted cDNA region in all positions, except a single base, and serves as an internal standard. The cDNA/competitor mixture is PCR amplified and is subjected to a post-PCR shrimp alkaline phosphatase (SAP) enzyme treatment, which results in the dephosphorylation of the remaining nucleotides. After inactivation of the alkaline phosphatase, the PCR products from the competitor and cDNA are subjected to primer extension, which generates distinct mass signals for the competitor- and cDNA-derives PCR products. After purification, these products are dispensed on a chip array, which is pre-loaded with components needed for analysis with matrix-assisted laser desorption ionization time-of-flight mass spectrometry (MALDI-TOF MS) analysis. The cDNA present in the reaction is then quantified by analyzing the ratios of the peak areas in the mass spectrum generated. For further details see, e.g. Ding and Cantor, Proc. Natl. Acad. Sci. USA 100:3059 3064 (2003).
[0135] Further PCR-based techniques include, for example, differential display (Liang and Pardee, Science 257:967 971 (1992)); amplified fragment length polymorphism (iAFLP) (Kawamoto et al., Genome Res. 12:1305 1312 (1999)); BeadArray™ technology (Illumina, San Diego, Calif.; Oliphant et al., Discovery of Markers for Disease (Supplement to Biotechniques), June 2002; Ferguson et al., Analytical Chemistry 72:5618 (2000)); Beads Array for Detection of Gene Expression (BADGE), using the commercially available Luminex100 LabMAP system and multiple color-coded microspheres (Luminex Corp., Austin, Tex.) in a rapid assay for gene expression (Yang et al., Genome Res. 11:1888 1898 (2001)); and high coverage expression profiling (HiCEP) analysis (Fukumura et al., Nucl. Acids. Res. 31(16) e94 (2003)). The contents of each of which are incorporated by reference herein in their entirety.
[0136] In certain embodiments, variances in gene expression can also be identified, or confirmed using a microarray techniques, including nylon membrane arrays, microchip arrays and glass slide arrays, e.g., such as available commercially from Affymetrix (Santa Clara, Calif.). Generally, RNA samples are isolated and converted into labeled cDNA via reverse transcription. The labeled cDNA is then hybridized onto either a nylon membrane, microchip, or a glass slide with specific DNA probes from cells or tissues of interest. The hybridized cDNA is then detected and quantified, and the resulting gene expression data may be compared to controls for analysis. The methods of labeling, hybridization, and detection vary depending on whether the microarray support is a nylon membrane, microchip, or glass slide. Nylon membrane arrays are typically hybridized with P-dNTP labeled probes. Glass slide arrays typically involve labeling with two distinct fluorescently labeled nucleotides. Methods for making microarrays and determining gene product expression (e.g., RNA or protein) are shown in Yeatman et al. (U.S. patent application number 2006/0195269), the content of which is incorporated by reference herein in its entirety.
[0137] In some embodiments, mass spectrometry (MS) analysis can be used alone or in combination with other methods (e.g., immunoassays or RNA measuring assays) to determine the presence and/or quantity of the one or more biomarkers disclosed herein in a biological sample. In some embodiments, the MS analysis includes matrix-assisted laser desorption/ionization (MALDI) time-of-flight (TOF) MS analysis, such as for example direct-spot MALDI-TOF or liquid chromatography MALDI-TOF mass spectrometry analysis. In some embodiments, the MS analysis comprises electrospray ionization (ESI) MS, such as for example liquid chromatography (LC) ESI-MS. Mass analysis can be accomplished using commercially-available spectrometers. Methods for utilizing MS analysis, including MALDI-TOF MS and ESI-MS, to detect the presence and quantity of biomarker peptides in biological samples are known in the art. See for example U.S. Pat. Nos. 6,925,389; 6,989,100; and 6,890,763 for further guidance, each of which is incorporated by reference herein in their entirety.
[0138] Protein sequences for use with the methods, sequence constructs, and systems of the invention can be determined using a number of techniques known to those skilled in the relevant art. For example, amino acid sequences and amino acid sequence reads may be produced by analyzing a protein or a portion of a protein with mass spectrometry or using Edman degradation. Mass spectrometry may include, for example, matrix-assisted laser desorption/ionization (MALDI) time-of-flight (TOF) MS analysis, such as for example direct-spot MALDI-TOF or liquid chromatography MALDI-TOF mass spectrometry analysis, electrospray ionization (ESI) MS, such as for example liquid chromatography (LC) ESI-MS, or other techniques such as MS-MS. Edman degradation analysis may be performed using commercial instruments such as the Model 49X Procise protein/peptide sequencer (Applied Biosystems/Life Technologies). The sequenced amino acid sequences, i.e., polypeptides, i.e., proteins, may be at least 10 amino acids in length, e.g., at least 20 amino acids in length, e.g., at least 50 amino acids in length.

INCORPORATION BY REFERENCE

[0139] References and citations to other documents, such as patents, patent applications, patent publications, journals, books, papers, web contents, have been made throughout this disclosure. All such documents are hereby incorporated herein by reference in their entirety for all purposes.

EQUIVALENTS

[0140] Various modifications of the invention and many further embodiments thereof, in addition to those shown and described herein, will become apparent to those skilled in the art from the full contents of this document, including references to the scientific and patent literature cited herein. The subject matter herein contains important information, exemplification and guidance that can be adapted to the practice of this invention in its various embodiments and equivalents thereof.
(57)

Claims

1. A method of identifying cancer-induced genetic mutations, comprising using a processor coupled to a non-transitory computer-readable medium to perform:
obtaining a first nucleic acid sequence corresponding to a nucleic acid in a non-cancerous sample from a subject;
identifying differences between the first nucleic acid sequence and a selected reference sequence;
representing, in the non-transitory computer-readable storage medium, the identified differences between the first nucleic acid sequence and the selected reference sequence as two or more alternative paths in a first reference directed acyclic graph (DAG) comprising nodes, wherein each alternative path is placed at a position in the first reference DAG where there is a difference between the first nucleic acid sequence and the reference sequence;
aligning one or more sequence reads from a second sequence corresponding to a cancerous sample from the subject to the first reference DAG, wherein the aligning considers two or more alternative paths by looking backward to any prior nodes on the first reference DAG to find a maximum score for the one or more sequence reads; and
identifying, based on the aligned one or more sequence reads to the first reference DAG, differences between the second sequence and the first reference DAG as new mutations correlated with the cancer.
2. The method of claim 1, further comprising:
preparing a second reference DAG representing the differences between the first sequence, the second sequence, and the reference sequence as two or more alternative paths at positions in the second reference DAG where there is a difference between the first sequence and the reference sequence or where there is a difference between the second sequence and the first sequence;
aligning one or more reads from a third sequence corresponding to an advanced cancerous sample from the organism to the second reference DAG; and
identifying differences between the third sequence and the second reference DAG as mutations due to the advanced cancer.
3. The method of claim 1, wherein the second sequence represents a major genetic clone due to the cancer.
4. The method of claim 2, wherein the third sequence represents a minor genetic clone due to the cancer.
5. The method of claim 1, wherein the cancer is selected from breast, lung, skin, prostate, thyroid, pancreatic, bladder, or ovarian cancer.
6. The method of claim 1, wherein the cancer is leukemia or lymphoma.
7. The method of claim 1, wherein the sequence reads are at least about 50 bp in length.
8. The method of claim 1, wherein the differences between the second sequence and the first reference DAG comprise insertions, deletions, polymorphisms or structural variants.
9. The method of claim 1, wherein the reference DAG represents a genome.
10. The method of claim 1, further comprising collecting the non-cancerous non-diseased sample from the subject.
11. The method of claim 10, wherein the first nucleic acid sequence is obtained by sequencing a nucleic acid isolated from the non-cancerous sample.
12. The method of claim 1, further comprising collecting the cancerous sample from the subject.
13. The method of claim 12, wherein the one or more sequence reads from the second sequence are obtained by sequencing a nucleic acid isolated from the cancerous sample.
14. A method of identifying mutations due to an advanced stage of cancer in a subject, comprising:
obtaining a first sequence corresponding to a non-cancerous sample from the subject and a second sequence corresponding to a cancerous sample from the subject organism;
identifying differences between the first sequence and the second sequence;
representing the identified differences between the first sequence and the second sequence as two or more alternative paths in a reference directed acyclic graph (DAG) comprising nodes, wherein each alternative path is placed at a position in the reference DAG where there is a difference between the first sequence and the second sequence;
aligning a sequence read corresponding to an advanced cancerous sample from the subject to the reference DAG, wherein the aligning considers the two or more alternative paths by looking backward to any prior nodes on the reference DAG to find a maximum score for the sequence read; and
identifying, based on the aligned sequence read to the reference DAG, differences between the sequence read and the reference DAG as new mutations correlated with an advanced stage of the cancer.
15. The method of claim 14, further comprising diagnosing the subject as having an advanced stage of the cancer.
16. The method of claim 14, wherein the cancer is selected from breast, lung, skin, prostate, thyroid, pancreatic, bladder, or ovarian cancer.
17. The method of claim 14, wherein the advanced stage of the cancer is metastatic cancer.
18. The method of claim 1, further comprising representing the new mutations correlated with the cancer as alternative paths in the reference DAG.
19. The method of claim 18, further comprising aligning one or more sequence reads from a third sequence corresponding to a sample later collected from the same subject to the first reference DAG, and monitoring cancer progression or recurrence based on the alignment.
20. The method of claim 1, further comprising reporting the new mutations correlated with the cancer to a health care provider.
*****

Sign in to the Lens

Feedback