![]() |
|
![]() |
||||
Thread | Thread Starter | Forum | Replies | Last Post |
PubMed: Targeted sequencing using Affymetrix CustomSeq Arrays. | Newsbot! | Literature Watch | 0 | 05-27-2011 04:50 AM |
Enrichment/targeted sequencing using very short tags | DNA Sorcerer | Sample Prep / Library Generation | 0 | 05-06-2011 05:55 AM |
Targeted sequence capture - Nimblegen + GS Junior | sutree | 454 Pyrosequencing | 1 | 10-19-2010 06:53 AM |
Rapid library prep and Nimblegen enrichment | Tom Haltern | 454 Pyrosequencing | 1 | 08-05-2010 06:55 AM |
Illumina/Solexa sequencing of human exons enriched with Nimblegen arrays | ECO | Illumina/Solexa | 9 | 06-05-2009 02:58 PM |
![]() |
|
Thread Tools |
![]() |
#1 |
--Site Admin--
Location: SF Bay Area, CA, USA Join Date: Oct 2007
Posts: 1,358
|
![]() ![]() Two papers were recently published in Nature Methods detailing an astonishing new technique which combines the flexibility of on-demand array technology with the power of next gen sequencing instrumentation. These papers demonstrate the specific capture and sequencing of specific genomic regions, using a programmable microarray platform (from Nimblegen) Direct selection of human genomic loci by microarray hybridization (Albert, et al) and Microarray-based genomic selection for high-throughput resequencing (Okou, et al) Both studies capture defined regions of the genome (either ~6,700 exons or up to a 5Mb contiguous region) using Nimblegen arrays, and present 454 sequencing data of the enriched regions. ~65-75% on target reads were reported, with median single base coverage of 5-7 fold. The Albert study validated their technique against four samples from the HapMap collection, and were able to identify 79-94% of known variants in the target regions. A third paper in the same issue of Nature Methods from the Church lab ("Multiplex amplification of large sets of human exons", Porreca et al) demonstrates another technique for capture. Rather than capturing the targets using hybridization directly on the array, this study uses a programmable array to synthesize "molecular inversion" capture probes that are cleaved from the array and used to fish out small regions of interest (~60-191 bp exon fragments). Enriched fractions were then sequenced with the Solexa Genome Analyzer. The results reported in this study were less than impressive, with only 28% on target hits. There was also a significant problem with calling heterozygous polymorphisms, however the authors hope this can be optimized at the reaction level. This technique, which relies on a dual hybridization event surrounding the region of interest followed by gap-filling/ligation, is much more complicated and seems that it will require intense optimization to approach the success had with direct capture. In any event, this enrichment technology will make a significant impact on any study examining a defined subset of the genome, such as candidate region sequencing workflows. What once was a laborious process of PCR primer design, optimization, amplification, and normalization, has become a simple one-pot hybridization event. |
![]() |
![]() |
![]() |
#2 |
--Site Admin--
Location: SF Bay Area, CA, USA Join Date: Oct 2007
Posts: 1,358
|
![]()
There was just a fourth Nimblegen paper released...I put down some thoughts about it here:
http://seqanswers.com/forums/showthread.php?t=11 |
![]() |
![]() |
![]() |
#3 |
--Site Admin--
Location: SF Bay Area, CA, USA Join Date: Oct 2007
Posts: 1,358
|
![]()
I'd love to hear some comments from people about how this is affecting experiments in their labs.
Completely obsoleting huge PCR-driven resequencing studies? Anyone talked to Nimblegen to get an idea of availability? At GME2007 a 454 rep told me they will be launching the exon platform soon (6 mo.?), but perhaps on the higher density chip so they can get it all on one. Anyway, if you're reading this, don't be shy... ![]() |
![]() |
![]() |
![]() |
#4 |
Junior Member
Location: SF bay area Join Date: Dec 2007
Posts: 6
|
![]()
Two points about the recently published target enrichment methods.
1. Comparing the methods from the different papers is not easy. First, Okou et al. used a resequencing array, which gives no measure of how many “off target” sequences were captured. Second, there is no set method of measuring recovery (how many of the targeted sequences were found) or specificity (what fraction of the reads mapped to the targeted regions). Table 1 in the Hodges et al. Nature Genetics paper shows how their numbers for recovery can vary from 54% to 67% depending on whether flanking regions are included. Also, to clarify ECO’s statement, the Church lab recovered only 28% of their targets across 2 experiments, while the run numbers from the methods give the % of on-target reads as 2694087/4794590 = 56%. The Hodges paper had a comparable raw specificity of 1993201 on-target reads out of 4198280 total reads = 47%. 2. There are 3 important workflow considerations for the average lab which the papers gloss over: Time, sample required, and money. Nimblegen: 65 hour hyb of 20 ug DNA to a custom array Church lab: 1.5 hour hyb of 1.5ug DNA to pool of oligos (admittedly, with more molecular biology steps). Depending on the sequencing platform speed and experimental design, the time may not matter. But 20ug of DNA is hard to get from human samples without amplification, while 1.5ug is feasible for humans and easily available for HapMap cell lines. Considering cost, Nimblegen claims the arrays can be reused at least once (but how can you be sure that a variant sequence came from your sample, and is not carryover from the previous sample?) However, the Church lab oligo pool was amplified and presumably could be aliquoted for dozens or even hundreds of individual samples. Clearly George Church is interested in driving the cost of sequencing way down, which may be in part why he uses this method. In the end we will probably have to wait for the next round of papers to see how well the different methods can be optimized. |
![]() |
![]() |
![]() |
#5 |
Member
Location: Sydney, Australia Join Date: Jan 2008
Posts: 83
|
![]()
I'm a postdoc employed to sequence a candidate region of interest via next-gen sequencing methods.
Here are my thoughts on Nimblegen capture as bullet points ripped from a presentation I gave to my group in November. Sorry, you don't get to the enjoy the accompanying figures ![]() Multiplexity: The number of independent capture reactions performed simultaneously in a single reaction. Specificity: The fraction of captured nucleic acids that derive from targeted regions. Uniformity: The relative abundances of targeted sequences after selective capture. BAC Pulldown (Bashiardes et al. (2005))
LR-PCR
Last edited by sci_guy; 01-24-2008 at 06:51 PM. |
![]() |
![]() |
![]() |
#6 |
Member
Location: Sydney, Australia Join Date: Jan 2008
Posts: 83
|
![]()
I've been in contact with Roche and I am lead to believe that a Nimblegen sequence capture service will be available around March, not long after human custom HD2 chips hit the market.
I have some open-ended questions regarding this "new and improved" formula ![]() The HD2 array in order to go from 385K probes to 2.1M takes up much more real estate on the slide but also incorporates a feature shrink, from 16 to 13 micron. Presumably, there are less total oligos in this smaller 13 micron region. Will the feature shrink and increased numbers of probes make HD2 "noisier" than the 385K chip? Will we still see reasonable specificity? Why have Hodges et al. (2007) used a custom sequencing primer approach to their downstream Solexa sequencing? Will the 1G instrument handle a longer primer OK? Why do they not include the Solexa adaptor as an adaptor in the Nimblegen protocol? I ask these questions as I would like to capture 500 bp fragments on the Nimblegen chip and then alter the protocol to reduce the fragment sizes down before a SOLiD or Solexa/Illumina run. I'm in the process of emailing Solexa and the corresponding authors of these papers to get some answers. If you can inform me on any of this, please post! Last edited by sci_guy; 01-24-2008 at 06:53 PM. |
![]() |
![]() |
![]() |
#7 |
--Site Admin--
Location: SF Bay Area, CA, USA Join Date: Oct 2007
Posts: 1,358
|
![]()
Great posts sci_guy! If you are willing to share your full presentation I would be happy to host it for the community (either just images or a powerpoint slideshow). Send me a PM or email if you are interested.
I have heard from Agilent that they are about to release an improved capture method (specifically NOT the Church circularization method) at the AGBT conference next month. This will supposedly be in a kitted format including all the reagents for capture, combined with a custom array. Great stuff...hopefully we can pull some of the Nimblegen folks in here to respond ... ![]() |
![]() |
![]() |
![]() |
#8 |
--Site Admin--
Location: SF Bay Area, CA, USA Join Date: Oct 2007
Posts: 1,358
|
![]()
For future reference, sci_guy's powerpoint deck is in this thread: http://seqanswers.com/forums/showthread.php?t=46
|
![]() |
![]() |
![]() |
#9 |
Member
Location: Dublin, Ireland Join Date: Nov 2007
Posts: 28
|
![]()
Does anyone know if the HD2 chips will be available for species other than human/mouse? Or if Nimblegen will make custom chips to order?
|
![]() |
![]() |
![]() |
#10 |
Member
Location: Sydney, Australia Join Date: Jan 2008
Posts: 83
|
![]()
pmcget - It's my belief that Nimblegen will offer custom HD2 CGH chips around March.
|
![]() |
![]() |
![]() |
#11 |
Junior Member
Location: GA Join Date: Jan 2008
Posts: 6
|
![]()
SCI guy,
what a great job on comparing information from the various paper. However, I would like to bring some correction to the info provided on nimblegen capture chip. for Okou, et al, 2007. T heir enrichment derivation (10 Ct’s enrichment of DNA. (2^10 ~ 1000)) was applied on 300k region not 50k region. This should have been 3x10^9 genome / 300x10^3 region = 10,000 fold enrichment for pure capture. therefore A 1000x fold enrichment of 300 kb target DNA means that 1 in 10 (not 1 in 60) captured DNA sequences are from the target region. for Albert, et al. 2007. Their enrichment calculation was based on the percentage of read that mapped back the to the target for the triplicate result of 5 Mb region. it was calculated as follow. In theory, 3x10^9 genome / 5x10^6 region = 600 fold enrichment = 100% enrichment. But with average of 72% read that mapped to the target [(75%+65%+77%)/3 = 72.33% =~72%) the enrichment represent 432 fold [(600 x 72/100]. If Okou, et al had derived their enrichment the same as that Albert et al. I think they found 99.1% of their read to map back to the target, then this represents 9910 fold enrichment [(10000 x 99.1)/100]. Ii that is true, whose estimation of enrichment or selectivity is accurate? Last edited by ivoire; 01-28-2008 at 03:02 PM. |
![]() |
![]() |
![]() |
#12 |
Junior Member
Location: GA Join Date: Jan 2008
Posts: 6
|
![]()
With the Nimblegen capture array, the larger the size the more economical it is. the price of the chip should be the same regardless of the size (of course the larger the size the more spaced the probes on the chip will be). I also do not understand why Nimblegen capture chip required 3 days hybe. this is just hybridization of oligo and it should in theory happen overnight (~14 - 18 hrs). Additionally what up with the 42 degree hybe?
Last edited by ivoire; 01-28-2008 at 03:03 PM. |
![]() |
![]() |
![]() |
#13 |
Junior Member
Location: SF bay area Join Date: Dec 2007
Posts: 6
|
![]()
My guess for the reason for the long hybs to Nimblegen arrays is that at 42 degrees, it takes a LONG time for the hybridization reaction to approach equilibrium. While short oligos can hyb quickly in solution, the oligo probes on the array are fixed and immobile, so the reaction relies on diffusion of the target genomic DNA fragments. These target DNAs are at least several hundred nt, and it will take each one a lot longer than overnight to find its complement on the array surface, esp. if the concentration is limiting (note the large sample requirement in these protocols). It would be much faster if the targets were short, like microRNAs. The combination of large sample+slow hyb times for the array based methods may give an edge to solution-based capture methods if the bugs get worked out (the solution capture probes could find their targets more quickly than fixed capture probes).
I would bet that the 42 degree hyb temperature could be determined not by biochemistry, but by engineering. For Nimblegen probe lengths, the hyb could more stringent (and diffusion is faster) at 65 degrees, but its harder to contain a small volume of liquid at a higher temperature for a long time. |
![]() |
![]() |
![]() |
#14 | ||
Member
Location: Sydney, Australia Join Date: Jan 2008
Posts: 83
|
![]() Quote:
Quote:
My presentation is biased towards our intended application, where we need high enrichment across a large genomic region in a small number of samples for, most likely, SNP discovery in exons. Affymetrix sequencing chips cost around $23K AUD for mask design. The 49-format chip has a capacity to sequence 300 kb and one must order one wafer of chips minimum (49 chips @ ~$600 AUD). Given that:
|
||
![]() |
![]() |
![]() |
#15 | |
Member
Location: Sydney, Australia Join Date: Jan 2008
Posts: 83
|
![]() Quote:
3.84 / 0.666 =~ 5.7x potential probes by area By actual probes calculation: 2.1M probes / 385,000 probes = 5.45x more probes Probe spacing (as I have understood your point) seems relatively the same. The feature shrink means probe is even more limiting so hyb kinetics will differ. Likewise, the addition of 5.45x more probes. I would love to see some empirical data on enrichment with HD2. HD2 chips are twice the price of 385,000 probe chips. See costs and array size data in my presentation here: http://seqanswers.com/forums/showthread.php?t=46 |
|
![]() |
![]() |
![]() |
#16 | |
Member
Location: Palo Alto, CA Join Date: Feb 2008
Posts: 16
|
![]() Quote:
Also, does anyone have an idea how much amount of oligos per feature (spot) on Nimblegen array? It seems that it is extremely low so that they need 3 day hyb time and 20ug of gDNA to drive the kinetics. |
|
![]() |
![]() |
![]() |
#17 | |
Junior Member
Location: Hefei, China Join Date: Feb 2008
Posts: 6
|
![]() Quote:
|
|
![]() |
![]() |
![]() |
#18 |
Junior Member
Location: New York Join Date: Feb 2008
Posts: 1
|
![]()
I'm not sure if you ever received the answers from Solexa or the authors you contacted, but I'm happy to answer. In the Hodges et al. paper, the DNA was prepared in two different formats, the Roche Nimblegen way (fragment sizes 500-600 or more bp) - which is suitable for 454 (conveniently since both 454 and Nimblegen are owned by Roche). Also, for these experiments common linkers were chosen based on previous experience (the sequence ligates well and amplifies well) and the fact that they do not add a lot more sequence onto the already large fragments. Having these linkers requires that additional adaptation be performed for Illumina sequencing. Therefore, the custom sequencing primer had to include the first set of adaptors, as they are the most internal and adjacent to the insert to be sequenced.
In the same paper, an alternative DNA preparation was also reported. It turned out to be much more suitable for Solexa, and bypassed the extra ligation step. Basically, the libraries were prepared according to the Solexa kit protocol prior to hybridization and the fragment sizes were much shorter (150-300), a much more suitable size for short read platforms. This resulted in 90% base pair coverage and a pretty high sequencing depth for such a large target size (6 Mb of one exon chip). Most importantly, it is imperative to have the DNA fragments ligated to common oligo adaptors before hybridization. The reason for this is that thermal elution of the slides at 95 degrees, may result in oligo detachment. Therefore, having the DNA ligated with common oligos allows enrichment of the genomic DNA and avoids sequencing of oligos and non-specific material. |
![]() |
![]() |
![]() |
#19 |
Member
Location: Sydney, Australia Join Date: Jan 2008
Posts: 83
|
![]()
Reseq: Many thanks for the info!
Horigen: No one got back to me. But I have heard anecdotally that Nimblegen is on track to offer 385,000 probe chip enrichment as a service around the end of March. There are some delays with the HD2 chip. Agilent will be releasing liquid phase sequence capture product later this year. |
![]() |
![]() |
![]() |
#20 |
Member
Location: Palo Alto, CA Join Date: Feb 2008
Posts: 16
|
![]()
It seems to me that shearing the genomic DNA is quite critical. If one can reliably shear DNA to ~100bp or 200bp, the genomic DNA will move faster (thus increase hybridization efficiency), and the library construction for sequence whole exon is easier (one may simply sequence ends of a fragment).
Anybody has a robust protocol to shear human genomic DNA to 100-200bp? Thanks! |
![]() |
![]() |
![]() |
Thread Tools | |
|
|