![]() |
|
![]() |
||||
Thread | Thread Starter | Forum | Replies | Last Post |
bowtie error: extra parameters specified | BioSlayer | Bioinformatics | 1 | 10-07-2011 11:38 AM |
PCR duplicates increase when excess of beads | tdm | SOLiD | 10 | 03-31-2011 09:48 AM |
Number of PCR How many PCR cycles to enrich adapter-modified DNA fragments | MGH Man | Sample Prep / Library Generation | 5 | 07-26-2010 06:15 AM |
titrate adapter & extra bands in library | xenia.zhang | Sample Prep / Library Generation | 0 | 01-13-2010 10:55 AM |
PCR enrichment of libraries in 12 cycles or less? | seqgirl123 | Illumina/Solexa | 6 | 07-05-2009 03:53 PM |
![]() |
|
Thread Tools |
![]() |
#1 |
Senior Member
Location: Purdue University, West Lafayette, Indiana Join Date: Aug 2008
Posts: 2,317
|
![]()
So the dogma goes:
Use as few PCR cycles during library construction as possible. One can imagine various rationales behind this one.
But is #3 even true? Before delving into it, I would like to exclude issues having to do with overrunning the supply of reactants in the PCR. If the amount of final product approaches the total supply of dNTPs in the reaction, I can easily imagine higher levels of misincorporation. By "errors" here, we mean errors per total bases, right? If the PCR polymerase used is like Taq polymerase it likely has an error rate of about 1 in 10,000 bases polymerized. Templates: One thousand 100 base amplicons. The question: will the error rate per amplicon be higher after 20 cycles than it was after 10 cycles? Again, presuming reactants are not limiting and each cycle is 100% efficient -- that is, doubling the number of amplicons: After 10 cycles there will be 1 million (2^10 * 1000) amplicons (~0.1 pg of DNA). After 20 cycles there will be 1 billion (2^20 *1000) amplicons (~0.1 ng of DNA). Doesn't intuition tell us that after 10 additional cycles the errors will have compounded and the overall error rate we might detect by sequencing 1000 of the amplicons will have increased? Ignoring indels, I don't see this is the case. Sure, Taq polymerase would tend to misincorporate a base in 1% of the product strands that it creates, and that erroneous template will then be amplified each cycle. But the 99% of product strands that did not contain an error will also be amplified each cycle. Meaning, your error rate in your product strands will just be the error rate of the polymerase -- 1 in 10,000 bases, 1% of the amplicons in this case. Don't get me wrong: I have sequenced PCR products. The error rate is way higher than 1 in 10,000 bases. So what is wrong with my logic? -- Phillip |
![]() |
![]() |
![]() |
#2 |
Member
Location: Groningen Join Date: Feb 2010
Posts: 78
|
![]()
So I'm rather new in bioinformatics, I have little choice other than to go what I'm taught and told. One of our lab vetarans once expressed his opinion on PCR prior to sequencing: he said that back in the days when he was using "Taq version 0.1" he hardly saw any bias. That he didn't expect that PCR caused much trouble now, even though there's much discussion on the subject of PCR duplicate reads and errors introduced by PCR.
And that is why I'd be interested to read reactions to Philips post. |
![]() |
![]() |
![]() |
#3 |
Senior Member
Location: Bethesda MD Join Date: Oct 2009
Posts: 509
|
![]()
Hi Phillip,
The problem arises b/c the errors are cumulative. Once you generate a mutation, it essentially becomes fixed at that fraction of the population since it's now template for all subsequent rounds of amplification. And every round of amplification generates additional errors, so the fraction of mutant molecules is always increasing. By calculating the probability, it should be clear. Consider a single 100bp amplicon. The likelihood that a single base replicates faithfully is 1-0.0001 (the error rate) = 0.9999. The likelihood that the entire amplicon is correct is 0.9999^100 (the amplicon size) = .99. After 10 rounds of PCR, the likelihood of being error-free is .99^10 = .90 (still pretty good odds), but after 20 cycles that drops to .99^20 = .82. The remainder will contain one or more errors. -Harold |
![]() |
![]() |
![]() |
#4 | |
Senior Member
Location: Boston area Join Date: Nov 2007
Posts: 747
|
![]() Quote:
|
|
![]() |
![]() |
![]() |
#5 | ||
Senior Member
Location: Purdue University, West Lafayette, Indiana Join Date: Aug 2008
Posts: 2,317
|
![]()
I wrote:
Quote:
Quote:
Using this method, one could calculate a best case scenario for detection of minor variants in pools of samples. -- Phillip |
||
![]() |
![]() |
![]() |
#6 | |
Junior Member
Location: San Francisco Join Date: Aug 2011
Posts: 2
|
![]() Quote:
I used your principle above to calculate error rates in a library preparation for sequencing (I am getting a high number of false heterozygous calls and I am trying to investigate why). My calculations are below: enzyme error rate: 5.5x10-6 base accuracy: 1-5.5x10-6 likelihood of correct amplicon: accuracy x length (304) = 0.989077 likelihood of correct amplicon after 35 cycles: 0.680865 I thought that this number is way too high! Can you imagine doing variant calling with such error rate? When I try to calculate the total number of amplicon molecules that does not have an error in my library at the end, I get stuck. I know that the input DNA is 50ng in the reaction, so using a simple calculation considering 3.3pg = 1 genome, I should have around 15,100 copies of my target in the beginning of the reaction. Theoretically, the calculations above show me that over 30% of my product will contain an error, not to mention sequencing errors! I know that in practice I don't get that, it is up to 1% both library prep and sequencing error. Where am I mistaken? Best, Camila. |
|
![]() |
![]() |
![]() |
#7 | |
Member
Location: Bay area Join Date: Jun 2012
Posts: 77
|
![]() Quote:
FWIW. |
|
![]() |
![]() |
![]() |
#8 | |
Junior Member
Location: uk Join Date: Dec 2009
Posts: 6
|
![]() Quote:
If you run 35 cycles of PCR, you're not going to get 35 doublings of DNA, because 1) each cycle is less than 2-fold even early in the reaction, and 2) reactants run out. Consider that 50ng * 2^35 = 1.7 kg of DNA. If you know the yield of the PCR reaction you can work out how many actual replications you're getting, on average. |
|
![]() |
![]() |
![]() |
#9 |
Super Moderator
Location: Walnut Creek, CA Join Date: Jan 2014
Posts: 2,707
|
![]()
It's important to note here that with every generation, only the new reads can have additional errors. So, for example, (correct rate)^(cycles) is not quite accurate; the actual number is better, and increasingly better as the duplication ratio per cycle drops from 100%.
|
![]() |
![]() |
![]() |
#10 |
Senior Member
Location: Bethesda MD Join Date: Oct 2009
Posts: 509
|
![]()
Brian is absolutely correct that replication efficiency drops at every cycle as reagents are consumed. If not, then 35 cycles of 50 ng input would yield ~1.7 kg of product :-).
However, Phillip's original query asked to exclude the issue of reagent limitation. And the concept underlying Muller's ratchet, although not directly applicable to PCR, explains why the mutational load increases with each round of replication. |
![]() |
![]() |
![]() |
#11 |
Junior Member
Location: San Francisco Join Date: Aug 2011
Posts: 2
|
![]()
This makes sense, thanks for all the answers.
|
![]() |
![]() |
![]() |
Thread Tools | |
|
|