|Home | About | Journals | Submit | Contact Us | Français|
Quantitative real-time polymerase chain reactions (qRT-PCR) have become the method of choice for rapid, sensitive, quantitative comparison of RNA transcript abundance. Useful data from this method depend on fitting data to theoretical curves that allow computation of mRNA levels. Calculating accurate mRNA levels requires important parameters such as reaction efficiency and the fractional cycle number at threshold (CT) to be used; however, many algorithms currently in use estimate these important parameters. Here we describe an objective method for quantifying qRT-PCR results using calculations based on the kinetics of individual PCR reactions without the need of the standard curve, independent of any assumptions or subjective judgments which allow direct calculation of efficiency and CT. We use a four-parameter logistic model to fit the raw fluorescence data as a function of PCR cycles to identify the exponential phase of the reaction. Next, we use a three-parameter simple exponent model to fit the exponential phase using an iterative nonlinear regression algorithm. Within the exponential portion of the curve, our technique automatically identifies candidate regression values using the P-value of regression and then uses a weighted average to compute a final efficiency for quantification. For CT determination, we chose the first positive second derivative maximum from the logistic model. This algorithm provides an objective and noise-resistant method for quantification of qRT-PCR results that is independent of the specific equipment used to perform PCR reactions.
Quantitative real-time polymerase chain reaction (qRT-PCR) provides a rapid and sensitive quantification of transcript levels that has essentially replaced previous techniques including Northern blots and RNase protection assays (Bustin, 2000). Moreover, qRT-PCR has become the method of choice for validating microarray data in basic research, molecular medicine, and biotechnology (Bustin, 2002; Freeman et al., 1999; Ginzinger, 2002; Klein, 2002). Given its importance in biomedical research, calculating the correct transcript levels from the raw data is essential.
Three conditions must be met for accurate real-time PCR quantification: 1) fluorescent amplicon label intensity must be proportional to its concentration; 2) amplification efficiencies of all samples must be similar to allow comparisons across samples; and 3) amplification threshold for quantification must be set within the exponential phase (EP) of the PCR.
Reagents have been developed (Bustin, 2000, 2002) that can produce fluorescent signals proportional to the quantity of the amplicon over an appropriate range (Higuchi et al., 1993). Similarly, improved primer design and enzymes allow PCR efficiencies of different samples to be approximately similar, although this is not guaranteed. The threshold used to compute the CT must be within the EP so it reflects initial template differences rather than just a change in reaction kinetics (Kainz, 2000; Peirson et al., 2003). Moreover, it is essential that the same threshold fluorescence be set for all samples so that a comparable CT or crossing point (CP) is measured equally for each sample. However, in many current real-time PCR systems, choosing the threshold depends on subjective judgment.
The original method for relative quantification was to assume an ideal amplification efficiency (100%) for all samples and then to compare directly exponential curves (e.g., 2 to the power of ΔCT (2ΔCT) (Livak and Schmittgen, 2001). This method was subsequently modified to include calculation of PCR efficiency from standard curves made using plasmids or pooled sample cDNA (Pfaffl, 2001; Rutledge and Cote, 2003; Stahlberg et al., 2003). However, creating a standard curve is time consuming and requires production of repeatable and reliable standards (Pfaffl, 2001). Moreover, the standard curve method requires 1) that the starting concentration of the standards are accurate; 2) that the efficiency of amplification for each sample is constant, which can rarely be achieved or verified in real experiments; and 3) that there are no errors from contamination, sample dilution, or variable competitive effects due to template concentration differences.
Since the shape of the exponential curve of the raw PCR fluorescence data (Peccoud and Jacob, 1998) contains information about amplification efficiency, computing efficiency from the kinetics of individual PCR reactions is theoretically possible. Other methods have approached this problem including use of the amplification plot (Liu and Saint, 2002a), the first derivative maximum of sigmoid model fitting (Liu and Saint, 2002b; Tichopad et al., 2002), the interactive window-of-linearity algorithm (Ramakers et al., 2003), the mid-value point regression (Peirson et al., 2003), the studentized residual statistics followed by four-parameter logistic model fitting algorithm (FPLM) (Tichopad et al., 2003), and the kinetic outlier detection (KOD) method (Bar et al., 2003). Only the last two methods used statistical techniques to estimate the efficiency objectively. However, both assumed that the baseline can be calculated accurately from only the ground cycles prior to amplification onset, which is often incorrect due to noise in the signal as the PCR reaction begins. Moreover, none of these algorithms address the issue of the noise in the fluorescent signal during exponential amplification, which influences the estimation of efficiency. Finally, the KOD method does not objectively automate the efficiency determination for the so-called “training samples,” which is critical for later efficiency estimations for real samples.
Here we describe a novel, objective, and noise-resistant algorithm to calculate the efficiency and CT for qRT-PCR from individual PCR reactions. This algorithm (Real-time PCR Miner) uses strictly objective criteria, automates all calculations, and is independent of the real-time PCR platform. Miner is available online for processing qRT-PCR (http://www.miner.ewindup.info/).
We used tissue from an African cichlid fish, Astatotilapia (Haplochromis) burtoni, bred from wild-caught stock (Fernald, 1977; Fernald and Hirata, 1977) and raised in laboratory aquaria. Animals were maintained under conditions that mimicked those of the natural habitat (27°C; 12:12 light/dark cycle with full spectrum lights; pH 7.6–8.0) and were fed daily (Wardleys, Secaucus, NJ) (Fernald, 1977). All RNA used for qRT-PCR was prepared from the retina and all fish used in this study were approximately 3 cm (standard length) with body weight approximately 1 g. Fish were sacrificed by rapid cervical transection and the eyes collected for analysis. All procedures were in accordance with the National Institutes of Health protocol for animal experimentation and approved by the Animal Care and Use Committee of Stanford University.
Fish were moved from their home tanks and acclimated in new tanks for three days in a room that allowed entry around the clock. Fish were sacrificed every three hours for 24 hours, the eyes isolated and the sclera, cornea, and lens removed. The two retinas from each animal with the pigment epithelium cell layer attached were placed in 1 ml Trizol (Invitrogen, Carlsbad, CA) followed by 250 µl chloroform (Sigma, St. Louis, MO) to isolate RNA. Total RNA was precipitated by adding 500 µl isopropanol followed by a 75% ethanol washing and air-drying. Five µg total RNA of each sample was used for reverse transcription by Superscript II reverse transcriptase with random primers (Invitrogen).
Based on A. burtoni sequences, primers for proliferating cell nuclear antigen (PCNA), actin, glycer-aldehyde 3-phosphodehydrogenase (G3PDH), and 18S rRNA were designed to avoid dimers or hairpin template structures, to have similar melting temperatures (ca. 60°C), and to generate amplicons with similar length (ca. 150–200 bps). Each amplicon was found only in single peak in melt curves (MyIQ software, v1.04, Bio-Rad Laboratories, Hercules, CA), indicating no dimer or multiple products. More details for the primer sequences are described in Table 1.
The qRT-PCR was performed using 30 µl triplicate reactions with 1 × IQ SYBR Green Supermix (Bio-Rad Laboratories, Hercules, CA), 0.5 µM of each primer, and 2.5 ng/µl cDNA (RNA equivalent) for each experimental time point and each gene using the MyIQ Single-Color Real-Time PCR Detection System (Bio-Rad Laboratories). PCR parameters were: 5 minutes at 95°C, 40 cycles of 30 seconds at 95°C, 30 seconds at 60°C, and 30 seconds at 72°C, followed by melt curve analysis. We detected the fluorescence at 490 nm at the start of the annealing step (60°C) in each cycle. Blank reactions without DNA template did not show any fluorescent signal above background levels. Dilutions from pooled samples (25.0, 12.5, 6.25, 2.5, 1.25, 0.625, and 0.25 ng/µl cDNA-RNA equivalent) were used to construct a standard curve for comparison. To minimize the influence of the variable pipetting error for dilutions and the variable competitive effects due to the differences in initial template concentration, we chose a narrow dilution range (only 1 to 100). To minimize the differences of the PCR conditions for the standards across genes, we used the same range of serial dilutions for all genes. To minimize the differences between the experimental samples and the standard curve, for each gene, we processed these together on the same plate.
For the standard curve method, both efficiency and CT were calculated from the “baseline subtracted curve fitting data” computed in MyIQ software (Taqman-Std) or the first positive second derivative maximal (SDM) of FPLM (SDM-Std). For our algorithm, we exported the so-called “background subtracted data” provided in the MyIQ software, which are actually the raw fluorescence values as a function of cycles, into our program (Real-time PCR Miner). The term “background” here, as defined by the manufacturer, refers to the background fluorescence read from outside of each well, rather than the baseline or ground fluorescence of each reaction before the amplification. It does not refer to the auto-fluorescence value of the reaction, but rather the auto-fluorescence of the plate. By subtracting this “background” from the fluorescence reading of the inside of the reaction well, the resulting data are the raw fluorescence data of the PCR reactions.
All curve fitting was done using SigmaPlot 8.0 (Systat Software, Point Richmond, CA). Raw data of the time sequence fluorescence values were imported from the real time PCR machine into Miner as described above to calculate efficiency, CT, and associated standard errors or deviations as well as the mean coefficient variation (CV%). Tests on platforms other than MyIQ used data from colleagues without any information about experimental design.
Our algorithm is implemented in stages as illustrated in Fig. 1: First we fit a mathematically defined “S” shaped curve to each PCR reaction and use this to identify the exponential phase (EP) of the PCR reaction. We use the noise level of the ground phase before amplification to determine the starting point of the EP (SPE), and the first positive second derivative maximum (SDM) of the logistic model as the end point of the EP (EPE). Second, we estimate efficiency by using an iterative nonlinear regression followed by the weighted average to fit the appropriate curve to the raw data (left branch, Fig. 1). Third, we determine CT by the SDM (right branch, Fig. 1).
To find a suitable mathematical representation of the complete kinetic curve for each PCR reaction, we compared several equations that generate S-shaped curves that are widely used for fitting enzymatic reactions: four-parameter logistic (Tichopad et al., 2003), sigmoid (Liu and Saint, 2002b; Tichopad et al., 2002, 2004), Gompertz (Schlereth et al., 1998), and Chapman models (Glover et al., 1997) (see Table 2). Figure 2A is an example for one typical sample where y0 is the ground fluorescence value and a is the difference between maximum and ground fluorescence values. We tested how well each model fit whole kinetic curves using 1,200 samples.
Although all models fit the entire extent of the raw fluorescence curve quite well (R2 > 0.999 on average, column 1, Table 3), they did not all fit the exponential phase equally well as seen in Fig. 2A. By comparing the mean square errors (MSE) in the EP of all S-shaped models, we found that the four-parameter logistic model (Logistic) fits the EP most closely in all cases (see below).
To identify where the EP of the fluorescence curve begins, we used the standard error of the ground fluorescence signal (y0) as the noise level of the ground phase (RNoise) when calculating the regression over the whole curve (f (x)). After amplification begins, the fluorescence value is above the noise level so the end point of the ground phase, which is the starting point for the EP (CP(SPE)), can be calculated by solving the function f (x) using RNoise (Table 2). Note that y0 here is not the baseline of the EP, but the base fluorescence level of the reaction, although these two values are typically close to each other. Previously, a studentized residual statistics model (Tichopad et al., 2003) has been suggested for SPE calculation by using the first outlier from the baseline (Outlier-SPE). We compared samples (n > 150) for eight platforms and found that Outlier-SPE usually identifies a later SPE (e.g., overestimated baseline) than our noise based SPE (Noise-SPE), resulting in a smaller EP on most popular PCR platforms (Table 4). When there are higher noise fluctuations in the ground phase, the Outlier-SPE will disrupt later calculations (Fig. 2B as an example). Also, the Outlier-SPE method is less stable than the Noise-SPE method (larger standard deviation) because either over- or underestimation of the baseline will result in variance of the CP(SPE).
To fit the exponential part of the curve, optimally, we need to know exactly where the exponential phase of the PCR ends. Previous work suggested that the first positive second derivative maximum CP(SDM) of function f (x) in either of the sigmoid or logistic models could be used to approximate the EPE (Tichopad et al., 2003, 2004). CP (SDM) values can be computed for each S-shaped model by setting the third derivative maximum to zero (Table 2).
To evaluate which S-shape model is best for EP determination, we compared the mean square errors (MSE) within the EP for each model using the corresponding CP(SPE) and CP(SDM) values. The results from 1,200 samples showed that the logistic model produced much lower mean MSEs than other S-shaped models (Table 3). The Gompertz and Chapman models did not match the kinetics of the PCR reaction very well, especially for the EP (high mean MSE, second column in Table 3) producing a very late start point for the EP (Fig. 2A). The sigmoid model is a point symmetric S-shaped model, and since the kinetics of the PCR reactions do not match the point symmetric rule this method overestimated EP (~10 cycles on average) resulting in a high mean MSE (second column in Table 3). The logistic model produced a window with a reasonable number of amplification cycles (7–8 on average) with the lowest mean square errors (second column in Table 3).
To test whether the identified EPs are exponential, we fitted the three-parameter simple exponent model (Fig. 3B, Equation (2)) to the 1,200 windows as defined by each S-shaped model. Based on the averaged MSE differences, the logistic model produced the best fit with very low MSE. The Gompertz model produced somewhat smaller MSE values, but only because 3–4 cycles are used for the three-parameter simple exponent model regression (column 4 in Table 3). Based on these comparisons, we chose the FPLM for the whole kinetic curve fitting and EP determination.
To estimate efficiency, existing algorithms use either a linear (Peirson et al., 2003; Ramakers et al., 2003) or nonlinear regression (Tichopad et al., 2003) for the points found in the EP (Fig. 3). Comparing the corresponding efficiency with the one derived from the standard curve (Table 5 and Fig. 4, and also see below), we found that a nonlinear regression fits better than a linear regression because the over-or under-baseline subtraction (OBS or UBS; cf. Fig. 2C as an example [Bar et al., 2003]) produces a miscalculation of the efficiency when performing baseline subtraction for linear regression.
Although the nonlinear regression provided a better fit than did linear regression to the EP, noise can influence the whole curve fitting, identification of the EP, and the variance of the fluorescence value for every cycle. To minimize this influence, we developed a noise-resistant method using iterative nonlinear regression followed by weighted average analysis. First, we defined all possible windows within the exponential phase containing at least four cycles. For each window, a nonlinear regression is performed to produce a candidate efficiency (Ei) as well as a P-value (Pi) of the regression. The P-value of the regression represents the probability of being wrong in concluding that there is an association between the dependent (fluorescence value) and independent (cycle) variables based on the regression equation. The smaller the P-value, the greater the probability is that there is an association. Traditionally, one can conclude that the independent variable can be used to predict the dependent variable when the P-value < 0.05 (SPSS, 2002). We first discard the efficiencies with a P-value ≥ 0.05. Then, we used the equation below to compute a relative weighting factor (wi) for the candidate efficiencies
We computed the weighted average for all candidate efficiencies whose P-value < 0.05 by Equation (2) as the final efficiency.
We compared this noise-resistant regression method (with weighted values) with the single regression method (without weighted values) using datasets for both linear and nonlinear regression without any visual correction. In the case of nonlinear regression, we also tested the different methods for SPE determination, Outlier-SPE versus Noise-SPE (Table 5 and Fig. 4) using the standard curve data as a reference. We performed an unpaired two-tailed student t-test assuming unequal variance to test whether there are significant differences (P (t) < 0.05) between the standard curve method and other methods. Among all tested data, the weighted average method using the EP identified by Noise-SPE (Noise-SPE-weight) produced very similar efficiency to that computed by the standard curve method (all P (t) > 0.13). The corresponding standard deviations for the Noise-SPE-weight method were also usually smaller than that of others (Table 5). Moreover, in all regression methods, whenever performing weighted averaging, the resulting efficiency will be closer to the referenced efficiency and will usually have a larger t-test value. In addition, points in the smaller EP found by the Outlier-SPE (Table 4) were all included in the Noise-SPE-Weight method. In some cases, the corresponding P-value of regression for such a small window was even higher than 0.05 and would be excluded in the Noise-SPE-Weight method, but not in the Outlier-SPE-No-Weight (Tichopad’s) method (data not shown).
To identify the optimal method for calculating CT, we compared the Taqman threshold method (MyIQ software) (Holland et al., 1991), the first derivative maximum (FDM) method (Tichopad et al., 2004), the second derivative maximum (SDM) method (Tichopad et al., 2003, 2004; Wittwer et al., 2001), and the mid-value point method using known serial dilutions (Table 6 and Fig. 5). The equations for computing CP(FDM) and CP(SDM) are listed in Table 2. The position of the mid-value point (CP(MP)) in the EP can be determined by Equation (3):
The mean square errors (MSE) of each method calculated from the known serial dilutions showed that the SDM method is the most accurate method since it produced the lowest MSE. Also, the SDM method provided a very good estimate among replicates with a lower mean CV% for CT than did the Taqman threshold and mid-value method. The FDM method results in an even smaller mean CV%; however, this is because the FDM values are already out of the EP where all kinetic curves tend to be saturated and hence merge (Fig. 5A). Furthermore, CT determination is extremely critical for the standard curve method, which is imperative to the computation of efficiency. The efficiency calculated from the CTs based on SDM (SDM-Std) is more stable (smaller standard deviation) than that from the CTs based on the Taqman threshold (Taqman-Std) (Table 7). The similar SDM method is also used in Roche LightCycler and Corbett Research’s Rotor-gene Real-time PCR machines.
Once efficiency and CT have been computed, we used these values for quantification (Fig. 1). The starting fluorescence (R0) of each sample is proportional to the starting template quantity (Liu and Saint, 2002a). We compared the results from the standard curve method and those from our algorithm by examining the daily rhythm in PCNA expression level which has been previously reported for A. burtoni (Chiu et al., 1995). We used the geometric mean of three control genes (Actin, G3PDH, and 18S rRNA) for accurate normalization (Vandesompele et al., 2002) to minimize errors from potential variance in a single control gene (Schmittgen and Zakrajsek, 2000; Suzuki et al., 2000). Both the new algorithm and standard curve method gave similar results showing a daily rhythm in PCNA mRNA expression level (Fig. 6) consistent with a previous report (Chiu et al., 1995). Our algorithm produced efficiencies similar to those of the standard curve method but showed consistently smaller standard deviations for efficiency estimation and mean coefficient of variance for CT (Table 7).
This method has been widely tested with thousands of samples processed in different real-time PCR systems (Bio-Rad MyIQ, iCycler, ABI PRISM™ 7700, 7900, Stratagene MX3000, MX4000, MJ Research DNA Engine 2 Opticon, and Roche LightCycler) and produces accurate quantification results (data from other real-time PCR platforms are not shown).
The value of qRT-PCR depends critically on appropriate analysis and interpretation of the outcome of cDNA amplification. Important parameters for this analysis include 1) the magnitude of the noise due to stochastic properties of PCR amplification and fluorescence detection; 2) the identification and selection of the EP; 3) the calculation of the amplification efficiency; 4) the selection of the threshold for CT determination; and 5) the choice of whether to use individual or average efficiencies for data analysis. The method presented here, based on quantitative assessment of the kinetics of individual reactions, offers an accurate and convenient method to address all of these issues and eliminates the need for estimates or qualitative judgments often embedded in other methods.
Noise, or stochastic variations in the detected fluorescence level, exists in all qRT-PCR reactions. In the initial cycles, because the fluorescence is low, the influence of the noise will be more pronounced. Typical quantification procedures fit a straight line to the signal before amplification to be used as the baseline. In software supplied by most real time PCR machines (e.g., MyIQ, iCycler from BioRad, ABI systems, Stratagene MX systems, etc.), the baseline is taken as a straight line fitted to data from the first few cycles, typically 10 cycles or a value arbitrarily chosen by the user. However, at high concentrations of starting template, PCR product can be amplified earlier than the cycles used by such algorithms. Moreover, linear regression algorithms (Peirson et al., 2003; Ramakers et al., 2003) require transformation of the raw fluorescence data from linear to logarithmic form using subtraction of the baseline. Nevertheless, since the baseline is usually calculated only from the initial or ground phase before amplification begins, it is not reasonable to subtract all later cycles from this value since this violates the basis of linear regression. Tichopad et al. recommended nonlinear regression using parameter y0 of the three-parameter simple exponent model as the baseline, a suggestion which avoided the violation of assumptions needed for linear analysis. However, the method they used for SPE determination, Outlier-SPE (Tichopad et al., 2003) is still based on the assumption that a perfect baseline can be fit by the points only from the ground phase. Note that the SPE is not the baseline of the EP, but the first detectable point in the EP Since noise has maximal influence on the ground phase, systematic errors were introduced (Table 4 and Table 5, Fig. 2B and Fig 4). In addition, the variable noise levels of different platforms will also result in variance in outlier detection (Table 4). The fitted baseline usually has a positive slope due to the slight fluorescence increase of the ground phase. Because of the influence of noise, the baseline is often overestimated. In the KOD method, the authors subtracted a baseline that was the arithmetic average of the five lowest fluorescence readings (slope = 0), meaning that the baseline value is more likely to be underestimated as the authors reported (Bar et al., 2003). Although the authors claimed that in typical experiments OBS and UBS can be visualized, small OBS or UBS is actually very hard to see since the value of the baseline is low. Instead of using only the ground phase, we used the noise level of the ground phase (RNoise) based on all points (whole curve fitting) to determine the SPE. This ensures that any amplification cycles used for efficiency estimation are not drawn from points that have low fluorescence readings and occur within the noise level.
Another critical step is choosing the optimal part of the PCR amplification phase for calculation of efficiency. Liu and Saint have proposed an amplification plot method to address this problem, but their method does not provide an objective procedure for determining the EP before performing nonlinear regression (Liu and Saint, 2002a). Another group chose the “window-of-linearity” method (an iterative linear regression algorithm) in which they set the start and end points of the EP subjectively and search for a line in a logarithm plot with the highest R2 value and a slope close to the maximum slope (Ramakers et al., 2003). Peirson et al. suggested yet another linear regression method using the mid-value point in the logarithmic plot, emphasizing that the points chosen for regression should be equally distributed around the mid-value point to achieve the highest accuracy (Peirson et al., 2003). However, visual inspection of all amplification curves is required to choose suitable windows for EP for these methods. In Peirson’s method, as in the MyIQ software, the standard deviation of cycles 1 to 10 is used as RNoise after performing baseline subtraction, which may not be reasonable for all reactions. Moreover, due to the influence of noise, not all points that are equally distributed on both sides of the mid-value point are suitable to be used for the efficiency estimate. Also, Peirson et al. used the maximal fluorescence of the entire curve to calculate the mid-value point, which is actually the middle value of the whole curve and much higher than that of the exponential phase. As described above, we use a method based on the logistic model which is entirely objective (Noise-SPE and SDM) to identify the EP.
After defining the EP, the efficiency is calculated. Although the methods provided by Liu and Saint (2002b) and Rutledge (2004) suggested that the R0 can be calculated directly from the fitted sigmoid model instead of using efficiency and CT, the intrinsic mathematical calculation relies on how well the sigmoid model is matched with the PCR kinetics. As noted above, the point symmetric sigmoid model does not fit the curve of the overall reaction, especially for the EP. Even using an improved S-shaped model, such as the logistic model, we still found a much higher MSE within the EP, compared with the three-parameter simple exponent model, the theoretical curve of PCR kinetics (>20 times, Table 3). Generally, a suitable whole S-shaped curve fitting can do a very good job of defining the exponential phase, since it can accurately predict the cycles on both sides of the exponential phase and plateau phase (Fig. 2A) where the fluorescence changes are less dramatic (lower MSE, data not shown). The whole S-shaped curve fitting fails, however, to be fit to the cycle regions of most rapid change (e.g., the exponential and plateau phases, Fig. 2A) with high enough fidelity (larger MSE than the exponent model, Table 3).
To calculate the efficiency, we chose the three-parameter simple exponential nonlinear regression, also proposed by others (Tichopad et al., 2003). However, in a PCR reaction, the EP occurs in a very early part of the amplification, meaning there are relatively lower fluorescence levels, and is therefore more influenced by noise as discussed above. In our method, we used the P-value of the regression to control the contribution (the weight) of the candidate efficiency to the final efficiency estimation. In practice, we found that with greater noise, fewer candidate efficiencies have a P-value less than 0.05. Typically, the EP spans ca. 8 cycles and the iterative regression will find ca. 15 windows containing 4–8 cycles. The method will then calculate all the candidate efficiencies and their related P-values. In contrast, Tichopad’s method (Tichopad et al., 2003) performs only one regression based on all points found in the EP. Our approach makes the whole algorithm very robust over thousands of tested samples across different platforms.
As another important parameter for the qRT-PCR, the CT values should be determined within the EP to reflect the initial concentration of the template. Currently, there are several different methods for estimating CT: the fit point, Taqman threshold, SDM (Tichopad et al., 2002, 2003, 2004; Wittwer et al., 2001), and FDM (Tichopad et al., 2004) methods. Since the mid-value point is also located within the exponential phase, it potentially can be used for objective CT determination.
In the fit point method, an intersecting line is placed arbitrarily in a logarithm plot at the base of the exponential portion of the amplification curves. This method can result in systematic errors due to the baseline subtraction and subjective judgment. Another problem for this method is that all samples should come from the same experimental plate so that the unique threshold can be set across all samples. This constrains the number of samples that can be reliably compared.
The Taqman threshold method refines the fit point method by fitting a line at 10 times the standard deviation of the fluorescence in the ground phase (Holland et al., 1991). However, using 10 times the value is an arbitrary choice and does not guarantee that the CT will be in the exponential phase.
Since the FDM and SDM as well as the mid-value point methods calculate the CT from an individual sample based on its own kinetics, they can potentially be used for cross-plate analyses as long as the noise levels across plates are similar. Nevertheless, the FDM value is usually not in EP, while the fluorescence at the mid-value point is relatively small and more easily influenced by noise (Fig. 5A). In practice, we found that the CT for the samples with extremely low concentrations of initial template (e.g., CT > 32 in a reaction with a total of 40 cycles) are usually much less accurate than others, which can be easily discovered by much larger differences among replicates. In some cases, the corresponding reaction curves do not even reach the SDM before the last cycle, although the logistic regression might still give a number for SDM (underestimated). Without providing enough accurate information about the reaction, these data should be excluded no matter what kind of postmathematical computation (standard curve or Miner) is chosen, unless one is introducing more templates, optimizing the experimental condition, or increasing the total cycles carefully if the variance among replicates is acceptable.
Since knowledge of amplification efficiency is critical for accurate real-time PCR quantification, using the mean efficiencies of all samples for each gene is still recommended (Tichopad et al., 2004). Applying individual corrections can result in potential systematic error because only a small number (~8) of available data points within EP can be used for the individual efficiency calculation. Different efficiencies from only triplicate samples might result in a considerable effect on R0 because any error in the measured efficiencies will be exponentially magnified (Peirson et al., 2003). Alternatively, when the differences of inhibitors (hemoglobin, heparin, glycogen, fats, Ca2+, etc.) or enhancers (glycerol, BSA, gene 32 protein, Taq extender, E. coli ssDNA binding protein, etc.) of RT-PCR among samples need to be considered in a particular experiment (Rossen et al., 1992; Tichopad et al., 2004; Wilson, 1997), using more replicates to calculate the mean of efficiency for a different experimental group to acquire a comparable R0 is advised. On the other hand, the individual efficiency allows additional investigation of the quality of each reaction.
In summary, the algorithm described here uses the kinetics of individual reactions for accurate estimates of efficiency and CT without the need for preparing a standard curve. Furthermore, this method allows all the key parameters for the quantification procedure to be objectively estimated, which is especially convenient for beginning users and for large sample sizes. It is hence economical for qRT-PCR analysis and robust for samples with high noise levels across real-time platforms.
Thanks to V Parikh, M. Corty, T. Au, R. Henderson, and Drs. S. Burmeister, S. Halstenberg, and L. Harbott for useful discussions and to colleagues who provided the data from other platforms for testing. Supported by NIH-NEI EY05051 to RDF.