Results

The inference method was performed on both experimental and in-silico simulated data. Below we discuss the performance of the inference by looking at indicative plots such as joint and marginal posteriors, trace plots or Bayesian credible intervals.

Experimental data

The experimental data consists of qPCR amplification curves in the form of arrays of fluorescence reads. Each amplification curve represents the observation data used to infer the initial quantity of a target analyte X0 (number of mtDNA molecules in a single cell). The qPCR experiments are performed in parallel, with the contents of each target analyte being placed in a separate well.

Each amplification curve from experimental data has a corresponding X0 estimate obtained using the standard curve method. This value is used as an additional indicator for whether the inferred posterior distribution for X0 is centered around to the true value of the parameter. However, this value obtained from the standard curve method is simply an estimate, and should not be treated as the true X0.

In-silico simulation data

We perform in-silico simulations to obtain an amplification curve given a parametrisation θ = [X0, r, σ], a known fluorescence coefficient α (which is not inferred) and a number of qPCR cycles n. The simulated process starts with X=X0 molecules, and is updated by sampling from a Binomial distribution with probability r. The fluorescence reads F1:n are generated by adding noise sampled from a Gaussian distribution with variance σ (see Model).

The advantage of the in-silico data is that it allows us to compare the means and posterior distributions of all parameters X0, r, σ with the true values using simulated data F1:n.

Trade-off between X0 and the efficiency r

In the first instance, inference was performed on data from a single amplification curve. As discussed in Inference, the exponential phase is extracted from the amplification curve and only those fluorescence reads are used as observation data. For the first ~30 cycles, the background noise dominates over the fluorescence signal, which leaves only data from ~4-5 cycles being informative for the inference of X0 and r, before significant saturation occurs and the assumption of r being a fixed parameter is clearly unreasonable.

We found that this lack of information caused degeneracy in the inference of X0 and r by inspecting the joint posterior of the two parameters (see below). The explanation for this phenomenon is that for a short fluorescence curve, a similar fluorescence curve can be obtained by increasing the efficiency r and lowering the quantity of the target analyte X0 (and vice-versa). However, with more data from the "take-off" part of the curve, there would be more information to distinguish between X0 and r.

The solution to this problem was to perform parameter-pooling inference using the data from multiple wells, with the assumption that parameters r and σ are shared between different wells within the same experimental setting. The new parameter for inference became θ = [X01, X02, X03, r, σ] for joint inference on 3 wells. Using data from multiple fluorescence curves where r is constant results in a more accurate inference of r, which in turn restricts the X0j parameters. We see that the width of the marginals is strongly constrained through this alternative model structure.


Figure 23. Parameters X0 and r can be traded off against each other (first plot), causing underdetermination in the marginals resulting from single-experiment inference (Fig. 25). When parameter pooling is used instead (second plot), the samples for X0 and r are independent, which in turn gives more restricted (determined) marginal posteriors.

Marginal posterior distributions from real data

We first look at results for the data obtained from real qPCR experiments. The goal of the inference method is to perform sampling from the joint posterior distribution P(θ | F1:n, α). The fluorescence coefficient α relates fluorescence intensity to molecule copy number and is determined prior to the inference, using sigmoid fitting on samples where the initial copy number X0 is known. Sigmoid fitting was used instead of inference because these samples have large X0 values, causing the process to behave deterministically (more details in Model). The inference for P(θ | F1:n, α) produces a sequence of samples of θ that approximates the joint distribution. From this sequence we can extract samples for each component of θ (marginals) and plot the marginalised posterior distributions.

It should be noted that posteriors shown here are merely indicative due to the analysis of chain mixing, which indicates that inference needs to be run for longer (see Analysing the Markov chain).

The marginal posterior distributions are easy to visualize and encode all of our uncertainty in the parameter given the data. For the target analyte quantity X0 we additionally plot the estimate from the standard curve method on top of the posterior distribution. The plot in Fig. 24 shows the posterior distributions obtained from joint inference using target analytes X01, X02, X03 from 3 separate qPCR wells (253, 254, 255).

Figure 24. Marginal posteriors of copy numbers X0i for different qPCR experiments (wells) are centered close to the values indicated by the standard curve method. However, this standard curve estimate should not be treated as the real parameter value.

The plots in Fig. 25 below demonstrate the effect of additional fluorescence data on the marginal posteriors by showing in parallel the results from single-well and joint inference. It is visible how the spread of posteriors narrows with joint inference, better localising the parameter values. Additionally, the prior used for each parameter is plotted: uniform prior for X0, Jeffreys priors for r and σ (the choice of priors discussed in Priors).



Figure 25. Marginal posteriors of parameters X0, r, σ2 shown in parallel for the single-experiment and parameter-pooling methods show how additional fluorescence information solves the trade-off problem between X0 and r and narrows the marginals. The parameter pooling method shows a significant improvement in the error bar on estimates, and therefore in the credibility of estimates obtained from inference.

Marginal posteriors distributions from simulated data

In order to evaluate the inference method, we simulate a qPCR experiment in-silico and generate fluorescence data. We do so by choosing a true parameter to be the mean θ inferred from real data. An amplification curve is generated according to the random branching process model, using this fixed θ (for more details, see Model). For brevity, we only look at results from the joint inference, as it has been shown in the previous section to remove parametric degeneracy.

Testing the inference algorithm on in-silico simulated data allows the marginal posterior to be compared against a ground truth, the true parameter θ used to simulate the data, in order to check that the inference method works.


Figure 26. The marginal posteriors show that there is appreciable uncertainty in the quantity we are interested in, namely X0. This result motivates our model, as it is something not taken into account by the simpler standard curve method.

Stochasticity and uncertainty in fluorescence reads

A main advantage of the Bayesian inference approach in qPCR absolute quantitation is that it accounts for both the stochasticity of the amplification process, by modelling it as a random branching process with efficiency r, as well as for parametric uncertainty. In this section, we show how stochasticity and parametric uncertainty result in a range of plausible fluorescence reads.

The inference algorithm produces a sequence of θ samples that approximate sampling from the posterior distribution of θ. A qPCR experiment was simulated in-silico for each sample, resulting in a fluorescence read sequence. A 5%-95% confidence interval for the fluorescence reads across all simulated experiments was determined at each cycle. The plotted area (Fig. 27) shows the confidence interval for the fluorescence data (green). Additionally, the plot shows the real fluorescence data used for inference (red), as well as the mean fluorescence value at each cycle (blue).

Figure 27. The fluorescence curve has many plausible values due to the stochasticity inherent to the qPCR amplification and parameters not being known precisely (parametric uncertainty).

The effect of the fluorescence coefficient α on inference results

The inter- and intra-experimental variability of the fluorescence coefficient α were discussed in the Model section and are showed again (Fig. 28). Estimates for α were obtained by fitting sigmoid curves on data from standards with known initial quantities X0. Sigmoid fitting was used instead of inference because these standards had a large copy number X0, causing the amplification process to behave deterministically. It has been discussed that the inter-experimental variability in α is explained by the use of different primers (experiments 3 and 4 below use the same primer and have a similar distribution). Thus the α estimate can be used for inference on a new sample, if the sample uses the same primer as the standards.

Figure 28. The fluorescence coefficient α estimates show significant inter-experimental variability when different primers are used. When the same primer is used, the distribution of α is very similar, although experiments were performed on different days. More work is needed to formally establish this result.

We are now interested in how different values of α from the same qPCR experimental setting influence the posterior distribution of X0, and how the initial copy number of the standard used for sigmoid fitting may play a role in the shape of the marginal posterior.

The experiments used to estimate α have the following initial standard quantities: 3 wells with 100,000 molecules, 3 with 10,000 and 3 with 1,000. A total of 9 α estimates were obtained. The plots below (Fig. 29) show posteriors from inference runs with different α values, grouped by the initial number of molecules in the curve used to determine α. The inference results plotted in other sections of this page all use the mean value of α from estimates obtained with sigmoid-fitting.

Figure 29. The larger variance of the posterior in the third plot can be explained by the deterministic behaviour assumption in sigmoid-fitting not holding when the standard inital copy number is too small, X0 = 1,000 (details about role of copy number in stochasticity discussed in Choice of model). It is not evident why the marginal posterior is larger when the standard copy number is X0 and further work is needed to explain this result.

Analysing the Markov chain

A Metropolis-Hastings algorithm produces a Markov chain of samples with the joint posterior P(θ | F1:n, α) as its stationary distribution. The calibration of the covariance Σ of the proposal distribution Q(θ’|θ) (see General Metropolis-Hastings) is essential for producing a chain with the desired properties. Two common ways of assessing the inference method and the covariance Σ are traceplots and auto-correlation functions (ACF). We first explain the meaning of traceplots by looking at different types of Markov chains, then show the traceplots for our inference.

Background

Three types of mixing behaviour may occur:

  • Chain is too cold — occurs when the covariance is too small, causing the samples to be correlated. The samples appear like a continuous trace on the joint posterior. The traceplot has fluctuations on the macroscopic level and the ACF fails to drop to zero (Fig. 30a).
  • Chain is too hot — occurs when the covariance is too large, causing bad proposals and a low acceptance rate. The samples spend a long time in the same position on the joint posterior. The trace plot has a "Manhattan skyline effect" (Fig. 30b).
  • Chain is well mixed — occurs when the covariance is appropriate, causing samples to appear independent and identically distributed when plotted onto the posterior. The trace plot has the "caterpillar effect" and the ACF quickly drops to zero (Fig. 30c).

Cold chain
Cold chain
Figure 30a. Example cold chain trace plot and joint posterior plot generated using a simple implementation of MH, used to fit two normal distributions with unknown means (the model parameters) and known fixed variance on 2D observation data. Only the trace plot of the first parameter is shown and its true value is plotted as a horizontal line. The chain is cold when the covariance of the proposal is too small, causing samples to be correlated.

Hot chain
Hot chain
Figure 30b. Example hot chain trace plot ("Manhattan skyline") and joint posterior plot using the same algorithm and observation data as in Fig. 30a. The chain is hot when the covariance of the proposal is too large, causing bad proposals to be made and lowering the acceptance rate.

Good chain
Good chain
Figure 30c. Example of well-mixing chain trace plot ("caterpillar plot") and joint posterior showing samples that are independent and identically distributed. The trace plot and joint posteriors shown here are the goals in terms of inference results.

Mixing results

The chains for parameters X0 and r look cold on the macroscopic level (Fig. 31), indicating that the inference may be exploring different regions of the parameter space that are comparable in terms of their log posterior value. Inference was run for only 100,000 iterations, with equally spread 10,000 samples shown here. Therefore, the posteriors shown above are only indicative results. Running the simulations for longer may allow chains to fully mix, and should be done as future work.

Figure 31. The trace plots for X0 and r show that the chain is not fully mixed and iterations should be run for longer to allow the chain to fully mix.


Figure 32. The lag on the ACF plots for X0 and r does not immediately drop to 0, showing that the samples are not independent. This indicates that further work is needed to improve the inference method.