Abstract
Inference of interaction networks represented by systems of differential equations is a challenging problem in many scientific disciplines. In the present article, we follow a semi-mechanistic modelling approach based on gradient matching. We investigate the extent to which key factors, including the kinetic model, statistical formulation and numerical methods, impact upon performance at network reconstruction. We emphasize general lessons for computational statisticians when faced with the challenge of model selection, and we assess the accuracy of various alternative paradigms, including recent widely applicable information criteria and different numerical procedures for approximating Bayes factors. We conduct the comparative evaluation with a novel inferential pipeline that systematically disambiguates confounding factors via an ANOVA scheme.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
1 Introduction
A topical and challenging problem for computational statistics and machine learning is to infer the structure of complex systems of interacting units. This research area has been particularly motivated by the cognate research discipline of computational systems biology, where researchers aim to reconstruct the structure of biopathways or regulatory networks from postgenomic data; see e.g. Smolen et al. (2000), De Jong (2002) and Lawrence et al. (2010). Two principled approaches can be distinguished. The first paradigm aims to apply generic models like sparse Lasso-type regression, Bayesian networks, or hierarchical Bayesian models. A recent overview and comparative evaluation was published by Aderhold et al. (2014). The advantage of this approach is that the computational complexity of inference is comparatively low, and the application of these methods to problems of genuine interest is computationally feasible. The disadvantage is that interactions are modelled at a high level of abstraction, which ignores the detailed nature of the underlying mechanisms. The second paradigm is based on mechanistic models and the detailed mathematical description of the underlying interaction processes, typically in the form of ordinary or stochastic differential equations (DEs). Two pioneering examples of this approach were published by Vyshemirsky and Girolami (2008) and Toni et al. (2009). The advantage of this paradigm is a more detailed and faithful mathematical representation of the interactions in the system. The disadvantage is the substantially higher computational costs of inference, which stem from the fact that each parameter adaptation requires a numerical integration of the differential equations. A novel approach, presented by Oates et al. (2014) and termed ’chemical model averaging’ (CheMA), aims for a compromise that combines the strengths of both paradigms. The underlying principle is that of gradient matching, first proposed by Ramsay et al. (2007). Given the concentration time series of some quantities whose interactions are to be inferred, the temporal derivatives of the concentrations are directly estimated from the data. These derivatives are then matched against those predicted from the DEs. Formally, on the assumption that the mismatch can be treated like observational noise of known distributional form, we can derive the likelihood and thus apply standard statistical inference techniques. The model is effectively a non-linear regression model, whose computational complexity of inference sits between the two paradigms discussed above: it is lower than for proper mechanistic models, since the DEs do not have to be integrated numerically; it is higher than for standard models of the first category, since the model is non-linear in its parameters and an analytic marginalization is intractable. Henceforth, we refer to it as a ’semi-mechanistic’ model.
This article takes the work of Oates et al. (2014), which won the best paper award at the European Conference on Computational Biology (ECCB) in 2014, further in four respects, related to accuracy and efficiency, model selection, benchmarking, and application expansion. An overview can be found in Fig. 1.
1.1 Accuracy and efficiency
Robust gradient estimation is absolutely critical for semi-mechanistic modelling. The numerical differentiation proposed in Oates et al. (2014) is known to be susceptible to noise amplification. We here propose the application of Gaussian process (GP) regression and the exploitation of the fact that under fairly general assumptions, GPs are closed under differentiation. Our approach effectively implements a low-pass filter that counteracts the noise amplification of the differentiation step, and we quantify the boost in network reconstruction accuracy that can be achieved in this way. We further critically assess the influence of the parameter prior in the underlying Bayesian hierarchal model. In particular, we compare the g-prior with the ridge regression prior (see e.g. Chapter 3 in Marin and Robert (2007)) in the context of the proposed semi-mechanistic model and demonstrate that the latter significantly improves both accuracy and computational efficiency.
1.2 Model selection
Network reconstruction is effectively based on statistical model selection. The model selection paradigm applied in Oates et al. (2014)—computing the log marginal likelihood (MLL) with Chib’s method—is not uncontroversial. Conceptually, alternatives to the MLL based on predictive performance have been promoted (see e.g. Sect. 7.4 in Gelman et al. (2014a)). Numerically, Chib’s method can give inaccurate results, as discussed e.g. in Murphy (2012), Chapter 24. In this article, we assess four numerical approximation procedures for the MLL in the context of semi-mechanistic models: Chib’s original method (Chib and Jeliazkov 2001), Chib’s method with a numerical stabilization, thermodynamic integration (Friel and Pettitt 2008), and a numerically stabilized version of thermodynamic integration (Friel et al. 2013). We further carry out a comparative evaluation between the MLL and four information criteria (IC) as approximations to the predictive performance paradigm promoted in Gelman et al. (2014a): ’divergence IC’ (DIC), ’widely applicable IC’ (WAIC), ’cross-validation IC’ (CVIC), and ’widely applicable Bayesian IC’ (WBIC).
1.3 Benchmarking
Assessing methodological innovation calls for an objective performance evaluation. We have carried out a comprehensive comparative evaluation of the proposed semi-mechanistic model with 11 state-of-the-art network inference methods from computational statistics and machine learning, based on a realistic stochastic process model of the underlying molecular processes (Guerriero et al. 2012) and six distinct regulatory networks with different degrees of connectivity. The analysis of such a complex simulation study is hampered by the influence of various confounding factors, which tend to blur naive graphical representations. We therefore apply an ANOVA scheme, which enables us to disentangle the various effects and thereby extract clear trends and patterns in the results. In this way we can show that by integrating prior domain knowledge via a system-specific mathematical representation, the resulting semi-mechanistic model can significantly outperform state-of-the-art generic machine learning and computational statistics methods. We provide an application pipeline (Fig. 2) with which a user can objectively quantify this performance gain.
1.4 Application extension
Finally, we adapt the method from the modelling of protein signalling cascades in Oates et al. (2014) to transcriptional gene regulation and include an explicit model of transcriptional time delays. In Appendix 4 we provide a novel application of the proposed semi-mechanistic model to plant systems biology, where the objective is to infer the structure of the key gene regulatory network controlling circadian regulation in Arabidopsis thaliana.
2 Model
2.1 Interaction model
Inference of networks from data has become a topical theme in various scientific disciplines, particularly in systems biology. Here, rather than merely aiming for a descriptive representation of associations, the objective is a quantitative mathematical description of the processes that lead to the formation of an interaction network (e.g. a ‘biopathway’ or a ‘biochemical reaction network’). A standard approach is to model this network with a system of ordinary differential equations (ODEs):
Here, \(i\in \{1,\ldots ,n\}\) denotes one of n components of the system, which is called a ‘node’. In systems biology, this is typically a gene or a protein. The variable \(x_i(t)\) denotes a measurable concentration of node i at time t. This can, for instance, be a gene expression or mRNA concentration. The vector \(\pi _i(t)\) contains the concentrations of the regulators of node i. In a network, the regulators of node i are those nodes with a directed edge (or arrow) pointing to node i. Finally, the differential equations depend on a parameter vector \(\varvec{\theta }\). In systems biology, these parameters are typically reaction rates that determine the kinetics of the underlying reactions. A specific example, taken from Barenco et al. (2006), is given in Sect. 2.3. What makes network inference in the context of such a mechanistic description particularly challenging is the fact that the parameters \(\varvec{\theta }\) are typically not measurable, or that only a small fraction of them can be measured. Hence, the elucidation of the interaction network structure requires these parameters to be inferred from concentration time series, which are typically sparse and noisy. To avoid the computational complexity of numerically solving the ODEs, we follow Oates et al. (2014) and use gradient matching. The idea, first proposed by Ramsay et al. (2007), is to estimate the time derivatives \(\frac{d x_i}{dt}\) directly from the data, then treat the problem as nonlinear regression. On the assumption that the estimated derivatives can be treated like noisy data distributed around the predicted derivatives, and this distribution is iid normal, we obtain for the likelihood:
where \(y_i(t_j)= \frac{d x_i(t)}{dt}|_{t=t_j}\), and \(\mathscr {N}(.|\mu ,\sigma ^2)\) is a normal distribution with mean \(\mu \) and variance \(\sigma ^2\). Oates et al. (2014) obtained the temporal derivatives \(y_i(t_j)\) (\(j=1,\ldots ,T\)) by differencing the time series \(x_i(t_1),\ldots ,x_i(t_T)\), based on the Euler equation. However, differencing is known to lead to noise amplification (see e.g. Chatfield (1989)). In the present work, we apply a GP to smooth interpolation and exploit the fact that GPs are closed under differentiation, i.e. provided the kernel is differentiable, the derivative of a GP is also a GP, and its covariance matrix can be derived (Solak et al. 2002; Holsclaw et al. 2013).Footnote 1 We provide more details in the following section.
2.2 Rate (or gradient) estimation
The fundamental concept of the interaction model is the matching of gradients between the regulator variables on the right-hand side and the rate of mRNA concentration change \(\frac{dx_i(t)}{dt}\) on the left-hand side of Eq. (1). Since direct measurements of these rates are typically missing, we derive a rate estimate from the available concentration measurements at the time points \(t^{\star }\in \{t_1,\ldots ,t_T\}\). A common procedure, which is also used by Oates et al. (2014), is to calculate the slope of the concentration change at each time point \(t^{\star }\) with the finite difference quotients
This numerical procedure can yield good approximations to the true rates of concentration change if the data \(x_i\) is relatively precise, i.e. the signal-to-noise ratio is high. If the data is noisy, however, the rates from the difference quotient are susceptible to distortions as a consequence of noise amplification, as mentioned above. We here propose the application of GP regression to counteract this noise amplification. A GP defines a prior distribution over functions g(.) that transform input data points, defined here as time points \(\mathbf{t}= (t_1, \ldots , t_T)\), into output data points, defined here as the concentration vector \(\mathbf{x}_i= (x_i(t_1), \ldots , x_i(t_T))\) for species i such that \(x_i(t^*) = g(t^*)\). The joint prior distribution over the functions \(p(g(t_1), \ldots , g(t_T))\) is Gaussian distributed and commonly has a zero mean and a covariance matrix \(\mathbf{G}\) with independent and identically distributed (iid) additive noise \(\sigma ^2_n\):
where I is the identity matrix. The key idea of the GP is that the elements \(p, q \in \{1, \ldots , T\}\) of the covariance matrix \(\mathbf{G}\) are calculated from a kernel function with \(G_{pq} = \kappa (t_p, t_q)\), which is typically chosen in such a way that for similar points \(t_p\) and \(t_q\), the corresponding values \(x_i(t_p)\) and \(x_i(t_q)\) are stronger correlated than for dissimilar arguments. Widely used kernel functions that are also applied in this paper are the radial basis function (RBF), the periodic function (PER), or the Matérn class function (MAT); see Chapter 4 in Rasmussen and Williams (2006) for the explicit mathematical expressions. By taking the first derivative of the kernel function \(\kappa '(t_p, t_q)\) we obtain a prior distribution \(p(\mathbf{g})\) over functions that define the first temporal derivative, i.e. the concentration gradients, for each of the time points in \(\mathbf{t}\). Provided the kernel function is differentiable, this is again a valid GP. The simplest approach is to compute the expectation over these functions and thus obtain a mean estimate of the analytical solution for the gradients at each time point. For the explicit mathematical expression, see e.g. Eq. (1) in Holsclaw et al. (2013). This acts as a proxy for the missing rates \(y_i(t)\) on the left-hand side of Eq. (1).Footnote 2 In Appendix 2 we describe the details of the GP application and the software we used.
2.3 Model and prior distributions
Equation (1) typically takes the form (Barenco et al. 2006)
Setting \(c_i=0\) and employing Michaelis–Menten kinetics yields the CheMA approach (Oates et al. 2014):Footnote 3
where the sum is over all species u that are in the set of regulators \(\pi _i\) of species i, and the indicator functions \(I_{u,i}\) indicate whether species u is an activator (\(I_{u,i}=1\)) or inhibitor (\(I_{u,i}=0\)). The first term, \(-v_{0,i} x_i(t^{\star })\), takes the degradation of \(x_i\) into account, while \(v_{u,i}\) and \(k_{u,i}\) are the maximum reaction rate and Michaelis–Menten parameters for the regulatory effect of species \(u\in \pi _i\) on species i, respectively. Equation (6) represents the typical form of transcriptional regulation without complex formation; see e.g. the supplementary material in Pokhilko et al. (2010) and Pokhilko et al. (2012). We discuss the limitations caused by complex formation in Sect. 6.5. Without loss of generality, we now assume that \(\pi _i\) is given by \(\pi _i= \{x_1,\ldots ,x_s\}\). Eq. (6) can then be written in vector notation:
where \({\mathbf{V}}_i=(v_{0,i},v_{1,i}\ldots ,v_{s,i})^{\top }\) is the vector of the maximum reaction rate parameters, and the vector \({\mathbf{D}}_{i,t^{\star }}\) depends on the measured concentrations \(x_u(t^{\star })\) and the Michaelis–Menten parameters \(k_{u,i}\) (\(u\in \pi _i\)) via Eq. (6):
We combine the Michaelis–Menten parameters \(k_{u,i}\) (\(u\in \pi _i\)) in a vector \({\mathbf{K}}_i\), and we arrange the T row vectors \({\mathbf{D}}_{i,t^{\star }}^{\top }\) (\(t^{\star }\in \{t_1,\ldots ,t_T\}\)) in a T-by-\((|\pi _i|+1)\) design matrix \({\mathbf{D}}_i={\mathbf{D}}_i({\mathbf{K}}_i)\). The likelihood is then:
where \({\mathbf{y}}_i:=(y_i(t_1),\ldots ,y_i(t_T))^{\top }\) is the vector of the rates or gradients for species i. To ensure non-negative Michaelis–Menten parameters, truncated Normal prior distributions are used:
where \({\mathbf{1}}\) is a vector of ones, \({\mathbf{I}}\) is the identity matrix, \(\nu >0\) is a hyperparameter, and the subscript, \(\left\{ {\mathbf{K}}_i\ge 0 \right\} \), indicates the truncation condition, i.e. that each element of \({\mathbf{K}}_i\) has to be non-negative. In the original CheMA model (Oates et al. 2014) a truncated g-prior is imposed on the maximum reaction rate vectors \({\mathbf{V}}_i\):
where \({\mathbf{D}}_i={\mathbf{D}}_i({\mathbf{K}}_i)\), and a Jeffrey prior is used for the noise variance: \(p(\sigma _i^2)\propto \sigma _i^{-2}\). In Sect. 6.2 we demonstrate an intrinsic shortcoming of the g-prior, and we show that the model can be significantly improved by employing a truncated ridge regression prior instead:
where \(\delta ^2_i\) is a new hyperparameter which regulates the prior strength. For \(\sigma _i^2\) and \(\delta ^2_i\) we use inverse Gamma priors, \(\sigma _i^{2}\sim IG(a_{\sigma },b_{\sigma })\) and \(\delta ^2_i \sim IG(a_{\delta },b_{\delta })\). Graphical model representations for both models CheMA and iCheMA are provided in Fig. 3. For inference of the iCheMA model the MCMC sampling scheme in Oates et al. (2014) has to be modified. The new full conditional distributions can be derived from the equations in Sect. 3.2 of Marin and Robert (2007). The details are given in Sect. 3.1, and pseudo-code of the MCMC algorithm is provided in Table 1. Table 2 shows that the replacement of Eq. (11) by Eq. (12) yields a substantial reduction of the computational costs of the MCMC scheme. Figure 6 in Sect. 6.2 shows that this replacement can also lead to a significantly improved network reconstruction accuracy.
3 Inference
3.1 Posterior inference
We refer to the proposed new variant of the CheMA model, which employs an analytical rather than a numerical gradient and replaces the truncated g-prior in Eq. (11) by the truncated ridge regression prior in Eq. (12), as the improved CheMA (iCheMA) model. For iCheMA, as outlined in Sect. 2.3, the Metropolis-within-Gibbs Markov Chain Monte Carlo (MCMC) sampling scheme proposed by Oates et al. (2014) has to be modified. In the new variant (iCheMA) we replace the truncated g-prior on \({\mathbf{V }}_i\) by the truncated ridge regression prior, we use a conjugate inverse Gamma prior rather than a Jeffrey’s prior for the noise variance \(\sigma _i^2\), and we introduce a new hyperparameter \(\delta _i^2\). We thus have to revise the sampling steps of the original MCMC inference algorithm. We also replace the approximate collapsed Gibbs sampling step for the noise variance \(\sigma _i^2\) from Oates et al. (2014) by an exact uncollapsed Gibbs sampling step. For computing the posterior distribution of the noise variance \(\sigma _i^2\),
Oates et al. (2014) approximate the marginalization integral
with the closed form solution from Marin and Robert (2007), Chapter 3. This is exact if there are no restrictions on the integration bounds. However, given the underlying positivity constraint for \({\mathbf{V}}_i\), symbolically \(\{{\mathbf{V}}_i\ge 0\}\), the integral is no longer analytically tractable and the expressions for Eqs. (13, 14) used in Oates et al. (2014) become an approximation.Footnote 4 We therefore switch to an uncollapsed Gibbs sampling step, where \(\sigma _i^2\) is sampled from the full conditional distribution \(p(\sigma _i^2|{\mathbf{K}}_i,{\mathbf{V}}_i,{\mathbf{y}}_i)\) and the marginalization integral from Eq. (14) becomes obsolete.Footnote 5
For species i and a given regulator set \(\pi _i\) we have to sample the maximum reaction rate vector \({\mathbf{V }}_i\), the Michaelis–Menten parameter vector \({\mathbf{K}}_i\), the noise variance \(\sigma _{i}^{2}\), and the new hyperparameter \(\delta _{i}^{2}\) from the posterior distribution:
where \({\mathbf{y}}_i:=(y_i(t_1),\ldots ,y_i(t_T))^{\top }\) is the vector of rates or gradients. For the full conditional distribution of \({\mathbf{V}}_i\) we get:
Since \({\mathbf{K}}_i\), \(\sigma _i^2\), and \(\delta _i^2\) are fixed in Eq. (16) and the (truncated) Gaussian prior on \({\mathbf{V}}_i\) from Eq. (12) is conjugate to the likelihood in Eq. (9), we obtain:
where \(\tilde{\varvec{\varSigma }} = \delta _{i}^{2} ({\mathbf{I}} + \delta _{i}^{2} {\mathbf{D}}_i^{\top } {\mathbf{D}}_i)^{-1}\), \(\tilde{\varvec{\mu }} = \tilde{\varvec{\varSigma }} (\delta _{i}^{-2} {\mathbf{1}} + {\mathbf{D}}_i^{\top } {\mathbf{y}}_i )\), and \({\mathbf{D}}_i = {\mathbf{D}}_i({\mathbf{K}}_i)\) is the design matrix, built from the rows given in Eq. (8). For the full conditional distribution of \(\delta _i^2\) we have:
As \({\mathbf{V}}_i\) and \(\sigma _i^2\) are fixed in Eq. (18) and the inverse Gamma prior on \(\delta _i^2\) is conjugate for \(p({\mathbf{V}}_i|\sigma _i^2,\delta _i^2)\), defined in Eq. (12), we obtain:
with \(\tilde{b}_{\delta } = b_{\delta } + \frac{1}{2} \sigma _{i}^{2} ({\mathbf{V }}_i - {\mathbf{1}})^{\top } ({\mathbf{V }}_i - {\mathbf{1}})\), and \(\tilde{a}_{\delta } = a_{\delta } + \frac{1}{2} ((|\pi _i|+1)\). For the full conditional distribution of \(\sigma _i^2\) we have:
As \({\mathbf{K}}_i\), \({\mathbf{V}}_i\), and \(\delta _i^2\) are fixed in Eq. (20) and the Gaussian–Inverse–Gamma prior on \(({\mathbf{V }}_i,\sigma _i^2)\) is conjugate for the likelihood in Eq. (9), we get:
where \(\tilde{b}_{\sigma } = b_{\sigma } + \frac{1}{2} [ ( {\mathbf{y}}_i - {\mathbf{D}}_i{\mathbf{V }}_i)^{\top } ({\mathbf{y}}_i - {\mathbf{D}}_i{\mathbf{V }}_i) + \delta _{i}^{2} ({\mathbf{V }}_i - {\mathbf{1}})^{\top } ({\mathbf{V }}_i - {\mathbf{1}})]\), and \(\tilde{a}_{\sigma } = a_{\sigma } + \frac{1}{2} (T + |\pi _i|+1)\). For the mathematical details see, e.g., Chapter 3 of Marin and Robert (2007).
The full conditional distribution of \({\mathbf{K}}_i\) cannot be computed in closed-form so that the Michaelis–Menten parameters have to be sampled by Metropolis–Hastings (MH) MCMC steps. Given the current vector \({\mathbf{K}}_i\), a realization \({\mathbf{u}}\) from a multivariate Gaussian distribution with expectation vector \({\mathbf{0}}\) and covariance matrix \({\varvec{\varSigma }}=0.1 \cdot {\mathbf{I}}\) is sampled, and we propose the new parameter vector \({\mathbf{K}}_i^{\star }={\mathbf{K}}_i+{\mathbf{u}}\) subject to a reflection of negative values into the positive domain. The MH acceptance probability for the new vector \({\mathbf{K}}_i^{\star }\) is then \(A({\mathbf{K}}_i,{\mathbf{K}}_i^{\star })=\min \left\{ 1,R({\mathbf{K}}_i,{\mathbf{K}}_i^{\star })\right\} \), with
where the Hastings-Ratio (HR) is equal to one, and the prior probability ratio (PR) depends on the model variant.Footnote 6 For the original CheMA model (Oates et al. 2014) we obtain from Eq. (11):
For the proposed new variant (iCheMA) we get from Eq. (12):
If the move is accepted, we replace \({\mathbf{K}}_i\) by \({\mathbf{K}}_i^{\star }\), or otherwise we leave \({\mathbf{K }}_i\) unchanged. Pseudo code of the MCMC sampling scheme for the new model variant (iCheMA) is provided in Table 1. Table 2 shows that the replacement of Eq. (23) by Eq. (24) yields a substantial reduction of the computational costs of the MCMC inference.
3.2 Model selection
The ultimate objective of inference is model selection, i.e. to infer the n regulator sets \(\pi _i\) (\(i=1,\ldots ,n\)) of the interaction processes described by Eq. (6). We compare and critically assess five alternative paradigms: the divergence information criterion (DIC), proposed by Spiegelhalter et al. (2002), the ’widely applicable information criterion’ (WAIC), proposed by Watanabe (2010), the ’widely applicable Bayesian information criterion’ (WBIC), proposed by Watanabe (2013), the cross-validation information criterion (CVIC) , proposed by Gelfand et al. (1992), and the marginal likelihood (also called ’model evidence’). For the ’marginal likelihood’ paradigm, we compare two numerical methods: Chib’s method (Chib), proposed by Chib and Jeliazkov (2001), and thermodynamic integration (TI), proposed by Friel and Pettitt (2008). For the latter, we have further assessed the numerical stabilization of the numerical integration proposed by Friel et al. (2013) in a simulation study, which can be found in Appendix 3.
3.3 Posterior probabilities of interactions
For the improved variant of the CheMA model (iCheMA) we follow Oates et al. (2014) and perform ’model averaging’ to compute the marginal posterior probabilities of all regulator-regulatee interactions (i.e. the ’edges’ in the interaction graph). The marginal posterior probability for species u being a regulator of i is given by:
where \(\varPi \) is the set of all possible regulator sets \(\pi _i\) of species i, and \(\varPi ^{(u\rightarrow i)}\) is the set of all regulator sets \(\pi _i\) of i that contain the regulator u. For simplicity, we chose a uniform prior for \(\pi _i\) subject to a maximum cardinality of 3 for the set of regulators (’parents’) of a node.
3.4 Network inference scoring scheme
For the CheMA model (Oates et al. (2014)) and the novel model variant (iCheMA) the marginal interaction posterior probabilities in Eq. (25) can be used to rank the network interactions in descending order. If the true regulatory network is known, this ranking defines the receiver operating characteristic (ROC) curve (Hanley and McNeil 1982), where for all possible threshold values, the sensitivity (or recall) is plotted against the complementary specificity. By numerical integration we then obtain the area under the curve (AUROC) as a global measure of network reconstruction accuracy, where larger values indicate a better performance, starting from AUROC = 0.5 to indicate random expectation, to AUROC = 1 for perfect network reconstruction. A second well established measure that is closely related to the AUROC score is the area under the precision recall curve (AUPREC), which is the area enclosed by the curve defined by the precision plotted against the recall (Davis and Goadrich 2006). AUPREC scores have the advantage over AUROC scores that the influence of large quantities in false positives can be identified better through the precision. These two scores (AUROC and AUPREC) are widely applied in the systems biology community to score the global network reconstructions accuracy (Marbach et al. 2012).
3.5 Causal sufficiency
In Appendix 1 we discuss causal sufficiency and how its violation affects the inference of regulatory network structures.
4 Evaluation
4.1 ANOVA
Like other network reconstruction models, CheMA and the novel iCheMA model yield a ranking of the regulatory interactions. Hence, if the true interaction network is known, AUROC and AUPREC scores can be computed, as explained in Sects. 3.3 and 3.4. For our performance evaluation on realistic network data, described in Sect. 5.2, we were running simulations for different settings, e.g. related to different regulatory network structures, shown in Fig. 4, and different inference methods. In order to distinguish the relevant effects from the confounding factors, we adopted the DELVE evaluation procedure for comparative assessment of classification and regression methods in Machine Learning (Rasmussen 1996; Rasmussen et al. 1996). That is, we set up a multi-way analysis of variance (ANOVA) scheme to disentangle the factors of interest. For instance, if there are three main effects (A, B, and C), and \(y_{ijkl}\) is the l-th AUROC (or AUPREC) score obtained for the constellation \(A=i\), \(B=j\) and \(C=k\), then we set up a 3-way ANOVA scheme:
where \(\varepsilon _{ijkl}{\sim }N(0,\sigma ^2)\) is zero-mean white additive Gaussian noise. We then computed ANOVA confidence intervals, e.g. for the effects of \(A=1\), \(A=2,\ldots \) on the AUROC scores (i.e. confidence intervals for the parameters \(A_1, A_2,\ldots \)); see, e.g., Brandt (1999) for details.
Thereby the main effects of the ANOVA models varied in dependence on the addressed research question. Throughout Sect. 6 we use the following five main effect symbols:
-
\({\mathbf{N_n}}\) (\(n=1,\ldots ,6\)) is the effect of the Network structure, see Fig. 4 for the six structures,
-
\({\mathbf{K_k}}\) (\(k=1,\ldots 4\)) is the effect of the GP Kernel, used for the computation of the analytical gradient (’RBF’, ’PER’, ’MAT32’, or ’MAT52’),
-
\({\mathbf{G_g}}\) (\(g=1,2\)) is the effect of the Gradient type, i.e. numerical (’difference quotient’) versus analytical (’GP interpolation’),
-
\({\mathbf{M_m}}\) (\(m=1,\ldots ,12\)) is the effect of the inference Method, i.e. the iCheMA model and eleven competing methods, listed in Table 5,
-
and \({\mathbf{P_p}}\) (\(p=1,2\)) is the effect of the Prior on \({\mathbf{V}}_i\) (i.e. the g-prior from Eq. (11) versus the ridge regression prior from Eq. (12).
In Appendix 2, we test the statistical assumptions of the ANOVA scheme in Eq. (26), and we include additional results related to the improvement of GP regression over numerical differentiation, and the influence of the network topology.
4.2 Simulation details
In our study we have included the four IC: DIC, WAIC, CVIC, and WBIC, and we have employed four different numerical methods to approximate the marginal likelihood: Chib’s original method (Chib and Jeliazkov 2001) (Chib naive), a stabilized version of Chib’s method (Chib), proposed here, thermodynamic integration with the trapezoid rule (TI), see Eq. (38), and thermodynamic integration with the numerical correction (TI-STAB), see Eq. (41).
The computation of the model selection scores (DIC, WAIC, CVIC, WBIC and the MLL with both Chib’s method and TI) requires MCMC simulations; pseudo code can be obtained from Table 1. We monitored the convergence of the MCMC chains with standard convergence diagnostics based on potential scale reduction factors (Gelman and Rubin 1992). The application of Chib’s method is based on the selection of a particular ’pivot’ parameter vector \(\tilde{\varvec{\theta }}\), as described under Eq. (34). Initially, we chose \(\tilde{\varvec{\theta }}\) to be the MAP (maximum a posteriori) estimator from the entire MCMC simulation. This was found to lead to numerical instabilities, though, as seen from the right panel of Fig. 8. We found a way to numerically stabilize Chib’s method, which we discuss in Appendix 1. We also found that the transition from TI to TI-STAB proposed by Friel et al. (2013) can be counter-productive, as seen from Table 4. We have investigated this unexpected effect in more detail in a simulation study in Appendix 3. We also found that the improved variant of CheMA (iCheMA) substantially reduces the computational costs of the MCMC-based inference, as shown in Table 2 and discussed in more detail in Appendix 1.
5 Data
5.1 Synthetic data
We generate \(T=240\) data points \(x_s(t_1),\ldots ,x_s(t_T)\) for \(n=4\) species (\(s=1,\ldots ,4\)) from iid standard Gaussian distributions. Subsequently, to obtain non-negative concentrations, the observations of each individual species are shifted such that the lowest value is equal to 0, before we follow Oates et al. (2014) and re-scale the observations of each species to mean 1. With \(x_1\) taking the role of the degradation process and \(x_2\) being an activating regulator (\(I_{2,y}=1\)) of a target species, whose gradient y(.) we here assume to be directly observable, we generate target observations \(y(t_j)\) with Eq. (6). We then have for \(j=1,\ldots ,T\):
where \({\mathbf{V}}_y=(v_{0,y},v_{2,y})^\mathsf{T}\) is the vector of maximum reaction rate parameters, \({\mathbf{K}}_y=(k_{2,y})\) contains the Michaelis–Menten parameter(s), and \(\epsilon _{t_j} \sim \mathscr {N}(0,\sigma ^2)\) is additive iid Gaussian noise (\(j=1,\ldots ,T\)). We keep the Michaelis–Menten parameter fixed at \(k_{2,y}=1\), while we vary the rates \(v_{0,y}\) and \(v_{2,y}\), as indicated in Table 3. Our goal is to infer the set of regulators \(\varvec{\pi }\!_{y}\) of y out of all subsets of \(\left\{ x_2,x_3,x_4 \right\} \), where \(\varvec{\pi }\!_{y}=\left\{ x_2 \right\} \) is the true regulator set. The effect of the degradation, taken into account by \(x_1\), is included in all 8 models.
5.2 Realistic data
For an objective model evaluation, we use the benchmark data from Aderhold et al. (2014), which contain simulated gene expression and protein concentration time series for ten genes in the circadian clock of A. thaliana. The time series correspond to measurements in 2-h intervals over 24 h, and are repeated 11 times, corresponding to different experimental conditions. We use time series generated from six variants of the circadian gene regulatory network in A. thaliana, shown in Fig. 4; these variants correspond to different protein, i.e. transcription factor, knock-downs. The molecular interactions in these graphs were modelled as individual discrete events with a Markov jump process, using the mathematical formulation from Guerriero et al. (2012) and practically simulated with Biopepa (Ciocchetta and Hillston 2009), based on the Gillespie algorithm (Gillespie 1977). For large volumes of cells, the concentration time series converge to the solutions of ODEs of the form in Eq. (5). However, for smaller volumes, time series simulated with Markov jump processes contain stochastic fluctuations that mimic the mismatch between the ODE model and genuine molecular processes, and the volume size was chosen as described in Guerriero et al. (2012) so as to match the fluctuations observed in real quantitative reverse transcription polymerase chain reaction (qRT-PCR) profiles. For the network reconstruction task, we only kept the gene expression time series and discarded the protein concentrations; this emulates the common problem of systematically missing values for certain types of molecular species (in our case: protein concentrations).
5.3 Real data
We have applied the iCheMa model to gene expression time series obtained with real-time polymerase chain reaction experiments to predict the circadian regulatory network in Arabidopsis thaliana. All details and the results can be found in Appendix 4.
6 Results
This section discusses the effect of gradient approximation (Sect. 6.1), the influence of the prior (Sect. 6.2), the accuracy of model selection (Sect. 6.3), the relative performance compared to the current state of the art (Sect. 6.4), and the problem of model mismatch (Sect. 6.5).
6.1 Evaluating the effect of the gradient computation
To illustrate the difference in the accuracy of network inference between a numerically calculated gradient using the difference quotient defined in Eq. (3) and an analytical gradient using a GP, we applied both gradient types to the realistic data of Sect. 5.2 and evaluated the performance of the methods listed in Table 5 together with iCheMA. The difference quotient was calculated with a time difference of \(\delta _t = 2\) hFootnote 7, and the analytical gradient was calculated with a GP using a RBF kernel as described in Sect. 2.2. The results of a preliminary study, in which we investigated the effect of the GP kernel on the network reconstruction accuracy, can be found in Appendix 2. We have recorded the AUROC and AUPREC scores for all the different conditions mentioned in Sect. 5.2, and summarize the outcome with an ANOVA analysis that treats the different conditions and methods as distinct effects. Fig. 5a shows that the results for the data with the analytical rate estimation significantly improves the performance over the numerical difference quotient while taking all methods of Table 5 and iCheMA into account. The same trend can be observed in Fig. 5b, which only considers the results for iCheMA. Hence, we conclude that network reconstruction accuracy significantly improves when an analytically derived gradient using a GP is used instead of the numerical difference quotient used in Oates et al. (2014). This can be observed for both a broad variety of different network reconstruction methods as well as specifically for the iCheMA method.
6.2 Evaluating the influence of the parameter prior
To evaluate the influence of the parameter prior on model selection, we computed the MLL for the g-prior in Eq. (11) and the ridge regression prior in Eq. (12) for the synthetic data from Sect. 5.1. For each of the 9 parameter settings, shown in Table 3, and 4 different noise variances \(\sigma ^2\) we generated 10 independent data instantiations from Eq. (27). We then applied the stabilized Chib method (Chib) for approximating the MLLs for all possible regulator–regulatee configurations, and computed the logarithmic Bayes factors (i.e. the differences of the MLLs) between the true and each wrong model. Fig. 6 gives the average logarithmic Bayes factors (averaged across the \(9\times 4=36\) parameter settings) for 10 independent replications. In Fig. 6 the distributions of the MLL score differences are represented with boxplots, where positive values indicate that the true structure is correctly selected, whereas negative values indicate that the wrong alternative structure is erroneously selected. It can be seen that regulator sets that do not contain the true regulator (corresponding to the three rightmost box plots) are clearly rejected. However, for the over-complex alternative models (three leftmost box plots), which contain spurious regulators, the MLL score difference obtained with the g-prior fails to consistently favour the true model. Two out of three differences are negative in Fig. 6a, indicating that an over-complex model is preferred to the true one. For the proposed ridge regression prior, on the other hand, the MLL score difference does succeed in consistently favouring the true structure, as displayed in Fig. 6b. We had a closer look at the results for the individual parameter sets, shown in Fig. 7. This figure reveals that the g-prior performed well for some parameter settings, but not for others. In particular, we found that the g-prior systematically fails when ’\(v_{0,y}<1\le v_{2,y}\)’ (see Table 3). In Appendix 1 we provide a theoretical explanation for this trend.
6.3 Model selection
We used the synthetic data from Sect. 5.1 to cross-compare the performance of the model selection schemes; for an overview see Appendix 1. The MLL based selection procedures score the individual models with respect to the differences in the MLL (or log Bayes factors). We compare these results with the score differences of the various IC. For inference we used the novel iCheMA model, and we varied the hyperparameters of the prior distributions so as to obtain increasingly diffuse prior distributions. In a first scenario we set the hyperparameter \(\delta ^2\) in Eq. (12) to a value sf, which we refer to as ’spread-factor’, and we kept \(\nu =0.5\) in Eq. (10) fixed. In the second scenario we set both hyperparameters \(\delta ^2\) and \(\nu \) to the spread factor sf. Figure 8 shows boxplots of the log score differences between the true model and an over-complex alternative model (with one redundant regulator-variable) for both scenarios and increasing spread factors (sf). Again positive differences indicate that the true structure is correctly selected, whereas negative differences indicate that the alternative over-complex structure is erroneously selected. The overall trend revealed in Fig. 8 is that the log score difference decreases with the spread-factor (i.e. with the diffuseness of the prior) for most of the model selection criteria.
The parameter priors in Eqs. (10) and (12) are Gaussians centred on \(\mu =1\), with different variances. For low spread-factors sf (i.e. for low prior variances), both groups of criteria (IC and MLL) clearly favour the true model, since the prior ’pulls’ the spurious interaction parameter from its true value of zero towards a wrong value of \(\mu =1\). As the prior becomes more diffuse, the score differences become less pronounced, but still select the true model up to spread factors of about \(sf\approx 100\). As the prior becomes more diffuse, with the spread factor exceeding \(sf> 100\), the IC occasionally fail to select the correct model. A more detailed representation focusing on the IC and larger spread factors is given in Fig. 9. It is seen that among the IC it is mainly DIC that repeatedly fails to select the true model (the central inter-quartile range of the score difference distribution, between the first and third quartile, includes negative values), whereas for the other information criteria the selection of the wrong model is relatively unlikely (the central inter-quartile range does not include negative values). Two of the four MLL methods, namely TI-8 and Chib, start to increasingly favour the true model as the spread factor further increases beyond \(sf> 1000\). This is a consequence of Lindley’s paradox, whereby MLL increasingly penalizes the over-complex model for increasingly vague priors. TI-4, in principle, shows a very similar trend but the score difference is lower than for TI-8, indicating that the choice of the discretization points (i.e. the applied temperature ladder) implied by the power \(m\in \{4,8\}\) in Eq. (42) can critically affect the result.
Among the MLL methods, the naive application of Chib’s method, Chib naive, as proposed in Chib and Jeliazkov (2001), shows a completely different pattern and systematically fails to select the correct model for large spread factors. A theoretical explanation for this instability is provided in Appendix 1. We achieve a stabilization of Chib’s method, referred to as Chib, by selecting the pivot parameter set \(\tilde{\varvec{\theta }}\) with the highest posterior probability within the set of actually sampled parameters (excluding the parameter states from the burn-in phase). We refer to Appendix 1 for details.
The left panel of Fig. 8 shows that the different ways of computing the MLL give very similar results up to a prior spread factor of about 1e+08. For spread factors exceeding this value, the results differ. The MLL computed with Chib’s method (Chib) monotonically increases, as expected from Lindley’s paradox. The MLL computed with TI is obtained without numerical stabilization and reaches a plateau, with different values obtained for different trapezium sum discretization schemes (determined by m in Eq. (42)). This is a numerical discretization error that results from the form of the integrand in Eq. (37), which has most of its area concentrated on values near \(\tau =0\). We tried to stabilize TI with the corrected trapezium rule, replacing Eq. (38) by Eq. (41). Interestingly, this transition from TI to TI-STAB turned out to be occasionally counter-productive, as shown in Table 4. We have investigated the effect of the numerical “stabilization” more thoroughly in a simulation study based on a Bayesian linear regression model. We found that TI-STAB can fail for small numbers of discretization points and diffuse priors. The study and its results can be found in Appendix 3. In our subsequent simulations we use the numerically stabilized variant of Chib’s method (Chib), which has a 10-fold lower numerical complexity compared to TI (because we used 10 different temperatures \(\tau \) for TI).
6.4 Comparison with state-of-the-art network reconstruction methods
We have compared the prediction accuracy of the proposed new novel variant (iCheMA) of the semi-mechanistic CheMA model of Oates et al. (2014) with the original CheMA model and 11 state-of-the-art machine learning methods, assessed in Aderhold et al. (2014). These methods are listed in Table 5 and were applied as described in Aderhold et al. (2014). The network reconstruction accuracy performance was tested on the realistic gene expression profiles from Sect. 5.2, based on the six network structures from Fig. 4. The results in terms of AUROC and AUPREC scores are shown in Fig. 11 and demonstrate that the iCheMA model outperforms all alternative methods. The CheMA model of Oates et al. is not included in this comparison. Due to the substantially higher computational costs, the simulations for all six networks in Fig. 4 would require several weeks of computing time on a medium-size cluster, as indicated by Table 2. In order to keep the computational complexity manageable, we compared CheMA and iCheMA in a separate study, using only data from the wildtype network, proposed by Pokhilko et al. (2010) and shown in the top left panel of Fig. 4. Note that for all methods included in Fig.11, the gradient for the response \(\frac{d x_i(t)}{dt}\) of Eq. (6) is derived from an analytical solution of the derivative using a GP with RBF kernel, as described in Sect. 2.2. The original definition of CheMA, on the other hand, uses a numerical gradient estimation. To separate out the effects of gradient estimation and the other differences between the methods, we applied both CheMA and iCheMA with both gradients: the numerical and the analytical gradient. The results are shown in Fig. 10. The two findings of this study are: iCheMA consistently outperforms CheMA, and the analytical gradient leads to a significant improvement in the prediction accuracy over the numerical gradient.
6.5 Model selection for network identification
As a final test, we evaluated the accuracy of model selection for network identification, using the data from Sect. 5.2. These data contain gene expression time series from the six gene regulatory networks of Fig. 4, which contain one wildtype and five mutant networks from protein knock-down experiments. We computed the MLL (with Chib’s method) for each of the candidate networks, for each data set in turn. The evaluation was repeated over five independent data instantiations. For each data set, all models were ranked based on the MLL, and the ranks were averaged over all data sets. This leads to the six-by-six confusion matrix of Fig. 12a, where the rows represent the networks used for data generation, and the columns represent the candidate networks ranked with the MLL. Note that we have emulated the mismatch between data generation and inference characteristic for real applications. For data generation, molecular interactions corresponding to the edges in the network were modelled with complex Markov jump processes, as described in Sect. 5.2, with the intention to mimic real biological processes. For inference and model selection, interactions were modelled with Michaelis–Menten kinetics, corresponding to Eq. (6), and the interactions between different regulators were modelled additively, by adding the Michaelis–Menten terms—this reduction in complexity is required for general computational tractability and scalability. The results are shown in Fig. 12a. The diagonal elements of the matrix show the average ranks for the correct network. Two of the six network structures are consistently correctly identified (average rank 1), but for the other four structures, the average ranks vary between 1.6 and 3. This failure to consistently identify the true network tallies with the fact that the AUROC scores in Fig. 11 are significantly below 1.0. The explanation is that the iCheMA model is currently restricted to additive interactions, as shown in Eq. (6). The data generation process, summarized in Sect. 5.2 and available in complete mathematical description from Guerriero et al. (2012), contains molecular processes related to complex formation (e.g. protein heterodimerization). Complex formation involving two transcription factors a and b acting on target gene i is mathematically described by a product of Michaelis–Menten terms of the form
where the symbols have the same meaning as in Eq. (6); see Pokhilko et al. (2010) for explicit mathematical expressions.
We included prior knowledge about molecular complex formation and expanded the iCheMA model accordingly to include the corresponding product of Michaelis–Menten terms in Eq. (6). We then computed the MLL as before and repeated the analysis. The results are shown in Fig. 12b and demonstrate that, by making the model more faithful to the data-generating process, model selection has substantially improved: the average ranks of the true network (shown in the diagonal elements of the matrix) are never worse than a value of 2 (out of 6), and reach the optimal value of 1 in 50 % of the cases. The corresponding results for the various IC are displayed in Table 6. When restricting the model to additive terms (greater mismatch between data and model), MLL outperforms the IC, as presumably expected. Interestingly, when reducing the mismatch between data and model by including the product terms, two IC, WAIC and CVIC, are competitive with MLL and perform even slightly better. DIC is substantially outperformed by MLL and the competitive information criteria WAIC and CVIC; these findings are consistent with the earlier results from Sect. 6.3. The performance of WBIC lies between WAIC /CVIC and DIC.
7 Discussion
Automatic inference of regulatory structures in our study is based on a bi-partition of the variables into putative regulators (transcription factor proteins) and regulatees (mRNAs) and a physical model of the regulation processes based on Michaelis–Menten kinetics. This effectively conditions the inference on assumed prior knowledge and is, as such, contingent on the accuracy of these assumptions. In our study we have allowed for a mismatch between the assumed prior knowledge and the ground truth. First, the assumed model is deterministic, defined in terms of ordinary differential equations, while the data-generation mechanism is stochastic (simulated with a Markov jump process). Second, the interaction model is additive, while the data-generation mechanism includes multiplicative terms. Third, we have allowed for the possibility of missing data (missing protein concentrations). Our results show that due to this mismatch, the true causal system cannot be learned (see e.g. Fig. 11, which shows AUROC and AUPREC scores clearly below 1). However, our work suggests that causal inference based on a simplified physical model achieves significantly better results than inference based on an empirical model. (See Fig. 11. Note that only iCheMA is based on a physical model; all the other methods use machine learning methods based on empirical modelling). Our study also quantifies how the performance improves as the physical model is made more realistic; see Fig. 12.
Semi-mechanistic modelling is a topical research area, as evidenced by the recent publication by Babtie et al. (2014). Our article complements this work by addressing a different research question. The objective of Babtie et al. is to investigate how uncertainty about the model structure (i.e. the interaction network defined by the ODEs) impacts on parameter uncertainty, and how parameter confidence or credible intervals are systematically underestimated when not allowing for model uncertainty. Our article addresses questions that have not been investigated by Babtie et al. : how accurate is the network reconstruction or ODE model selection, which factors determine it, and to what extent? Our work has been motivated by Oates et al. (2014), and we have shown that the authors’ seminal work, which won the best paper award at ECCB 2014, can be further improved with two methodological modifications: a different gradient computation, based on GP regression, and a different parameter prior, replacing the g-prior used by Oates et al. by the ridge regression prior more commonly used in machine learning. These two priors have e.g. been discussed in Chapter 3 of Marin and Robert (2007), but without any conclusions about their relative merits. Our study provides empirical evidence for the superiority of the ridge regression prior (Fig. 6) in the context of semi-mechanistic models, and a theoretical explanation for the reason behind it (Sect. 6.2). Table 2 shows that the new iCheMA variant reduces the computational costs drastically; a theoretical explanation for the reduction is provided in Appendix 1.
Our work has led to deeper insight into the strengths and shortcomings of different scoring schemes and numerical procedures. We have investigated the effectiveness of DIC as a method of semi-mechanistic model ranking. DIC is routinely used for model selection in Winbugs (Lunn et al. 2012), and the paper in which it was introduced (Spiegelhalter et al. 2002) has got over 5000 citations at the time of the submission of the present article. However, our findings that in the context of network learning DIC often prefers a model with additional spurious complexity over the true model (Fig. 9) questions its viability as a selection tool for semi-mechanistic models.
We have further compared different methods for computing the MLL. We have shown that Chib’s method (Chib and Jeliazkov 2001) can lead to numerical instabilities. These instabilities have also been reported by Lunn et al. (2012) and are presumably the reason why Chib’s method is not available in Winbugs. We have identified the cause of the numerical instability (see Appendix 1), and propose a modified implementation that substantially improves the robustness and practical viability of the method. This modification appears to be even preferable to thermodynamic integration, which at higher computational complexity shows noticeable variation with the discretization of the integral in Eq. (37) and the number of ‘temperatures’. It has been suggested (Friel et al. 2013) that the accuracy of thermodynamic integration can be improved by including second-order terms in the trapezium sum—see Eq. (41)—but the findings of our study are that this correction is no panacea for a general improvement in numerical accuracy, and that there are scenarios where the second-order correction can be counter-productive. It has come to our attention that a more recent method for improving thermodynamic integration has been proposed by Oates et al. (2016). Including this method in our benchmark study would be an interesting project for future research.
Due to the high computational complexity and potential instability of the MLL computation, several articles in the recent computational statistics literature have investigated faster approximate but numerically more stable alternatives. In our work, we have included WAIC, CVIC and WBIC as alternatives to MLL and evaluated their potential for model selection in two benchmark studies. It turns out that these more recent IC significantly outperform DIC (Figs. 8 and 9), and that WAIC and CVIC are compatible in performance with model selection based on the MLL (Table 6). It is advisable that several independent studies for different systems be carried out by independent researchers in the near future, but our study points to the possibility that statistical model selection in complex systems may be feasible at a comparable degree of accuracy but with substantially lower computational costs than with MLL.
Notes
Within this paper we refer to the resulting gradients as the numerical (Oates et al. 2014) and the analytical gradient, proposed here.
Using the whole distribution, as on page 58 of Holsclaw et al. (2013), would give us additional indication of uncertainty akin to a distribution of measurement errors. Due to the increased computational costs (additional matrix operations) this has not been attempted, though.
In the original CheMA model, a different decay term is used (for protein dephosphorylation), and no inhibitory interactions are included for numerical reasons. The linear decay term in Eq. (6) is more appropriate for transcriptional regulation, and the inclusion of inhibitory interactions achieves better results (as shown in Appendix 2).
This issue applies to the CheMA model, as proposed by Oates et al. and to the new improved variant (iCheMA), proposed here. For both models we implement the exact uncollapsed rather than the approximate collapsed Gibbs sampling step for \(\sigma _i^2\).
The truncation of \({\mathbf{V}}_i\) is then automatically properly taken into account, as it will only be conditioned on \({\mathbf{V}}_i\) that fulfil the required constraint.
The HR is equal to 1, as the proposal moves are symmetric. New candidates \({\mathbf{K}}_i^{\star }\) with negative elements are never accepted, as they have the prior probability zero, \(P({\mathbf{K}}_i^{\star })=0\).
In the realistic data study, we assume gene measurements to take place in intervals of \(\delta _t = 2\) h. This mimics typical rates for the qRT-PCR sampling experiments in Flis et al. (2015).
\(\int _a^b f(x) dx = {(b-a)} \frac{f(b)+f(a)}{2} - \frac{(b-a)^3}{12} f``(c)\) for some \(c\in [a,b]\).
Loosely speaking, under ’ideal circumstances’ the MCMC sample should contain parameters with high posterior probabilities, while parameters deviating from the sampled ones, such as \(\tilde{\varvec{\theta }}\), should be assumed to be very unlikely.
Note that \(\mu =1\) is the prior expectation of \(v_{u,i}\) and \(v_{0,i}\), see Eq. (11).
In real-world applications the maximal number of regulators for each species can be restricted to a maximal ’fan-in’ (or ’in-degree’) value, but this fan-in is rarely set to to a value lower than 3. Hence, taking the degradation process (i.e. the first column of the design matrix \({\mathbf{D}}_i\)) into account, the MCMC inference requires at least 4-dimensional multivariate Gaussian integrals to be computed. For completeness, we note that substantially more effective algorithms for computing multivariate Gaussian CDFs are only available for the bivariate and trivariate case, see, e.g., the algorithms in Drezner and Weslowsky (1989) and Genz (2004).
The score in Eq. (60) is the deviation between the true and an approximated log Bayes factor.
A full list of growing conditions and strains is available from the MocklerLab (Mockler et al. 2007) at ftp://www.mocklerlab.org/diurnal/expression_data/Arabidopsis_thaliana_conditions.xlsx
The equation for protein translation and degradation of P only depends on the status of light and darkness. Thus, protein P is produced throughout darkness with a sharp decline at dawn. By multiplication with a continuous (or binary) light indicator in the range [0,1] we obtain a sharp peak at dawn for Light*P. This peak might act as an initial ‘start of day’ impulse for some of the clock genes.
References
Aderhold, A., Husmeier, D., Grzegorczyk, M.: Statistical inference of regulatory networks for circadian regulation. Stat. Appl. Genet. Mol. Biol. 13(3), 227–273 (2014)
Babtie, A.C., Kirk, P., Stumpf, M.P.H.: Topological sensitivity analysis for systems biology. PNAS 111(51), 18,507–18,512 (2014)
Barenco, M., Tomescu, D., Brewer, D., Callard, R., Stark, J., Hubank, M.: Ranked prediction of p53 targets using hidden variable dynamic modeling. Genome Biol. 7(3), r25 (2006)
Brandt, S.: Data Analysis: Statistical and Computational Methods for Scientists and Engineers. Springer, New York (1999)
Chatfield, C.: The Analysis of Time Series. Chapman & Hall, Boca Raton (1989). iSBN 0-412-31820-2
Chib, S., Jeliazkov, I.: Marginal likelihood from the Metropolis-Hastings output. J. Am. Stat. Assoc. 96(453), 270–281 (2001)
Ciocchetta, F., Hillston, J.: Bio-PEPA: a framework for the modelling and analysis of biological systems. Theor. Comput. Sci. 410(33), 3065–3084 (2009)
Davis, J., Goadrich, M.: The relationship between precision-recall and ROC curves. In: Proceedings of the 23rd International Conference on Machine Learning (ICML), ACM, pp. 233–240 (2006)
De Jong, H.: Modeling and simulation of genetic regulatory systems: a literature review. J. Comput. Biol. 9(1), 67–103 (2002)
Drezner, Z., Weslowsky, G.: On the computation of the bivariate normal integral. J. Stat. Comput. Simul. 35, 101–107 (1989)
Flis, A., Fernández, A.P., Zielinski, T., Mengin, V., Sulpice, R., Stratford, K., Hume, A., Pokhilko, A., Southern, M.M., Seaton, D.D., et al.: Defining the robust behaviour of the plant clock gene circuit with absolute rna timeseries and open infrastructure. Open Biol. 5(10), 150,042 (2015)
Fogelmark, K., Troein, C.: Rethinking transcriptional activation in the arabidopsis circadian clock. PLoS Comput. Biol. 10(7), e1003,705 (2014)
Friel, N., Pettitt, A.: Marginal likelihood estimation via power posteriors. J. R. Stat. Soc.: Ser. B (Stat. Methodol.) 70, 589–607 (2008)
Friel, N., Hurn, M., Wyse, J.: Improving power posterior estimation of statistical evidence. Stati. Comput. 24(5), 709–723 (2013)
Gelfand, A.E., Dey, D.K., Chang, H.: Model determination using predictive distributions with implementation via sampling-based methods. Bayesian Stat. 4, 147–167 (1992)
Gelman, A., Rubin, D.B.: Inference from iterative simulation using multiple sequences. Stat. Sci. 1, 457–472 (1992)
Gelman, A., Carling, J.B., Stern, H., Dunson, D.B., Vehtari, A., Rubin, D.: Bayesian Data Analysis, 3rd edn. Chapman & Hall, Boca Raton (2014a)
Gelman, A., Hwang, J., Vehtari, A.: Understanding predictive information criteria for Bayesian models. Stat. Comput. 24(6), 997–1016 (2014b)
Genz, A.: Numerical computation of rectangular bivariate and trivariate normal and t probabilities. Stat. Comput. 14(3), 251–260 (2004)
Genz, A., Bretz, F.: Numerical computation of multivariate t probabilities with application to power calculation of multiple contrasts. J. Stat. Comput. Simul. 63, 361–378 (1999)
Genz, A., Bretz, F.: Comparison of methods for the computation of multivariate t probabilities. J. Comput. Graph. Stat. 11(4), 950–971 (2002)
Gillespie, D.T.: Exact stochastic simulation of coupled chemical reactions. J. Phys. Chem. 81(25), 2340–2361 (1977)
Grzegorczyk, M., Aderhold, A., Husmeier, D.: Inferring bi-directional interactions between circadian clock genes and metabolism with model ensembles. Stat. Appl. Genet. Mol. Biol. 14(2), 143–167 (2015)
Guerriero, M.L., Pokhilko, A., Fernández, A.P., Halliday, K.J., Millar, A.J., Hillston, J.: Stochastic properties of the plant circadian clock. J. R. Soc. Interface 9(69), 744–756 (2012)
Hanley, J.A., McNeil, B.J.: The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology 143(1), 29–36 (1982)
Holsclaw, T., Sansó, B., Lee, H.K., Heitmann, K., Habib, S., Higdon, D., Alam, U.: Gaussian process modeling of derivative curves. Technometrics 55(1), 57–67 (2013)
Lawrence, N.D., Girolami, M., Rattray, M., Sanguinetti, G.: Learning and Inference in Computational Systems Biology. Computational Molecular Biology. MIT Press, Cambridge (2010)
Lindley, D.V.: A statistical paradox. Biometrika 1, 187–192 (1957)
Lunn, D., Jackson, C., Best, N., Thomas, A., Spiegelhalter, D.: The BUGS Book: A Practical Introduction to Bayesian Analysis. Chapman & Hall, Boca Raton (2012)
Marbach, D., Costello, J.C., Küffner, R., Vega, N.M., Prill, R.J., Camacho, D.M., Allison, K.R., Kellis, M., Collins, J.J., Stolovitzky, G., et al.: Wisdom of crowds for robust gene network inference. Nat. Methods 9(8), 796–804 (2012)
Marin, J.M., Robert, C.P.: Bayesian Core: A Practical Approach to Computational Bayesian Statistics. Springer, New York (2007)
Mockler, T., Michael, T., Priest, H., Shen, R., Sullivan, C., Givan, S., McEntee, C., Kay, S., Chory, J.: The DIURNAL project: DIURNAL and circadian expression profiling, model-based pattern matching, and promoter analysis. In: Cold Spring Harbor Symposia on Quantitative Biology, Cold Spring Harbor Laboratory Press, vol. 72, pp. 353–363 (2007)
Murphy, K.: Machine Learning: A Probabilistic Perspective. MIT Press, Cambridge (2012)
Oates, C.J., Dondelinger, F., Bayani, N., Korkola, J., Gray, J.W., Mukherjee, S.: Causal network inference using biochemical kinetics. Bioinformatics 30(17), i468–i474 (2014)
Oates, C.J., Papamarkou, T., Girolami, M.: The controlled thermodynamic integral for Bayesian model evidence evaluation. J. Am. Stat. Assoc. (2016). doi:10.1080/01621459.2015.1021006
Ota, K., Yamada, T., Yamanishi, Y., Goto, S., Kanehisa, M.: Comprehensive analysis of delay in transcriptional regulation using expression profiles. Genome Inform. 14, 302–303 (2003)
Pokhilko, A., Hodge, S., Stratford, K., Knox, K., Edwards, K., Thomson, A., Mizuno, T., Millar, A.: Data assimilation constrains new connections and components in a complex, eukaryotic circadian clock model. Mol. Syst. Biol. 6(1), 416 (2010)
Pokhilko, A., Fernández, A., Edwards, K., Southern, M., Halliday, K., Millar, A.: The clock gene circuit in Arabidopsis includes a repressilator with additional feedback loops. Mol. Syst. Biol. 8, 574 (2012)
Pokhilko, A., Mas, P., Millar, A.: Modelling the widespread effects of TOC1 signalling on the plant circadian clock and its outputs. BMC Syst. Biol. 7(1), 1–12 (2013)
Ramsay, J., Hooker, G., Campbell, D., Cao, J.: Parameter estimation for differential equations: a generalized smoothing approach. J. R. Stat. Soc.: Ser. B 69(5), 741–796 (2007)
Rasmussen, C.E.: Evaluation of Gaussian processes and other methods for non-linear regression. PhD thesis, University of Toronto, Dept. of Computer Science (1996)
Rasmussen, C.E., Williams, C.K.I.: Gaussian Processes for Machine Learning. MIT Press, Cambridge (2006)
Rasmussen CE, Neal RM, Hinton GE, van Camp D, Revow M, Ghahramani Z, Kustra R, Tibshirani R (1996) The DELVE manual. http://www.cs.toronto.edu/~delve
Smolen, P., Baxter, D.A., Byrne, J.H.: Modeling transcriptional control in gene networks—methods, recent results, and future directions. Bull. Math. Biol. 62(2), 247–292 (2000)
Solak E, Murray-Smith R, Leithead WE, Leith DJ, Rasmussen CE (2002) Derivative observations in Gaussian process models of dynamic systems. In: Proceedings of Advances in Neural Information Processing Systems (NIPS), MIT Press, Cambridge (2002)
Spiegelhalter, D.J., Best, N.G., Carlin, B.P., Van Der Linde, A.: Bayesian measures of model complexity and fit. J. R. Stat. Soc.: Ser. B (Stat. Methodol.) 64(4), 583–639 (2002)
Toni, T., Welch, D., Strelkowa, N., Ipsen, A., Stumpf, M.P.: Approximate Bayesian computation scheme for parameter inference and model selection in dynamical systems. J. R. Soc. Interface 6(31), 187–202 (2009)
Vanhatalo, J., Riihimäki, J., Hartikainen, J., Jylänki, P., Tolvanen, V., Vehtari, A.: GPstuff: Bayesian modeling with Gaussian processes. J. Mach. Learn. Res. 14(1), 1175–1179 (2013)
Vyshemirsky, V., Girolami, M.A.: Bayesian ranking of biochemical system models. Bioinformatics 24(6), 833–839 (2008)
Watanabe, S.: Asymptotic equivalence of Bayes cross validation and widely applicable information criterion in singular learning theory. J. Mach. Learn. Res. 11, 3571–3594 (2010)
Watanabe, S.: A widely applicable Bayesian information criterion. J. Mach. Learn. Res. 14, 867–897 (2013)
Yang, Z.: Maximum likelihood phylogenetic estimation from DNA sequences with variable rates over sites: approximate methods. J. Mol. Evol. 39, 306–314 (1994)
Acknowledgments
This project was partially supported by a grant from the Engineering and Physical Sciences Research Council (EPSRC) of the UK, grant reference number EP/L020319/1, and a grant from the European Commission FP7, “Timet”, grant agreement 245143. We would like to thank Andrew Millar for helpful discussions.
Author information
Authors and Affiliations
Corresponding author
Appendices
Appendices
1.1 Appendix 1: Methodology
1.1.1 Model selection
The ultimate objective of inference is model selection, i.e. to infer the n regulator sets \(\pi _i\) (\(i=1,\ldots ,n\)) of the interaction processes described by Eq. (6). Thereby each regulator set \(\pi _i\) also includes a binary variable that indicates, for each regulator, the type of regulation: activation versus inhibition. For notational clarity, let \(\varvec{\theta }_i=({\mathbf{V }}_i,{\mathbf{K}}_i,\sigma _{i}^{2},\delta _{i}^{2})\) denote the model-specific parameters, and recall that \({\mathbf{y}}_i=(y_i(t_1),\ldots ,y_i(t_T))^{\top }\) denotes the observed (or approximated) gradients for species \(x_i\).
DIC is an improvement on the classical information criteria AIC and BIC and is defined as
where \( \overline{\varvec{\theta }_i} = \int \varvec{\theta }_i p(\varvec{\theta }_i|\pi _i,{\mathbf{y}}_i) d\varvec{\theta }_i\) is the posterior mean of the parameters. In practice, the integrals are approximated by sums over parameters approximately sampled from the posterior distribution \(p(\varvec{\theta }_i|\pi _i,{\mathbf{y}}_i)\) with MCMC. The marginal likelihood is defined as
The essential difference between Eqs. (28) and (29) is that DIC is defined as an expectation with respect to the posterior distribution, \(p(\varvec{\theta }_i|\pi _i,{\mathbf{y}}_i)\), whereas the marginal likelihood is defined as an expectation with respect to the prior distribution, \(p(\varvec{\theta }_i|\pi _i)\). This has two consequences. First, the marginal likelihood is affected by the choice of prior, and in particular is known to increasingly penalize a more complex model as the prior \(p(\varvec{\theta }_i|\pi _i)\) becomes more diffuse (Lindley’s paradox, see Lindley 1957). DIC, on the other hand, is unaffected by diffuse priors: the more diffuse a prior, the less of an influence it has on the posterior. The second consequence is that computing the log marginal likelihood (MLL) is more onerous, as we will discuss below.
DIC is routinely used for model selection in Winbugs (Lunn et al. 2012), and the paper in which it was introduced (Spiegelhalter et al. 2002) has got over 5000 citations at the time of the submission of the present article. However, a limitation of DIC is that it can only be applied to non-singular models. A statistical model is said to be non-singular if the map between parameters and probability distributions is one-to-one and if its Fisher information matrix is positive definite. This excludes several classes of important models, like mixture models. In singular statistical models, the maximum likelihood estimator does not satisfy asymptotic normality, and DIC loses its asymptotic justification. To intuitively see from where the difficulty arises, consider a likelihood with two equivalent modes, resulting e.g. from different label choices of a mixture model. The posterior expectation \(\overline{\varvec{\theta }_i}\) will then not be representative of the data, as it will lie in a region of low likelihood between the two modes.
A generalization of DIC that can be applied to singular models is the WAIC proposed in Watanabe (2010). Recall that the data vector \({\mathbf{y}}_i\) contains T individual observations \(\{y_i(t_1),\ldots ,y_i(t_T)\}\). Then WAIC is defined as
where \(\mathbb {E}(.)\) and \(\mathbb {V}(.)\) stand for the posterior mean and variance, i.e. the mean and variance with respect to the posterior distribution \(p(\varvec{\theta }_i|{\mathbf{y}}_i,\pi _i)\). The essential difference between DIC and WAIC is that for the latter, the posterior expectation and variance are defined in data space, \(y_i\), and not in parameter space, \(\varvec{\theta }_i\). This leads to greater numerical stability and is the reason why WAIC is not restricted to non-singular models.
Both DIC and WAIC are approximations to the Bayesian leave-one-out cross-validation (BLOOCV) estimate:
where \({\mathbf{y}}_i^{[t_j]} = \{y_i(t_1),\ldots ,y_i(t_{j-1}),y_i(t_{j+1}),\ldots ,y_i(t_T)\}\) is the leave-one-out data with sample \(y_i(t_j)\) removed. In fact, WAIC can be shown to be asymptotically equivalent to BLOOCV (Watanabe 2010), and the relation between DIC and BLOOCV is discussed in Gelman et al. (2014b). In practice, the integral on the right-hand side of Eq. (30) is replaced by a sum over a sample from an MCMC simulation:
where \(\{\varvec{\theta }_{i,j,k}\}_{k=1,\ldots ,K}\) is a sample from \(p(\varvec{\theta }_i|{\mathbf{y}}_i^{[t_j]})\). The practical difficulty is the high computational cost, as a consequence of having to rerun the MCMC simulations T times, on T separate data sets \({\mathbf{y}}_i^{[t_1]},\ldots ,{\mathbf{y}}_i^{[t_T]}\). An approximation to BLOOCV that only requires a single MCMC run (on \({\mathbf{y}}_i\)) was proposed in Gelfand et al. (1992)
which in practice is computed by
where \(\{\varvec{\theta }_{i,r}\}_{r=1,\ldots ,R}\) is an MCMC sample from \(p(\varvec{\theta }_i|{\mathbf{y}}_i)\). From these distributions in Eq. (33) one obtains the so-called CVIC as follows:
For computing the MLL, it is known that approximating the integral in Eq. (29) with Monte Carlo based on a direct sample from the prior \(p(\varvec{\theta }_i|\pi _i)\) shows, in general, extremely poor convergence. In the present article, we compare two improved and established methods: Chib’s method and thermodynamic integration. Chib’s method is based on
where the posterior near some selected ‘pivot’ parameters \(\tilde{\varvec{\theta }}_i\), \(p(\tilde{\varvec{\theta }}_i|\pi _i,{\mathbf{y}}_i)\), is approximated with MCMC; see Chib and Jeliazkov (2001) for details. TI is based on the power posteriors
with
from which the MLL is computed via
Here \(\mathbb {E}_{\varvec{\theta }_i,\tau }\) is an expectation with respect to the power posterior in Eq. (35). In practice, these expectations are computed for various ’temperatures’ \(\tau \) simultaneously with population MCMC, as described in Friel and Pettitt (2008). The integral in Eq. (37) is one-dimensional and can be numerically approximated e.g. with a trapezium sum:
A potentially numerically more stable alternative, which we refer to as TI-STAB, was proposed in Friel et al. (2013). Friel et al. show that:
where \(\mathbb {V}_{\varvec{\theta }_i,\tau _k}\) is the variance w.r.t. the power posterior in Eq. (35). The second derivative of \(\mathbb {E}_{\varvec{\theta }_i,t}[\log (p({\mathbf{y}}_i|\varvec{\theta }_i,\pi _i))]\) at a point \(\tau \in [\tau _{k-1},\tau _{k}]\) can then be approximated by the differential quotient of \(\mathbb {E}_{\varvec{\theta }_i,t}[\log (p({\mathbf{y}}_i|\varvec{\theta }_i,\pi _i))]\)’s first derivative from Eq. (39):
Friel et al. (2013) then employ the corrected trapezoidal ruleFootnote 8 to compute each sub-integral \(\int _{\tau _{k-1}}^{\tau _{k}}\mathbb {E}_{\varvec{\theta }_i,\tau }[\log (p({\mathbf{y}}_i|\varvec{\theta }_i,\pi _i))] d\tau \). This yields:
Since the contributions to the integral come essentially from the ’temperature’ range \(\tau \rightarrow 0\), the choice of discretization points is typically chosen of the form:
for \(\quad 0\le k \le K-1\), where K is the number of discretization points. For our study, we selected \(K=10\) and \(m \in \{4,8\}\).
Thermodynamic integration is more expensive than the other methods reviewed in this section, due to the need to run MCMC simulations for a range of smoothing levels \(\tau _k\). In principle a simplification could be achieved based on the mean value theorem of differential calculus
where \(\tau ^* \in [0,1]\), and \(Z({\mathbf{y}}_i,\pi _i,\tau )\) was defined in Eq. (36). The value \(\tau ^*\) is unknown. An asymptotically optimal approximation, derived in Watanabe (2013), is
The corresponding score is known as the WBIC:
In summary, we compare five model selection methods: DIC, WAIC, CVIC, WBIC, and the MLL. For MLL, we compare four alternative numerical methods: Chib’s method as proposed in Chib and Jeliazkov (2001) (Chib naive), a stabilized version of Chib’s method (Chib), proposed here, thermodynamic integration with the trapezoid rule (TI), see Eq. (38), and thermodynamic integration with a numerical correction (TI-STAB), see Eq. (41).
1.1.2 Causal sufficiency
Causal inference is a fundamental research topic, which is beyond the scope of the present study. What we pursue is model selection: given a set of candidate mathematical models, find the one that is most consistent with the data. This concerns the particular challenge of selecting explanatory variables within the proposed mathematical modelling context, which in terms of transcriptional regulation are the putative transcriptional regulators.
Ultimately, statistical inference faces two separate model selection problems. The first one is: given a mathematical description of the regulatory processes, find the network structure that is most consistent with the data. This problem is well-defined and the focus of our investigations. However, there is a second model selection problem underneath: what is the correct mathematical description of the regulatory processes? This problem is open-ended, in that there is an infinite number of regulatory processes that one could potentially consider. What we have focussed on is transcriptional regulation. It could be argued that this view is limited, and that additional processes related to micro-RNA regulation and unobserved species should ideally be taken into consideration. Besides the higher degree of computational complexity, beyond what is practically feasible with our computational resources, the set of interaction processes will potentially always have to be further extended, as our current knowledge of molecular biology is limited, and new scientific insights gained in the future, related to new molecular processes hitherto unknown, will have to be accommodated.
We have studied and quantified the consequences of model misspecification in two ways. First, we have compared our additive model of transcriptional regulation with models that are biologically less realistic. This is discussed in Sect. 6.4. Secondly, we have evaluated the effect of a misspecification related to ignoring protein interactions. This influence is quantified in Sect. 6.5. The consequences of ignoring processes related e.g. to micro-RNA regulation could in principle also be tested in a simulation study, but that is beyond the scope of the present article paper.
In more general terms, the question is how much we can trust the model selected from a model family that is incomplete and, most likely, oversimplified. But that is a fundamental problem affecting science in general. Our view of the world is never complete, and we need to base model selection on what is known or computationally feasible at the time the work is carried out. It is worth to recall George Box famous citation in this context, that all models are wrong, but some are useful. The considerable agreement of the network structures inferred from the real gene expression data with those independently published in the mathematical biology literature, as discussed in Appendix 4, confirms the usefulness of the presented inference framework despite its inevitable model simplifications.
1.1.3 The instability of Chib’s method
In Sect. 6.3 we have seen that Chib’s method, naively applied as proposed in Chib and Jeliazkov (2001), can yield unstable results under certain circumstances. In our study we considered diffuse prior distributions, which can yield diffuse posterior distributions with undesired attractor states in the configuration space of the parameters. To see that, assume that in an MCMC chain, the parameters of the CheMA (or iCheMA) model have reached, by chance, a very large Michaelis–Menten parameter \(k_{u,i}^{(r)}\) for a redundant species u, symbolically \(k_{u,i}^{(r)}\rightarrow \infty \). We then have:
for all time points \(t^{\star }\). That is, the design matrix column, corresponding to the redundant species u, becomes effectively a zero column having only a minor effect on the gradients of the response species i. The corresponding reaction rate parameter \(v_{u,i}^{(r)}\) can then also take very large values without any significant negative effect on the model likelihood in Eq. (29). This, in turn, can yield an undesired positive feedback mechanism between \(v_{u,i}^{(r)}\rightarrow \infty \) and \(k_{u,i}^{(r)}\rightarrow \infty \). Consequently, a naive choice of the pivot parameter vector \(\tilde{\varvec{\theta }}\) in Eq. (34) of Chib’s method, where the selected pivot parameters \(\tilde{v}_{u,i}\) and \(\tilde{k}_{u,i}\) for \(v_{u,i}\) and \(k_{u,i}\), respectively, substantially deviate from the MCMC sampled values with \(v_{u,i}^{(r)}\rightarrow \infty , k_{u,i}^{(r)}\rightarrow \infty \), yields a significant underestimation of the denominator in Eq. (34), and thus a significant overestimation of the MLL for models with redundant species. The reason for the underestimation of \(P(\tilde{\varvec{\theta }}|G,D)\) is that the selected pivot parameter vector \(\tilde{\varvec{\theta }}\) (\(\tilde{v}_{u,i}\in \tilde{\varvec{\theta }}\) and \(\tilde{k}_{u,i}\in \tilde{\varvec{\theta }}\)) is substantially different from the MCMC sampled parameters, where \(v_{u,i}^{(r)}\rightarrow \infty \), and \(k_{u,i}^{(r)}\rightarrow \infty \). Consequently, Chib’s method, as proposed in Chib and Jeliazkov (2001), significantly underestimates the posterior probability of the pivot parameters \(\tilde{\varvec{\theta }}\).Footnote 9 Attractor states indicative of diffuse posterior distributions cannot be avoided when diffuse prior distributions are used. However, a stabilization of Chib’s method can be achieved by selecting a pivot parameter vector \(\tilde{\varvec{\theta }}\) that is representative for the sampled parameter values. In our simulation studies we never observed unstable results when selecting the pivot parameters \(\tilde{\varvec{\theta }}\) with the highest posterior probability out of the set of actually sampled parameters (excluding the parameter states from the burn-in phase).
1.1.4 The inappropriateness of the g-prior
The inappropriateness of the g-prior, as demonstrated in Sect. 6.2, stems from the fact that the design matrices \({\mathbf{D}}_i\), defined in Sect. 2.3, have a very characteristic structure. The rows of the matrix \({\mathbf{D}}_i\) are given in Eq. (8) and it can be seen that the first column of \({\mathbf{D}}_i\), which represents the degradation, contains exclusively negative values, \((-x_i(t_1),\ldots ,-x_i(t_T))^{\top }\), while the subsequent columns contain only positive elements, \(\frac{I_{u,i} x_u(t^{\star })+(1-I_{u,i})k_{u,i}}{x_u(t^{\star })+k_{u,i}}\) (\(u\in \pi _i\) and \(t^{\star }\in \{t_1,\ldots ,t_T\}\)). Hence, the covariance structure \(({\mathbf{D}}_i^{\top }{\mathbf{D}}_i)^{-1}\) of the g-prior imposes strong positive correlations between the degradation parameter \(v_{0,i}\) and the other reaction rate parameters \(v_{u,i}\) (\(u\in \pi _i\)) in the original CheMA model. Parameter constellations, such as \(v_{0,i}<1<v_{u,i}\), where \(v_{u,i}\) and \(v_{0,i}\) are negatively correlated w.r.t. the prior in Eq. (11), are a priori penalized.Footnote 10 The strength of those correlations diminishes with the number of regulators in \(\pi _i\) (i.e. the number of covariates in \({\mathbf{D}}_i\)), which is the reason why the g-prior systematically favours over-fitting models. Replacing the truncated g-prior from Eq. (11) by the ridge regression prior in Eq. (12), as proposed here, avoids a priori imposed correlations among the parameters altogether. This was found to significantly improve the model selection performance, as demonstrated empirically in Fig. 6.
1.1.5 Computational costs of CheMA and iCheMA
In addition to the better network reconstruction accuracy, e.g. demonstrated in Figs. 10 and 11, the improved variant of CheMA (iCheMA) also substantially reduces the computational costs of the MCMC-based inference, as shown in Table 2. The computational bottleneck of the original CheMA model of Oates et al. (2014) stems from the particular form of the prior distribution on the unknown parameters. The (truncated) multivariate Gaussian g-prior on the maximum reaction rate parameters \(P_{\left\{ {\mathbf{V}}_i\ge 0\right\} }({\mathbf{V}}_i|\sigma _i^2,{\mathbf{D}}_i)\) in Eq. (11) depends on the values of the Michaelis–Menten parameters \({\mathbf{K}}_i\) via the design matrix \({\mathbf{D}}_i\)=\({\mathbf{D}}_i({\mathbf{K}}_i)\). As \({\mathbf{K}}_i\) can only be sampled with MH MCMC steps, see Eqs. (22, 23), each proposed move from \({\mathbf{K}}_{i}\) to \({\mathbf{K}}_{i}^{\star }\) also changes the prior distribution on \({\mathbf{V}}_i\). Hence, the MH acceptance probabilities in Eq. (23) includes the ratio:
where \(P(.|\sigma _i^2,{\mathbf{D}}_i({\mathbf{K}}_{i}))\) and \(F(.|\sigma _i^2,{\mathbf{D}}_i({\mathbf{K}}_{i}))\) are the probability density function and the cumulative distribution function (CDF) of the untruncated \(\mathscr {N}( {\mathbf{1}}, n \sigma _i^2 ({\mathbf{D}}_i' {\mathbf{D}}_i)^{-1} )\) Gaussian distribution, and \({\mathbf{D}}_i={\mathbf{D}}_i({\mathbf{K}}_i)\). Hence, each proposal move on the Michaelis–Menten parameters \({\mathbf{K}}_i\) requires two CDFs \(F({\mathbf{0}}|.)\) of multivariate Gaussian distributions to be computed. It is well-known that the computation of a Gaussian CDF with more than 3 dimensions (and an arbitrary covariance structure) is computationally challenging, and has for example to be done by randomized quasi Monte Carlo integration methods; see e.g. the algorithms proposed in Genz and Bretz (1999, 2002).Footnote 11 The computational bottleneck of the inference algorithm is caused by the fact that this Monte Carlo approximation is required in each step of the MCMC simulation.
In the new variant (iCheMA) the truncated g-prior is replaced by a truncated ridge regression prior, see Eq. (12), so that the prior distribution of \({\mathbf{V}}_i\) becomes independent of \({\mathbf{K}}_i\) (see the graphical models in Fig. 3). The prior PR in Eq. (23) then reduces to Eq. (24). Hence, PR does not depend on multivariate Gaussian integrals (CDFs), and the computational bottleneck is avoided. See Table 2 for a comparison of the average run-times. The replacement of the g-prior by the ridge regression prior also ensures that the prior probabilities of \({\mathbf{V}}_i\) can be computed by factorization. The mean vector, \(\varvec{\mu }={\mathbf{1}}\), and the covariance matrix, \(\varvec{\varSigma }=\delta ^2_i \sigma _i^2{\mathbf{I}}\), of the truncated Gaussian prior in Eq. (12) imply:
where \(|{\mathbf{V}}_i|\) is the dimension of \({\mathbf{V}}_i\), and \(F(.|\varvec{\mu },\varvec{\varSigma })\) denotes the CDF of a multivariate (or univariate) Gaussian distribution with mean \(\varvec{\mu }\) and variance \(\varvec{\varSigma }\).
1.2 Appendix 2: Extended results
1.2.1 Gradient calculation with the Gaussian process
The Matlab package gpstuff (Vanhatalo et al. 2013) provides a convenient implementation of GP regression including several kernel functions and a procedure to optimize the kernel parameters with scaled conjugate gradient optimization of the MLL (fminscg). Prior to applying the GP we partitioned the data of the full mRNA time series of each species into smaller fragments that correspond to individual experiments. This is necessary in order to avoid artefacts that can arise from transitions between unrelated experiments. The fragments contain 13 time points over 24 h for the realistic data set (Sect. 5.2) and 12 data points over 48 h for the real data set (Sect. 5.3). For all kernels, except the periodic kernel (PER), we set the initial length scale parameter to the time interval between two data points, which resulted in 4 h for the realistic data, and 8 h for the real data set. The initial periodic length for the PER kernel was consistently set to a 12 h period length for both data sets. The Matérn class kernel was used in two variants defined by setting the hyperparameter to \(\nu =3/2\) (MAT32), and \(\nu =5/2\) (MAT52). The initial signal variance \(\sigma ^2\) was set to 1 and the noise variance to \(\sigma ^2_n = 0.1\) for all kernels. Note that the above values serve as an input to the optimization function gp_optim, which uses the optimization procedure fminscg as the default. Finally, we took the first derivative of each of the implemented kernel functions, replaced the original covariance matrix with the one derived from the derivative functions, and then applied Eq. (1) in Holsclaw et al. (2013) to obtain the expectation of the derivative for each time point.
In a preliminary study, we investigated the effect of the GP kernel on the network reconstruction accuracy. We used the realistic data of Sect. 5.2, obtained from the six network structures shown in Fig. 4, and compared four widely applied standard kernels: the radial basis function (or RBF) kernel, the periodic kernel, and two Matérn class kernels with different hyperparameters: \(\nu =3/2\) (MAT32) and \(\nu =5/2\) (MAT52). For mathematical details of these kernels, see Chapter 4 in Rasmussen and Williams (2006). To keep the computational costs of this pre-study low, we used the Lasso method (see Table 5) for network reconstruction, as described in Aderhold et al. (2014). The results are shown in Fig. 13. The performance of the periodic kernel is clearly the worst. The differences in the results for the other kernels are not significant. However, the RBF kernel shows, overall, the best performance, and was therefore used in all studies.
1.2.2 Inhibitor kinetics and decay term in CheMA
Inhibitor kinetics. The inhibitor kinetics defined in Eq. (5) of Oates et al. (2014) are slightly different from Eq. (6) in our paper in that the inhibitory effects enter the Michaelis–Menten term via a sum in the denominator. However, this equation is only a theoretical construct that was not actually used in the authors’ work. In Sect. 3.1 of their paper the authors write that they found inference for inhibitor sets to be extremely challenging, and that they therefore did not include any inhibitory regulation in the reaction graph. This implies that the sum in the denominator of Eq. (5) of their paper is zero. Consequently, the model defined in Eq. (6) of our paper is a generalization of CheMA, which subsumes CheMA in the limiting case of no inhibition. We repeated the simulations reported in our paper with a fan-in restriction of zero imposed on the inhibitor set, i.e. with all inhibitors removed, like in CheMA. The results are shown in Fig. 14 and clearly demonstrate that this leads to a significant drop in performance.
Decay terms. The CheMA model includes a Michaelis–Menten decay term for dephosphorylation:
Dephosphorylation is an active form of regulation mediated by phosphatases, hence saturation processes have to be taken into consideration. The corresponding effect in transcriptional regulation is mRNA degradation. As opposed to dephosphorylation, this is a passive decay that is commonly modelled with a linear decay term (Barenco et al. (2006)). In order to adapt the original CheMA model to transcriptional regulation, we have therefore replaced the Michaelis–Menten decay term by a linear decay term:
see Eq. (5). This makes the model consistent with the biological literature; see e.g. Barenco et al. (2006) and Lawrence et al. (2010). To quantify the effect of this adaptation we have rerun the simulations reported in our manuscript with the linear decay term replaced by the Michaelis–Menten term used in CheMA. The results are shown in Fig. 15 and demonstrate that, in the context of transcriptional regulation, the linear term of Eq. (50) achieves overall better results than the Michaelis–Menten term of Eq. (49), with equal performance for four and a significant improvement for two of the six network structures included in the comparative analysis.
The upshot of these two studies is that if we apply CheMA exactly as reported in Oates et al. (2014), i.e. with Michaelis–Menten decay and no inhibition, the performance improvement achieved with the proposed new iCheMA model over CheMA would even be stronger than suggested by the results presented in our manuscript. However, to make the comparative analysis biologically realistic in the context of transcriptional regulation, it makes better sense to apply CheMA with the same modifications as for iCheMA, i.e. as reported in the manuscript.
1.2.3 Comparison with under-complex models
The results of a complementary study, where we compare the score differences (MLL versus IC) between the true and an under-complex model, are shown in Fig. 16. Unlike the previous study (Fig. 8), all methods succeed in identifying the true structure, indicating that the identification of an under-complex structure is substantially easier than the identification of an over-complex structure. While the IC score differences remain unaffected by the spread factor of the prior distribution (\(sf\)), the MLL score differences decrease with increasing \(sf\), as a consequence of Lindley’s paradox. However, for the two scenarios considered here, even for spread factors as large as \(sf=10^{20}\), MLL still allows the identification of the true structure, which suggests that the final outcome of model selection with MLL is not affected by Lindley’s paradox.
1.2.4 ANOVA
In Sects. 6.1, 6.2 and 6.5 of the main paper we evaluated the proposed improved CheMa model with respect to different effects, including the gradient response (Sect. 6.1), the choice of the prior distribution for the model parameters (Sect. 6.2), and the comparison of iCheMA to several state-of-the-art methods given realistic data sets (Sect. 6.5). These experiments involved a vast number of simulations and different setups. An example of one setup is shown in Fig. 17 for the method comparison of Sect. 6.5. This figure illustrates the complexity of the results, which makes it difficult to identify clear patterns and trends. In order to separate the effects of interest from the various confounding factors we adopted the DELVE evaluation scheme (Rasmussen (1996); Rasmussen et al. (1996)) and implemented a multi-way analysis (ANOVA, e.g. Brandt (1999)) as defined in Sect. 4.1.
To assess network reconstruction accuracy under different experimental conditions we first calculated AUROC and AUPREC scores that were used separately as response y in our ANOVA model:
The scores \(y_{gnml}\) were collected for all factors included in our simulation studies: For the method comparison study (Sect. 6.5) this involved the type of gradient calculation g, the network topology n, the network reconstruction method m, and the data instantiation l. The index range for these parameters are \(g \in \{0,1\}\), where \(g=0\) denotes a gradient obtained from the difference quotient, and \(g=1\) a gradient from a GP smoothing operation; \(m \in \{0,1,2,3,4,5\}\), where \(m=0\) represents the published network ‘wildtype’ and the five remaining indices represent modified topologies that simulate interventions of the wildtype network shown in Fig. 4; \(n \in \{0,1,2,\ldots ,11\}\), for the 11 network reconstruction methods shown in Table 5 and Fig. 20, and five different data instantiations \(l \in \{0,1,2,3,4\}\). The main effects were defined with discrete values denoted as \(G_g\) for the different gradients, \(N_n\) for the networks, and \(M_m\) for the various reconstruction methods. The additive error was defined as zero-mean Gaussian noise with \(\varepsilon _{ognmk} \sim N(0,\sigma ^2)\). Note that we modified Eq. (51) for the study in Sect. 6.2, where we considered the parameter prior (\(P_p\)) as main effect.
To confirm the validity of the ANOVA model, we tested the distribution of the residuals in terms of independence and identicality (i.i.d.). We carried out a standard residual analysis to assess whether the i.i.d. assumption was valid and the choice of this ANOVA model appropriate. A violation would indicate that the model in Eq. (51) was inadequate to fully capture the structure in the data, and that we would have to extend the model with, e.g. with higher-order terms.
To test the assumption of a Normal distribution, we created a QQ plot with the quantiles of the residuals of the AUROC and AUPREC scores plotted against the quantiles of a Normal distribution as shown in Fig. 18. Both plots show a good match along the dashed diagonal reference line with only minor deviations at the tails for the AUROC residuals (Fig. 18a). This suggests that the AUROC residuals have slightly heavier tails, although, an overall good agreement between the residual and Normal distributions can be observed.
The scatter plots in Fig. 19 show the residuals plotted against the fitted values from Eq. (51) for AUROC and AUPREC scores. The plots show overall even distributions that lack any discernible pattern or trend. This is consistent with the ANOVA model assumption that the distribution of the residuals does not depend on the main effects. Figs. 18 and 19 thus indicate that the ANOVA model of Eq. (51) is not in violation with the ANOVA model assumptions.
The plots in Fig. 20 and 21 show the results of the ANOVA approach for different effects in terms of the mean fitted AUROC and AUPREC values and corresponding confidence intervals. Fig. 20 displays the different methods \(M_m\) as the main effect of the model with iCheMA outperforming all other methods. A comparison of reconstruction accuracy with respect to the different network structures \(N_n\) in Fig. 21 shows that the accuracy increases with less complex interaction patterns. The effect of the gradient \(G_g\) is displayed in Fig. 5a. It can be seen that the GP gradient dominates over the difference quotient based gradient.
1.3 Appendix 3: Numerical stabilization
In Table 4 we have observed unexpected sharp increases of the MLL score difference obtained with the stabilized TI (TI-STAB) approach from Eq. (41). To explain the numerical instability of TI-STAB for large spread factors we consider a standard Bayesian linear regression model with a Gaussian distributed response variable y.
where \(\varvec{\pi }\) is a vector of \(m+1\) covariates (including a 1 for the intercept), \(\sigma ^2\) is the noise variance, and \(\varvec{\beta }=(\beta _0,\ldots ,\beta _m)^{\top }\) is the \((m+1)\)-dimensional vector of regression coefficients on which we impose a Gaussian prior:
where \(\delta ^2\) is a hyperparameter. The likelihood for T data points \((y_1,{\mathbf{x}}_1^{\top }),\ldots ,(y_T,{\mathbf{x}}_T^{\top })\) is:
where \({\mathbf{y}}=(y_1,\ldots ,y_T)^{\top }\) and \({\mathbf{X}}= ({\mathbf{x}}_1,\ldots ,{\mathbf{x}}_T)^{\top }\) is the design matrix. The marginal likelihood \(p({\mathbf{y}}|\delta ^2,M)\) and the power posteriors \(p(\varvec{\beta }|\delta ^2,{\mathbf{y}},\tau ,M)\) (\(\tau \in [0,1]\)) can then be computed in closed form:
where \(\mu _{\tau } = \frac{\tau }{\sigma ^2} \varvec{\varSigma }_{\tau } {\mathbf{X}}^{\top } {\mathbf{y}}\), and \(\varvec{\varSigma }_{\tau } = \sigma ^2 (\tau {\mathbf{X}}^{\top }{\mathbf{X}} + \frac{\sigma ^2}{\delta ^2} {\mathbf{I}})^{-1}\).
For each \(\tau \in \{\tau _1,\ldots ,\tau _K\}\) let \(\varvec{\beta }_{\tau }^{(1)},\ldots ,\varvec{\beta }_{\tau }^{(J)}\) be a sample from Eq. (55). The logarithm of the marginal likelihood in Eq. (54) can then be approximated by TI and TI-STAB via Eqs. (37–41), using:
For a comparison of the two thermodynamic integration approaches (TI and TI-STAB), see Eqs. (38–41), we compute Bayes factors \(\mathscr {B}\) for a (true) linear model M with one covariate (\(m=1\)) against an over-complex model \(\tilde{M}\) with two covariates (of which one is irrelevant). To this end, we we generate data from the true model. For \(i=1,\ldots ,T\):
where \(\epsilon _1,\ldots ,\epsilon _T\) are i.i.d. \(N(0,\sigma ^2\))-distributed. We set \(\beta _0=1\), \(\beta _1=-1\), \(\sigma ^2=1\), and \(T=100\). We sample the values \(x_{1,i},\ldots ,x_{1,T}\) of the covariate from standard Gaussian N(0, 1) distributions, and we build the true n-by-2 design matrix \({\mathbf{X}} = ({\mathbf{x}}_1,\ldots ,{\mathbf{x}}_T)^{\top }\), where \({\mathbf{x}}_i = (1,x_{1,i})^{\top }\). We compare the marginal likelihood \(p({\mathbf{y}}|\delta ^2,M)\) of this true model with the marginal likelihood \(p({\mathbf{y}}|\delta ^2,\tilde{M})\) of an over-complex model, for which we include a redundant covariable, having no effect on y. We sample its values \(x_{2,1},\ldots ,x_{2,T}\) from independent N(0, 1) distributions. Using Eq. (54) we compute the exact (logarithmic) Bayes factors:
for increasing hyperparameters \(\delta ^2=10^i\) \((i=-1,\ldots ,5\)), and we compare them with the approximations \(\mathscr {B}_{\delta ^2}(M,\tilde{M})\) obtained with TI and TI-STAB, see Eqs. (37–41), by plugging in the Monte Carlo estimates from Eqs. (56–57). As before we use the discrete temperatures \(\tau _k=\left( \frac{k}{K}\right) ^m\) \((0\le k \le K)\) with \(K=10\) and \(m=8\), and for each temperature \(\tau _k\) we take \(J=10000\) samples from the power posteriors in Eq. (55) to compute Eqs. (56–57). This yields two estimates \(\hat{\mathscr {B}_{\delta ^2}}(M,\tilde{M})_{TI}\) and \(\hat{\mathscr {B}_{\delta ^2}}(M,\tilde{M})_{TI-STAB}\) for each \(\delta ^2\).
In our study we compare the true log Bayes factors \(\mathscr {B}_{\delta ^2}(M,\tilde{M})\), computed with Eq. (59), with estimates \(\hat{\mathscr {B}_{\delta ^2}}(M,\tilde{M})\) approximated with TI and/or TI-STAB, respectively. Fig. 22a shows boxplots of the deviation scores:
for both approaches, TI and TI-STAB, applied to 10 independent data instantiations.Footnote 12 Fig. 23a shows the approximated Bayes factor \(\hat{\mathscr {B}_{\delta ^2}}(M,\tilde{M})\) for one individual (representative) data example. Both figures reveal that TI systematically under-estimates the log Bayes factor, while TI-STAB over-estimates it. Fig. 23a shows that for small values of \(\delta ^2\), TI-STAB gives indeed better predictions than its non-stabilized counterpart TI, as expected. However, for increasing values of \(\delta ^2\), this trend is reverted, and the performance of TI-STAB drastically deteriorates as \(log_{10}(\delta ^2)\) exceeds values of 3, rendering the numerical stabilization counter-productive. Since the systematic mismatches suggest that the approximation with the trapezium sum (without and with the correction term) with \(K=10\) might be too rough, we fix \(\delta ^2=10^5\) and increase the number of temperatures K. Fig. 22b shows boxplots of the \(\varDelta _{\mathscr {B}}(\delta ^2)\) from Eq. (60) with \(\delta ^2=10^5\) for the 10 data instantiations in dependence on K; again, the values \(\varDelta _{\mathscr {B}}(\delta ^2)\) for one representative data example are shown in Fig. 23b. As expected, for both approaches the estimation accuracy increases with K, and for large K \((K\ge 16)\) TI-STAB becomes more accurate than TI (even for this most diffuse prior with \(\delta ^2=10^5\)). However, the overall conclusions is that contrary to the findings in Friel et al. (2013), TI-STAB does not offer a consistent improvement on TI, and we have identified scenarios where the use of the ’stabilization method’ is counter-productive.
To understand the reason for the failure of the stabilisation method, note the conceptual difference between the computation of the first and second derivative. The estimation of the first derivatives can be regarded as accurate, as long as the sample size for the approximation of the variance in Eq. (39) is sufficiently large. However, the second derivatives are approximated by a difference quotient; see Eq. (40). To reduce this error, increasing the sample size drawn from the power posterior is not enough; the number of discretization points needs also to be increased. If both the sample size and the number of discretization points are increased together, then the quadratic correction will indeed lead to higher accuracy. However, in many applications, the number of discretization points is fixed, e.g. when running a population MCMC simulation with a fixed number of processors/temperatures. Our study has revealed that for a fixed number of 10 discretization points (which is what we could afford in terms of computational costs), the discretization error inherent in the computation of the second derivatives can turn out to be counterproductive if the prior is vague.
To explore in more detail why TI-STAB can substantially fail for diffuse priors, we fix \(K=10\) and \(\delta ^2=10^5\), and analyse the approximation of the MLLs in more detail. The temperatures \(\tau _0,\ldots ,\tau _{10}\) segment the interval [0, 1] into 10 disjunct sub-intervals, and we extract the interval-specific contributions to the MLL. Fig. 24 shows some diagnostics for the true model, applied to one typical data instantiation. The left panel compares the true interval-specific contributions and those approximated by TI. It can be seen that TI under-estimates the MLL by \(-11\), where especially the contribution of the third interval, \([\tau _2,\tau _3]\), is under-estimated. The centre left panel of Fig. 24 shows the interval-specific ratios of the right- and the left-hand side of Eq. (40). It can be seen that the approximation is critically inaccurate for the third interval. The centre-right panel of Fig. 24 compares the interval-specific TI-STAB corrections and the correction terms that are required for compensation of the discretization error with TI. For the third interval the inaccurate approximation via Eq. (40) yields a correction term that is too large, i.e. a ’hypercompensation’ of the mismatch obtained with TI (see Fig. 24, left panel). Finally, the right panel of Fig. 24 shows the resulting interval-specific contributions of TI-STAB, where the inaccurate approximation via Eq. (40) has led to a ’hypercompensation’ of the moderate mismatch of TI (Fig. 24, left panel), i.e. to a more drastic over-estimation.
1.4 Appendix 4: Real world application
1.4.1 Arabidopsis thaliana data
To predict a circadian regulatory network in Arabidopsis thaliana we use the DIURNAL data resource from the Mockler Lab (Mockler et al. 2007). The focus of the DIURNAL database are diurnal and circadian regulated genes that have been previously identified from microarray data using the Affymetrix ATH1 GeneChip platform (TAIR) in conjunction with the HAYSTACK tool (Mockler et al. 2007). Each selected gene in DIURNAL provides a compilation of up to 20 different time courses determined by different experimental conditions, which were produced by different laboratories. Each time series has gene expression samples from 12 consecutive measurements taken in 4 h intervals that cover 48 h. The experimental conditions include different settings such as 12 h light days, short days with 6 h of light, long days with 18 h of light, total darkness, constant light, different temperatures, and over-expressed genes. The background strain is predominantly Col-0Footnote 13. We extracted the gene profiles for the following A. thaliana genes: CCA1, LHY, PRR5, PRR7, PRR9, GI, TOC1, LUX, ELF3, ELF4, RVE8-1, and RVE8-2, where the last two correspond to two different Affymetrix probes of RVE8. Although the gene RVE8 is not an established component of the core clock, we are interested in the prediction of interactions with clock genes to compare our findings with a recent study by Fogelmark and Troein (2014). Since all 20 conditions were available for these genes, we were able to collect measurements for \(T=240\) timepoints for each gene. We normalized the data for each laboratory separately, since some of the time courses were found to be on a different scale. Furthermore we added an artificial light variable to mimic a light spike at the beginning of the day.
1.4.2 Modelling transcriptional delays
Transcriptional regulation can be subject to time delays. When a transcription factor binds to the promoter of a gene to initiate its transcription, the transcribed mRNA is not available immediately, but only after a certain transition time required to assemble and release it. Mathematically, this can be modelled by modifying Eq. (5) as follows:
where \(\varDelta \) is an additional parameter to allow for the fact that the effect of the transcriptional regulators on the regulatee is subject to the time delay \(\varDelta \). The ensuing mathematical modifications are straightforward; for instance, Eq. (6) becomes
and Eq. (8) generalizes to
In principle, we could now follow the same inference procedure as described before, with an augmented parameter vector in which \(\varDelta \) has been included: (\(\varvec{\theta }^\mathsf{T},\varDelta )^\mathsf{T}\). However, for the MLL we can use a ’trick’ from statistical phylogenetics to reduce the computational complexity. Denote by \(p(D|G,\varDelta )\) the marginal likelihood for a fixed time delay \(\varDelta \), obtained using the methods above with \(\varDelta \), e.g. in Eqs. (62, 63), fixed. We need to integrate out \(\varDelta \):
Here, \(p(\varDelta )\) is the prior distribution of the transcriptional delay \(\varDelta \), whose support is the domain of the nonnegative real numbers, and for which the gamma distribution is an appropriate choice (see Fig. 25):
The integral in Eq. (64) is analytically intractable. Interestingly, such integrals routinely occur in phylogenetics, where the standard procedure is to approximate them by a discrete gamma distribution, as first proposed by Yang (1994). The idea is to introduce \(k\) categories, each having approximately equal probability \(1/k\), and approximate Eq. (64) by
Yang (1994) suggested the following procedure for determining the \(\varDelta _i\)’s. Start with the percentage points corresponding to
and set
where \(G^{-1}(.|\alpha ,\beta )\) is the inverse of the cumulative distribution function of the gamma density \(g(.|\alpha ,\beta )\). Then, rescale all time delays, \(\varDelta _i = \lambda \tilde{\varDelta }_i\), \(\lambda > 0\), such that
An illustration is provided in Fig. 25. Yang (1994) showed that in the context of phylogenetic inference, the discrete gamma method gives reliable results for the number of discrete categories as low as \(k=3\), but the most common choice (which we use in our work) is \(k=4\).
For the inclusion of time delays via Eq. (64) we need their prior distribution. We obtained a comprehensive list of transcriptional time delays from a literature study (Ota et al. 2003), displayed the distribution as a histogram, shown in Fig. 25, and fitted a gamma density to it; see Eq. (65). We then approximated the integral in Eq. (64) with the discrete gamma distribution method of Yang (1994), as discussed above. We chose \(k=4\) discrete time delay categories, as recommended in Yang (1994); see Fig. 25.
1.4.3 Artificial light and data normalization
Light provides essential information that influences the regulatory system of the circadian clock. Proteins that are sensitive to light are known to convey this information to the clock. However, the real data described in Sect. 5.3 only provides “light versus darkness” information as binary descriptors and lacks quantities that more specifically reflect the action of light sensing components. The realistic Biopepa data simulations from Sect. 5.2 have shown that model prediction can be significantly improved by including an artificial light protein regulator called Light*P, as demonstrated in Fig. 26. Light*P resembles an artificial light spike in the beginning of the day and is composed of a binary light indicator multiplied by a hypothetical light sensing protein P; the latter was simulated with a Markov Jump Process following the procedure in Sect. 5.2. As an alternative to this approach we modelled the light influence as a change-point process that divides the data into a dark versus a light phase. Chib’s MLL is then calculated with iCheMA for each segment separately and summed up to the final MLL score. This model shows only minor and no significant improvement over the model that lacks any light information, as displayed in Fig. 26. Since the inclusion of Light*P has such a beneficial effect on model prediction, we decided to artificially generate this covariable with BiopepaFootnote 14 for the real data. The change of interaction probabilities, when including Light*P into the real data, is illustrated in the scatter matrix of Fig. 27.
Another important factor is data normalization, which is commonly carried out with a Z-score transformation of the whole time series of a gene. However, the DIURNAL MocklerLab data in Sect. 5.3 include time series that stem from different labs and exhibit measurements on different scales. To avoid distortions due to extreme values we decided to apply separate Z-score transformations to those sets of time series that originate from the same publisher or lab (“lab-wise” normalization). Note that a more reliable approach would be to scale the gene measurements based on the expression level of one of the house keeping genes before applying the Z-score transformation. However, we found that the different labs did not consistently use the same house keeping genes. The scatter matrix in Fig. 27 shows the difference of the marginal posterior interaction probabilities using “lab-wise” normalization compared to the standard approach that uses Z-scores over all time-series of a single gene (“gene-wise”). Although there are some changes in the medium probability range, there is little change at the very low and very high probability score spectrum.
1.4.4 Results
There is substantial interest in understanding the molecular mechanism of circadian regulation, i.e. an organism’s internal time keeping. For plants in particular circadian regulation is essential to align the plant’s metabolism to the diurnal rhythm of day and night. In our final study, we applied the improved CheMA method (iCheMA) to gene expression time series from the circadian clock genes in the model plant A. thaliana. Since this is a real data set, we allowed for transcriptional time delays, as discussed above. Based on the previous results, we selected the marginal likelihood as the most reliable model selection criterion overall, and we applied Chib’s method with our numerical stabilization.
Various hypotheses about the structure of the central circadian gene regulatory network in A. thaliana have been published in the biological literature. For the evaluation of our predictions, we used two recently published network structures, which we refer to as the P2013 (Pokhilko et al. 2013) and the F2014 network (Fogelmark and Troein 2014). Taking these structures as a gold standard, we can evaluate the network reconstruction accuracy with standard ROC curves. The results are shown in Fig. 28. The figure compares the performance of two different GP kernels: the RBF kernel and the periodic kernel (see Chapter 4 in Rasmussen and Williams (2006)). The figure also evaluates the effect of explicitly modelling time delays. A further subtlety is the fact that the F2014 network includes a gene that is not included in the P2013 network (RVE8). For that reason we have evaluated the reconstruction accuracy for the F2014 network twice: with RVE8 included and excluded. Fig. 28 also shows the AUROC, as well as the AUPREC as an alternative figure of merit.
With AUROC values ranging between 0.58 and 0.78, the agreement with the literature is significantly better than random expectation. The particular patterns found are as follows. The agreement with the F2014 network is better than with the P2013 network. The inclusion of time delays achieves a small improvement in terms of AUROC score (increase by 0.02), but not in terms of AUPREC scores. Unlike the results for the synthetic study, summarized in Fig. 13, the periodic kernel outperforms the RBF kernel.
To arrive at a particular network structure prediction, we took the marginal posterior probabilities obtained with the periodic kernel and the inclusion of time delays, corresponding to the top row in Fig. 28, and extracted all gene interactions above a selection threshold of 0.11; this value was found in Grzegorczyk et al. (2015) to lead to approximately the same number of gene interactions as in the P2013 network. The resulting network is shown in Fig. 29. To formally compare this prediction with the two networks from the literature, we created a Venn diagram, also shown in Fig. 29, which displays the number of interactions in the various intersections, e.g. the number of interactions predicted with iCheMA and found in P2013, but not in F2014 etc. Of the 15 gene interactions predicted with iCheMA in the way described above, shown in Fig. 29, 10 can be found in both the P2013 and the F2014 network, 2 can be found in the F2014 network, and only 3 do not agree with any of the published networks. This results are significantly better than random expectation and suggests that iCheMA provides a useful tool for hypothesis generation in molecular systems biology.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
About this article
Cite this article
Aderhold, A., Husmeier, D. & Grzegorczyk, M. Approximate Bayesian inference in semi-mechanistic models. Stat Comput 27, 1003–1040 (2017). https://doi.org/10.1007/s11222-016-9668-8
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11222-016-9668-8