
Research article

Open

 Published:
Identification of neutral biochemical network models from time series data
BMC Systems Biologyvolume 3, Article number: 47 (2009)
Abstract
Background
The major difficulty in modeling biological systems from multivariate time series is the identification of parameter sets that endow a model with dynamical behaviors sufficiently similar to the experimental data. Directly related to this parameter estimation issue is the task of identifying the structure and regulation of illcharacterized systems. Both tasks are simplified if the mathematical model is canonical, i.e., if it is constructed according to strict guidelines.
Results
In this report, we propose a method for the identification of admissible parameter sets of canonical Ssystems from biological time series. The method is based on a Monte Carlo process that is combined with an improved version of our previous parameter optimization algorithm. The method maps the parameter space into the network space, which characterizes the connectivity among components, by creating an ensemble of decoupled Ssystem models that imitate the dynamical behavior of the time series with sufficient accuracy. The concept of sloppiness is revisited in the context of these Ssystem models with an exploration not only of different parameter sets that produce similar dynamical behaviors but also different network topologies that yield dynamical similarity.
Conclusion
The proposed parameter estimation methodology was applied to actual time series data from the glycolytic pathway of the bacterium Lactococcus lactis and led to ensembles of models with different network topologies. In parallel, the parameter optimization algorithm was applied to the same dynamical data upon imposing a prespecified network topology derived from prior biological knowledge, and the results from both strategies were compared. The results suggest that the proposed method may serve as a powerful exploration tool for testing hypotheses and the design of new experiments.
Background
Mathematical models in modern molecular biology have become attractive as compactors of the massive amounts of multidimensional data produced by highthroughput techniques, thus following similar ideas that previously led from reductionism to quantitative inroads into physiology and ecology. In the smallerdimensional world described by the model structure and its parameters, new experiments are easier to conceive, hypotheses can be tested with greater clarity, and knowledge can be extended with inexpensive computational effort [1].
Generally, mathematical models are implemented with a set of parameters, which give them the flexibility of mapping a range of behaviors into a unifying mathematical framework. Except for singular cases where parameters are directly measured experimentally, parameter estimation from experimental data is an inevitable step in the process of constructing models [2]. A good, firsttier compromise between the need for a closedform, computable representation of the biological process and the risk of ignoring meaningful parameters of mechanistic, hypothesisdriven reductions can be found in the use of generic, "canonical" modeling frameworks. Specifically, within the framework of biochemical systems theory (BST) [3–6], Ssystem models (Equation 1) offer a particularly convenient solution because their parameters more or less directly describe the interactions between the components of the system of interest [6].
In the general Ssystem form (Equation 1), the time variation of the concentration or amount of each component X_{ i }of the system is given by the difference between production and degradation terms. The constant rates α_{ i }and β_{ i }represent the turnover rates of the production and degradation fluxes and the kinetic rates g_{ ij }and h_{ ij }quantitatively characterize the influence of the component X_{ j }on the production and degradation term of the system component X_{ i }, respectively [6, 7]. Thus, the network structure and the nature of the interactions driving the phenomenon under investigation are mapped essentially onetoone onto the parameter values of the model. This modeling framework has been successfully applied to many biochemical systems [6] and can generally be considered a good first default for representing complex biological systems, especially if the governing mechanisms are not well characterized [8]. Automation of the estimation of the highdimensional parameter set of an Ssystem model from multivariate time series data has therefore become a widely pursued computational challenge that has been addressed by a wide variety of optimization techniques: from relatively slow global heuristic optimization techniques like genetic algorithms and simulated annealing [9–12] to fast local optimization algorithms such as alternating regression and eigenvector optimization [13, 14] among others [15, 16]. Most of these optimization algorithms share the strategy of decoupling the differential equation system into a larger, nonlinear algebraic system [6, 11, 17, 18]. This strategy eliminates the need for numerical system integration at each step of the optimization process, which is expensive because Ssystems can be numerically stiff, just like most other nonlinear models.
Ironically, the difficulty of finding any numerically integrated Ssystem model that fits a given set of experimental time series data well is accompanied by the "opposite" problem: many recent publications have pointed out that multiparametric models tend to have the capacity of accommodating whole ranges of parameter values without much affecting the system dynamics [19–28]. Furthermore, it was found that there are typically welldefined directions in the parameter space to which the system dynamics is insensitive, a phenomenon that was termed "sloppiness" [22–25]. Since redundancy appears to be a widespread design feature of biological processes, exploring the admissible parameter space of a model and a dataset of interest has relevance that reaches well beyond typical sensitivity analyses of model parameters.
Although the concept of sloppiness has been discussed quite intensely, little attention has been paid to the question of whether or not sloppiness can be translated into the structure of the biological system itself. In other words, is the biological system in reality more or less uniquely parameterized or is there such significant interindividual variation that we could in principle find large "clouds" of parameter manifestations if we were able to determine the parameters in individual cells or organisms. The answer to this question is not without consequence, because it would affect the definition of what it means to have a good model fit to a given set of data.
In the case of Ssystems, the question has further implications. If an admissible parameter cloud, defined by a sufficiently accurate overall fit of some data, permits some kinetic order parameter to be positive, zero, or negative, the interpretation of the estimated model becomes distinctly different. In the first case, the effect is activating, in the second it is negligible, and in the third it is inhibiting. Is such a parameter cloud a computational artifact that would disappear if more data were available, or is it possible that a real biological system would actually allow such different effects of a variable on the system? To reformulate the question, do natural systems only allow for sloppiness in parameter values or also for sloppiness in structure? While the question itself is not new (e.g., [29]), it can presently not be answered in generality and with reliability.
In this report, we address the intrinsic redundancies in the interactions between biological system components by proposing an embedding method for Ssystem parameter space estimation based on a Monte Carlo process that is combined with an improved version of our previous parameter optimization algorithm. We apply this methodology to experimental time series data characterizing the glycolytic pathway in the bacterium Lactococcus lactis [30–32]. In the same context we explore the concept of sloppiness in Ssystems by studying the implications of admissible ensembles of models that dynamically represent the data well but lead to different interpretations.
Neutral solution analysis
This section describes the concepts of the proposed analysis of neutral solution spaces, i.e., of multiple model parameter sets with similar dynamics; all technical details are presented in the later Methods section. To characterize the neutral solution spaces, we propose a Monte Carlo (MC) random walk process [33], which is sped up by a nonlinear optimization algorithm that allows us to assess Ssystem parameter sets (forming the "neutral space") that give the system a similar dynamical behavior as it had been measured experimentally in the form of time series data. Differently from similar methodologies suggested in the literature [22, 23, 25, 26], the proposed approach is performed with the decoupled form of the system [34], which allows analysis of one system equation at a time. In this fashion, problems with numerical integration of differential equations, which is otherwise needed at each step of the MC process, are avoided. Thus, we suppose in the following that the time series of all components are available and have been smoothed, thus permitting the numerical estimation of their first derivatives at each point of the time series. For each system component, a series of steps is performed as follows.
First, using the smoothed time series of all components X_{ i }, as well as their numerically estimated derivatives, the system of differential equations is converted into a system of algebraic equations (Equation 2) [11, 18, 34]. Second, an optimization algorithm is applied to this algebraic system, leading to an optimal parameter set that matches the algebraic equations with the observed systems dynamics. This optimal parameter set is the starting point of the MC process. A cost function C is defined to quantify changes in behavior of the decoupled system resulting from perturbations in the parameter set (Equation 16). The Hessian matrix of C is calculated at the optimal parameter set and used to guide subsequent, artificial parameter perturbations, which collectively form the MC random walk [22, 25]. At each MC step, the parameter set is perturbed (using the eigenvectors of the Hessian matrix [35]) and then used as initial guess for the optimization algorithm, however, with an earlier stop criterion. This premature termination prevents the algorithm from converging to the same local optimal point and is accompanied by a (small) residual error. The cost function C is evaluated with the new local parameter set, and this parameter set is accepted for the next iteration with a certain probability (Equation 18). Any parameter sets satisfying the conditions of a predefined behavioral class (e.g., with a cost function value smaller than a threshold) are recorded. At the conclusion of the MC process, the recorded collection of parameter sets contains solutions of the decoupled system that adhere to the specified behavioral class.
After the MC process has been run for each system component, instances from the collections of parameter sets for all variables are randomly sampled to recouple the models by means of numerical integration. Although the decoupled, algebraic form offers the advantage of avoiding numerical integration problems, it is not guaranteed that the recoupled system will lead to an accurate, comprehensive solution. The Methods and Results sections provide detailed descriptions of the proposed techniques and outcomes.
Results
After preliminary, successful tests with simulated data (see supplementary material), we applied the proposed optimization algorithm to actual time series. These data consist of metabolite profiles from the glycolytic pathway of the bacterium Lactococcus lactis, which were obtained with in vivo NMR experiments [30, 31]. For modeling purposes, the concentrations of the metabolites were coded as follows: glucose – X_{1}; glucose 6phosphate (G6P) – X_{2}; fructose 1, 6bisphosphate (FBP) – X_{3}; phosphoenolpyruvate (PEP) – X_{4}; lactate – X_{5}; acetate – X_{6}. As a preprocessing step for the parameter optimization, the time series were smoothed/denoised and their slopes (numerical first derivatives) were estimated as shown in previous work on nonstationary noise filtering [36].
Initially, the proposed algorithm exclusively used the time series of all metabolites in the parameter optimization, representing the case where no knowledge about the network connectivity is at hand. The optimal parameter set of each metabolite was separately translated into the eigenspace of the solution and subsequently fed into the MC process. The parameters were then perturbed and an ensemble of models was created as described in the Methods section. The parameter sets were selected based on a behavioral class [37], which was defined by a residual less than 5 (see Equation 15). As an example, Figure 1 shows all parameter sets for the G6P production term that fall within the defined behavioral class. After ordering, these parameters have the same distribution pattern (Figure 1B). Similar results were found for all other metabolites.
As an alternative, we performed a MC random walk in the original parameter space, as opposed to the eigenspace. Interestingly, although not surprising in the end, the results did not present as wide a range as was found in the exploration of the parameter set in the eigenspace. The difference in outcomes is explained by the compensation of errors between production and degradation terms: perturbations in the eigenspace of the matrix W affect both production and degradation terms while perturbing only one parameter does not always maintain the balance between the two terms. In other words, given one of the Ssystem's terms (production or degradation) and the vector of slopes, the complementary term can be obtained by multiple linear regression, which has been shown not to be sloppy [24]).
As expected, eigenvalues of the Hessian matrix fall within a sparse range [22–25, 38], thereby elucidating the stiff and sloppy directions (see Additional file1). The region of the parameter space that produces similar dynamical model behaviors can be approximated as an ellipsoid whose main axes are given by the direction of the eigenvectors of the Hessian matrix (see Figure 2) and whose width is inversely proportional to the squared root of the corresponding eigenvalue [22–25, 38]. A projection of the ellipsoid into threedimensional space for the acetate production parameters g_{64}, g_{65} and g_{66} is shown in Figure 2. Revisiting reasons discussed for other modeling frameworks [22–25, 27], sloppiness can be explained in the proposed approach by the neutral space of solutions for Equation 6 and consequently Equation 8. Specifically, given the "right" linear combination of the eigenspace of the matrix W, any vector resulting from stretching or shrinking this combination is also a solution of Equation 6, however with different parameter values.
Ideally, all combinations of parameter sets found for the individual metabolites would generate recoupled models that fit the data upon numerical integration, within some error bound. Given the large number of parameters (at the order of 10^{14} combinations), a full exploration of this statement is not possible. In order to ameliorate the expensive combinatorial issue of assessing the trajectories of the recoupled systems, the parameter set for each metabolite generated from the MC process was randomly selected and the resulting system numerically integrated. This process was repeated 500 times and the results are presented in Figure 3.
As can be seen from Figure 3, the uncertainties and bounds in the prediction for most metabolites are actually close to the observed data. A notable exception is lactate, where the measured data contain little noise. Because PEP is the main precursor of lactate in the model [31] (pyruvate is not explicitly modeled), the class of predictions of the PEP dynamics was resampled for residuals less than 1. Furthermore, the newly sampled systems were integrated with glucose supplied in three different initial concentrations, namely 20, 30 and 40 mM. The results of these predictions are shown in Figure 4. For all three different scenarios the new ensemble model predictions provide accurate descriptions of the observed dynamics for the concentrations of lactate and the other metabolites in the model, except for the noticeable undershooting of PEP (see [30, 31] for information on the responses of the system to different initial glucose concentrations).
Although the models found by the proposed procedure accurately describe the dynamics observed in the experimental data, none of the parameter sets match well with the network topologies found in the literature, suggesting that distinctly different parameter combinations, and not just sloppy versions of some parameter set, area able to match these data. A variety of techniques can be applied to the ensemble in order to compare different models [37, 39], to cluster models by means of transformation groups, or even reduce them to a smaller subset [20]. One possible avenue for further analysis of this vast parameter space is by creating groups defined by different behavior classes based on biological and dynamical information [37, 39].
Regardless of specific followup analyses, one of the purposes of this work is to demonstrate that the proposed optimization algorithm can be effectively applied to the parameter identification of specific networks, for instance, by taking kinetic parameters of some component X_{ j }out of the optimization process of X_{ i }. Conversely, previous knowledge can be used to restrict the values of the g and h interaction parameters. For example, existing knowledge about Lactococcus lactis primary metabolism [30–32, 40] provides precise clues about which interactions are reasonable. To analyze the benefit of information outside the time series data, we applied the proposed optimization algorithm to the same Lactococcus lactis data, this time however constraining some parameter values with the preexisting information. The simulation results with the resulting system (Figure 5; see equations in Additional file1) not only describe the data but also agree with the double pulse of glucose described in [32].
Exactly as described before, the system resulting from this combined approach could be used as the starting point of a MC random walk process, and the same analysis of sloppiness and behavior classes could be performed. In order to assess the accuracy of the solutions for large systems, the optimization algorithm was also tested with a symbolic genetic network model consisting of 30 components [41] (see Additional file1). Because the algorithm performs the parameter optimization with the decoupled form of the system, its complexity is linear with the number of the system's components [14]. Thus, rather than system dimension, the real limitation of the optimization algorithm is the time series dynamics. Poor dynamical variability (components close to the steadystate) and collinear time series will result in a conditioning deficiency of the matrix L, causing numerical problems with the inverse operations and misleading the convergence. Despite the successful retrieval of the 30dimensional system dynamics, the algorithm's limitation becomes more evident with large systems (higher possibility of having time series portions with collinear components, resulting in illconditioned blocks in the matrix L [42]). This drawback is partially resolved with the regularization technique presented in the Methods section. Also, problems of this nature can be prevented by removing the collinear components from the matrix L or placing them as independent variables (variables that do not take place on the parameter optimization [6]). Of course, this issue is drastically diminished when a chosen network topology constraints the matrix L.
Discussion and Conclusion
Within the new field of systems biology, an extraordinary effort has been devoted to kinetic modeling with the aim of understanding biological processes better [1, 43, 44]. A very significant part of this effort has been directly related to the identification of model parameters [2, 45–47]. Until recently, the quality of the estimated parameter values was judged by the fit of experimental data. However, new analyses have pointed out the importance of other criteria, such as extrapolability and error compensation among terms [48], which is in some sense related to the sloppiness of admissible parameter sets and the fact that different parameter values can generate similar dynamical behaviors in nonlinear biological systems models [22, 23, 25–27]. These observations have a direct impact on the robustness of models, which may translate into robustness of the biological system itself, which may become apparent in tolerance to mutations, changes in gene expression, and insensitivity to modest changes in environmental conditions. This need for physiological robustness implies that biochemical networks should be able to preserve their dynamical properties within reasonable ranges of their kinetic parameters [49].
In this report, we present an extension of our previous optimization algorithm for Ssystem parameter identification from time series data [14]. The proposed method turned out to be faster and more accurate than its predecessor (see Additional file1) and was used here in combination with a Monte Carlo random walk technique to explore the space of admissible parameter sets of Ssystem models. This strategy allowed us to explore the concept of sloppy models. The results indicate that both, a fully integrated and a decoupled model, can be sloppy. We also reanalyzed time series of the concentrations of six metabolites within the glycolytic pathway of the bacterium Lactococcus lactis and demonstrated how a sloppiness analysis can elucidate the admissible parameter space and ultimately lead to more reliable estimates.
The central result reported here is that a diversity of parameter sets may produce quasiisomorphic dynamics for Ssystem models. Most of the parameter variations extended to both positive and negative parts of the parameter space (Figure 1B). This result is interesting, because it could be the consequence of two distinctly different scenarios. First, it could reflect redundancy or sloppiness caused by insufficient data. In other words, the data are not informative enough to distinguish between alternative models that fit equally well. In the past, such situations have often been "resolved" by setting as many parameters to zero as possible with the set of admissible solutions, borrowing arguments of parsimony or Ockham's razor. The important feature of this scenario is that further experimental data, maybe obtained under similar yet sufficiently different conditions, would declare one of the candidate models the (sole) winner. The second possible root cause of distinct, wellfitting parameter sets is the actual natural coexistence of different regulatory signals between components (inhibition versus activation) or different regulatory networks (some with and some without particular regulatory signals). Thus, otherwise similar cells or organisms would function under slightly different regulatory regimens. The difference between these scenarios is conceptually similar to the distinction between uncertainty and variability, which was discussed intensely in the 1980s and 1990s within the fields of risk assessment and exposure analysis [50]. The former case of uncertainty (due to insufficient data) is only valid if the model is, in principle, uniquely identifiable and structurally distinguishable [19, 51] from among all feasible Ssystems. This aspect raises the possibility that a sloppy model could be unidentifiable and that the sloppy directions, which are given by the possible parameter combinations [22] could be a measure of nonidentifiability [51]. For Ssystems this argumentation can be extended. Our results show that an ensemble of Ssystems could be interpreted as a collection of structurally indistinguishable and unidentifiable models [19]. This conjecture would explain the range of variability (negative, zero, positive) of the parameter sets that were found. Furthermore, the identifiability characteristic of the mathematical framework could be associated with the robustness of the biological system to environmental changes. Analogously, the structure distinguishability characteristic could be, in some sense, associated with the robustness of the biological system to mutations that change the network of interaction among components. A more rigorous study of Ssystem identifiability and distinguishability will be needed to reveal more concrete conclusions and mathematical implications [52, 53].
Whether the distinction between parameter sloppiness and structural distinguishability in biological models is an important issue will have to await further investigation. Nevertheless, it is clear that extrapolability in the generic network identification problem from time series data is a more complex task than the computational fitting of a model. One should expect that this observation is true for any model structure, but it was shown here that Ssystem models allow its exploration in a most translucent fashion.
The assessment of sloppiness provides valuable information that can be extracted from the prediction of the ensemble of models and through the investigation of behavioral classes that differ in their dynamical features (e.g. overshoots or response time). These classes may be further reduced by biological knowledge such as biochemical, physicochemical, or thermodynamic constraints [54, 55]. The proposed techniques could also serve as a powerful exploration tool for the testing of hypotheses and the design of new experiments. Moreover, the union of the proposed optimization algorithm with statistical methods could also result in a robust network inference implementation [56].
Maybe the most important value of this report is the clear definition of a framework to explore sloppiness, robustness and evolutionary innovation [38, 49, 57], where neutral parameter spaces (system with essentially the same dynamical behavior) are merged with neutral networks [49, 57] into one unique structure. This approach could reveal insights into how different metabolic networks could possibly have been changed during the evolutionary process.
Methods
The method described here is an extension of our previous work in Ssystems parameter estimation [14]. Exactly as its predecessor, the proposed algorithm exploits the advantages of the numerical decoupling of Ssystems, which allows that each component has its own optimization process [6, 11]. However, the optimization is now performed in a transformed equation of the decoupled form. This transformation enhances the ability of the algorithm to reach local optimal points. Given a time series collection X_{ m }(t_{ i }) of the concentrations of species or metabolites m = [1,..., M] in the time interval t_{ i }= [t_{1},..., t_{ N }] and their respective slope values S_{ m }(t_{ i }) estimated in the same time interval, one can numerically decouple the Ssystem equations and write a semilinearization (only the lefthand side is linearized) form for each species m as follows
where
and Vp_{ m }= [log α_{ m }g_{m 1}g_{m 2}⋯ g_{ mM }] is the parameter vector for the production term. In Equation (2), the time point index t_{ i }is suppressed in order to keep a simple notation. If a multiple regression step is applied to the system in Equation (2), the production parameter vector can be written as
The matrix L^{+} is the MoorePenrose inverse of the matrix L, given by L^{+} = (L^{T}L)^{1}L^{T}. Substituting Equation (4) into Equation (2), we obtain the following eigenvector problem [14, 58]:
In this equation, the matrix W = LL^{+} is idempotent with M+1 unit eigenvalues, while the remaining eigenvalues are zero. It is clear from Equation (5) that not only but also are eigenvectors of W corresponding to eigenvalue 1. If Equations (2), (4) and (5) are rewritten in order to isolate the degradation parameter vector instead of the production parameter vector, the vector also belongs to the eigenspace. Standard routines for eigenvalue analysis can easily calculate the eigenvectors of matrix W. However, this does not imply that any of these eigenvectors will map onto the "correct" parameter set. In fact, these vectors form the eigenspace of W corresponding to eigenvalue 1, leading to the conclusion that the true parameter set can be formulated as a linear combination of these eigenvectors. Thus, let EigS be a matrix where the columns are the eigenvectors with correspondent eigenvalues 1. Then the following vectors can be defined:
or
In Equations (6) and (7), ψ and δ are (M+ 1)dimensional vectors that represent an arbitrary linear combination of the eigenvectors of W. We can use Equation (7) to write the decoupled form of the Ssystem as
Equation (8), once written as a function of the production and degradation parameter vectors, can now be seen as a function of the vectors ψ and δ. Using the uniqueness of the mapping between the vectors ψ and δ,
we can rewrite Equaiton (8) as a function of only the vector δ. For the estimation of vector δ and consequently of vector ψ, we define a cost function as the sum of squared residuals between the two sides of Eq. (8).
where is the righthand side of the Equation (8). Thus, Equation (10) leads to the following gradient equations
or
where
The symbols and represent the Hadamard and Shanghai Jiao Tong (SJT) product respectively [42, 59], defined as
In Equations (11) and (12), υ is the argument of the logarithm in Equation (10). We use the LevenbergMarquardt routine [60] for the minimization of F with the following nonlinear constraints
In these inequality constraints (see also Equation (7)), the pairs lcbucb and lkbukb are the intervals (lower and upper bounds) for the rate constants and kinetic orders respectively. Preliminary tests with a simulated systems show the potential and accuracy of the algorithm (see Additional file1). Without change in the gradient equations, a Tikhonov regularization [61] was also implemented in order to overcome potential problems with the illconditioning of the matrix L, resulting in the following matrix:
where λ is the regularization parameter and I the identity matrix.
Sloppiness in Ssystems models
Sloppiness has been proposed as a nearly universal characteristic of parameter sensitivity from multiparametric nonlinear models (see definition of 'universal' in [22]). In a nutshell, a model is sloppy if it is markedly more sensitive to some parameter combinations than to others. This feature has been shown to be the rule rather than the exception among biological systems models [22]. One characteristic attribute of sloppiness is a diversity of parameter sets that produce similar dynamical behaviors. If a cost function is defined to quantify the variation in dynamical behavior one can visualize sloppiness in terms of the eigenvalues of the Hessian of this cost function. These eigenvalues are typically very sparse within a large range, suggesting that the model is more sensitivity to certain parameter combinations (eigenvectors with largest eigenvalues – stiff combinations) than to others (eigenvectors with smallest eigenvalues – sloppy combinations) [22–25].
For Ssystem models, the diversity in the parameter space is to be interpreted not only as a variation in the kinetic constants but also as redundancy in the topology of the biological network or as alternative topologies associated with the same phenotype. Either interpretation has no effect on the usefulness of the model from the predictability point of view as we discuss in the Results section. Indeed, it has been shown that sloppy models can lead to accurate predictions (e.g., [24, 38]).
Decoupled ensembles of Ssystem models
Recently, a Monte Carlo random walk method [33] in the parameter space of multiparametric system biology models was suggested in order to assess uncertainties in model parameters and model structure [22, 23, 25, 26, 38]. In this context, Ssystem models are particularly convenient because they map any network topology directly onto its parameter values (g and h) with no change in model structure [6]. Therefore, we propose a walk in the parameter space of the Ssystem and proffer that it corresponds to a walk in the 'topology space'. This idea is explored here by combining of the Monte Carlo approach with a nonlinear optimization algorithm. In contrast to relevant work in the literature, but in accordance with our optimization method, the random walk is performed separately with the decoupled Ssystem format for each system component. This strategy avoids difficulties encountered in the numerical integration of stiff equations and poses no significant loss of information, under the reasonable assumption that there are no bifurcation points within the considered parameter ranges. Thus, we define the following cost function, which quantifies variations in the first derivatives of a variable with respect to variations in the parameter set Δδ:
Here, δ* is the optimal parameter set found using the proposed optimization algorithm. Evaluation of the Hessian matrix,
allows us to analyze the sensitivity of the first derivative of each component of the system in relation to its parameters. This Hessian is also used to guide the Monte Carlo random walk, which selects the next perturbations within the parameter set. Thus, the Monte Carlo random walk starts with the optimal vector δ* as initial condition, randomly selects one component of this vector and perturbs it proportional to the Hessian eigenvector (see [25, 35]), and uses the new vector as initial condition for the proposed optimization algorithm, which now has an early stop criterion (number of iterations). The new optimal vector Δδ (i.e., the outcome of the optimization algorithm) is accepted as the next step of the process with probability
[26], where
is the probability distribution of the parameter set δ given the model for the slopes δ m(δ*). In Equation (19), k is a normalization factor (that vanishes in the probability of acceptance [26], Eq. (18)) and σ is the standard deviation of the slopes S_{ m }. If a new vector δ is accepted for the next step, the hessian matrix is recalculated and the process follows as described above. This strategy permits a more detailed exploration of the parameter space where the MC steps are taken in the sloppy directions. A small range for the number of iterations of the optimization algorithm (early stop criteria) was tested and no significant differences were observed. Relative large numbers were avoided in order to prevent the algorithm convergence to the same neighborhood. After the optimization process, the vectors δ are mapped onto the parameter degradation vector Vd_{ m }= [log β_{ m }h_{m 1}h_{m 2}⋯ h_{ mM }]^{T}and consequently onto the parameter production term Vp_{ m }= [log α_{ m }g_{m 1}g_{m 2}⋯ g_{ mM }]^{T}, restating the decoupled system in its original parameters. All the numerical integrations presented in this report were carried out by the MATLAB^{®} ode23s routine. The rate constants were optimized within the range [0.1, 300] while the kinetic order were optimized within the range [2, 2]. All the results shown in this report can be reproduced with the freely available MATLAB^{®} scripts (see Additional file2).
References
 1.
Westerhoff HV, Kolodkin A, Conradie R, Wilkinson SJ, Bruggeman FJ, Krab K, van Schuppen JH, Hardin H, Bakker BM, Mone MJ, et al.: Systems biology towards life in silico: mathematics of the control of living cells. J Math Biol. 2009, 58: 12. 10.1007/s0028500801608
 2.
Banga JR: Optimization in computational systems biology. BMC Syst Biol. 2008, 2: 47 10.1186/17520509247
 3.
Savageau MA: Biochemical systems analysis. I. Some mathematical properties of the rate law for the component enzymatic reactions. J Theor Biol. 1969, 25 (3): 365369. 10.1016/S00225193(69)800263
 4.
Savageau MA: Biochemical systems analysis. II. The steadystate solutions for an npool system using a powerlaw approximation. J Theor Biol. 1969, 25 (3): 370379. 10.1016/S00225193(69)800275
 5.
Savageau MA: Biochemical systems analysis. 3. Dynamic solutions using a powerlaw approximation. J Theor Biol. 1970, 26 (2): 215226. 10.1016/S00225193(70)800133
 6.
Voit EO: Computational analysis of biochemical systems: a practical guide for biochemists and molecular biologists. 2000, Cambridge; New York: Cambridge University Press
 7.
Savageau MA: Biochemical systems analysis: a study of function and design in molecular biology. 1976, Reading, Mass.: AddisonWesley Pub. Co., Advanced Book Program
 8.
Hecker M, Lambeck S, Toepfer S, van Someren E, Guthke R: Gene regulatory network inference: data integration in dynamic modelsa review. Biosystems. 2009, 96 (1): 86103. 10.1016/j.biosystems.2008.12.004
 9.
Kikuchi S, Tominaga D, Arita M, Takahashi K, Tomita M: Dynamic modeling of genetic networks using genetic algorithm and Ssystem. Bioinformatics. 2003, 19 (5): 643650. 10.1093/bioinformatics/btg027
 10.
Gonzalez OR, Kuper C, Jung K, Naval PC, Mendoza E: Parameter estimation using Simulated Annealing for Ssystem models of biochemical networks. Bioinformatics. 2007, 23 (4): 480486. 10.1093/bioinformatics/btl522
 11.
Voit EO, Almeida J: Decoupling dynamical systems for pathway identification from metabolic profiles. Bioinformatics. 2004, 20 (11): 16701681. 10.1093/bioinformatics/bth140
 12.
Liu PK, Wang FS: Inference of biochemical network models in Ssystem using multiobjective optimization approach. Bioinformatics. 2008, 24 (8): 10851092. 10.1093/bioinformatics/btn075
 13.
Chou IC, Martens H, Voit EO: Parameter estimation in biochemical systems models with alternating regression. Theor Biol Med Model. 2006, 3: 25 10.1186/17424682325
 14.
Vilela M, Chou IC, Vinga S, Vasconcelos AT, Voit EO, Almeida JS: Parameter optimization in Ssystem models. BMC Syst Biol. 2008, 2: 35 10.1186/17520509235
 15.
Sorribas A, Cascante M: Structure identifiability in metabolic pathways: parameter estimation in models based on the powerlaw formalism. Biochem J. 1994, 298 (Pt 2): 303311.
 16.
Sorribas A, Samitier S, Canela EI, Cascante M: Metabolic pathway characterization from transient response data obtained in situ: parameter estimation in Ssystem models. J Theor Biol. 1993, 162 (1): 81102. 10.1006/jtbi.1993.1078
 17.
Voit EO, Savageau MA: Powerlaw approach to modeling biological systems; III. Methods of analysis. J Ferment Technol. 1982, 60: 233241.
 18.
Almeida JS, Voit EO: Neuralnetworkbased parameter estimation in Ssystem models of biological networks. Genome Inform. 2003, 14: 114123.
 19.
Walter E, Pronzato L: On the identifiability and distinguishability of nonlinear parametric models. Mathematics and Computers in Simulation. 1996, 42: 125134. 10.1016/03784754(95)001239.
 20.
Voit EO: Symmetries of Ssystems. Math Biosci. 1992, 109 (1): 1937. 10.1016/00255564(92)900507
 21.
Sands PJ, Voit EO: Fluxbased estimation of parameters in Ssystems. Ecol Modeling. 1996, 93: 7588. 10.1016/03043800(95)002154.
 22.
Gutenkunst RN, Waterfall JJ, Casey FP, Brown KS, Myers CR, Sethna JP: Universally sloppy parameter sensitivities in systems biology models. PLoS Comput Biol. 2007, 3 (10): 18711878. 10.1371/journal.pcbi.0030189
 23.
Gutenkunst RN, Casey FP, Waterfall JJ, Myers CR, Sethna JP: Extracting falsifiable predictions from sloppy models. Ann N Y Acad Sci. 2007, 1115: 203211. 10.1196/annals.1407.003
 24.
Waterfall JJ, Casey FP, Gutenkunst RN, Brown KS, Myers CR, Brouwer PW, Elser V, Sethna JP: Sloppymodel universality class and the Vandermonde matrix. Phys Rev Lett. 2006, 97 (15): 150601 10.1103/PhysRevLett.97.150601
 25.
Brown KS, Sethna JP: Statistical mechanical approaches to models with many poorly known parameters. Phys Rev E Stat Nonlin Soft Matter Phys. 2003, 68 (2 Pt 1): 021904
 26.
Battogtokh D, Asch DK, Case ME, Arnold J, Schuttler HB: An ensemble method for identifying regulatory circuits with special reference to the qa gene cluster of Neurospora crassa. Proc Natl Acad Sci USA. 2002, 99 (26): 1690416909. 10.1073/pnas.262658899
 27.
Piazza M, Feng XJ, Rabinowitz JD, Rabitz H: Diverse metabolic model parameters generate similar methionine cycle dynamics. J Theor Biol. 2008, 251 (4): 628639. 10.1016/j.jtbi.2007.12.009
 28.
Barbano PE, Spivak M, Flajolet M, Nairn AC, Greengard P, Greengard L: A mathematical tool for exploring the dynamics of biological networks. Proc Natl Acad Sci USA. 2007, 104 (49): 1916919174. 10.1073/pnas.0709955104
 29.
Voit EO, Savageau MA: Powerlaw approach to modeling biological systems; II. Application to ethanol production. J Ferment Technol. 1982, 60: 229232.
 30.
Ramos A, Neves AR, Santos H: Metabolism of lactic acid bacteria studied by nuclear magnetic resonance. Antonie Van Leeuwenhoek. 2002, 82: 14. 10.1023/A:1020664422633
 31.
Neves AR, Pool WA, Kok J, Kuipers OP, Santos H: Overview on sugar metabolism and its control in Lactococcus lactis – the input from in vivo NMR. FEMS Microbiol Rev. 2005, 29 (3): 531554. 10.1016/j.femsre.2005.04.005
 32.
Voit E, Neves AR, Santos H: The intricate side of systems biology. Proc Natl Acad Sci USA. 2006, 103 (25): 94529457. 10.1073/pnas.0603337103
 33.
Metropolis N, Ulam S: The Monte Carlo Method. J Amer Stat Assoc. 1949, 44: 335341. 10.2307/2280232.
 34.
Voit EO, Ferreira AEN: Computational analysis of biochemical systems: a practical guide for biochemists and molecular biologists. 2000, Cambridge; New York: Cambridge University Press
 35.
Brown KS, Hill CC, Calero GA, Myers CR, Lee KH, Sethna JP, Cerione RA: The statistical mechanics of complex signaling networks: nerve growth factor signaling. Phys Biol. 2004, 1: 34. 10.1088/14783967/1/3/006.
 36.
Vilela M, Borges CC, Vinga S, Vasconcelos AT, Santos H, Voit EO, Almeida JS: Automated smoother for the numerical decoupling of dynamics models. BMC Bioinformatics. 2007, 8: 305 10.1186/147121058305
 37.
Alves R, Savageau MA: Systemic properties of ensembles of metabolic networks: application of graphical and statistical methods to simple unbranched pathways. Bioinformatics. 2000, 16 (6): 534547. 10.1093/bioinformatics/16.6.534
 38.
Daniels BC, Chen YJ, Sethna JP, Gutenkunst RN, Myers CR: Sloppiness, robustness, and evolvability in systems biology. Curr Opin Biotechnol. 2008, 19 (4): 389395. 10.1016/j.copbio.2008.06.008
 39.
Alves R, Savageau MA: Comparing systemic properties of ensembles of biological networks by graphical and statistical methods. Bioinformatics. 2000, 16 (6): 527533. 10.1093/bioinformatics/16.6.527
 40.
Voit EO, Almeida J, Marino S, Lall R, Goel G, Neves AR, Santos H: Regulation of glycolysis in Lactococcus lactis: an unfinished systems biological case study. Syst Biol (Stevenage). 2006, 153 (4): 286298.
 41.
Kimura S, Ide K, Kashihara A, Kano M, Hatakeyama M, Masui R, Nakagawa N, Yokoyama S, Kuramitsu S, Konagaya A: Inference of Ssystem models of genetic networks using a cooperative coevolutionary algorithm. Bioinformatics. 2005, 21 (7): 11541163. 10.1093/bioinformatics/bti071
 42.
Magnus JR, Neudecker H: Matrix differential calculus with applications in statistics and econometrics. 1988, Chichester [England]; New York: Wiley
 43.
Fisher J, Henzinger TA: Executable cell biology. Nat Biotechnol. 2007, 25 (11): 12391249. 10.1038/nbt1356
 44.
Palsson B: Systems biology: properties of reconstructed networks. 2006, New York: Cambridge University Press
 45.
Peifer M, Timmer J: Parameter estimation in ordinary differential equations for biochemical processes using the method of multiple shooting. IET Syst Biol. 2007, 1 (2): 7888. 10.1049/ietsyb:20060067
 46.
BalsaCanto E, Peifer M, Banga JR, Timmer J, Fleck C: Hybrid optimization method with general switching strategy for parameter estimation. BMC Syst Biol. 2008, 2: 26 10.1186/17520509226
 47.
Kitayama T, Kinoshita A, Sugimoto M, Nakayama Y, Tomita M: A simplified method for powerlaw modelling of metabolic pathways from timecourse data and steadystate flux profiles. Theor Biol Med Model. 2006, 3: 24 10.1186/17424682324
 48.
Goel G, Chou IC, Voit EO: System estimation from metabolic timeseries data. Bioinformatics. 2008, 24 (21): 25052511. 10.1093/bioinformatics/btn470
 49.
Ciliberti S, Martin OC, Wagner A: Robustness can evolve gradually in complex regulatory gene networks with varying topology. PLoS Comput Biol. 2007, 3 (2): e15 10.1371/journal.pcbi.0030015
 50.
Cullen AC, Frey HC: Probabilistic techniques in exposure assessment: a handbook for dealing with variability and uncertainty in models and inputs. 1999, New York: Plenum Press
 51.
Cobelli C, Distefano JJ: Parameter and Structural Identifiability Concepts and Ambiguities – a CriticalReview and Analysis. American Journal of Physiology. 1980, 239 (1): R7R24.
 52.
Ljung L, Glad T: On Global Identifiability for Arbitrary Model Parametrizations. Automatica. 1994, 30 (2): 265276. 10.1016/00051098(94)900299.
 53.
Vinga S, Thomaseth K, Lemos JM, Neves AR, H S, AT F: Structural analysis of metabolic networks: a case study on Lactococcus lactis. 8th Portuguese Conference on Automatic Control: 2008; Vila Real, Portugal. 2008, 566571.
 54.
Liebermeister W, Klipp E: Bringing metabolic networks to life: convenience rate law and thermodynamic constraints. Theor Biol Med Model. 2006, 3: 41 10.1186/17424682341
 55.
Ederer M, Gilles ED: Thermodynamically feasible kinetic models of reaction networks. Biophys J. 2007, 92 (6): 18461857. 10.1529/biophysj.106.094094
 56.
Price ND, Shmulevich I: Biochemical and statistical network models for systems biology. Curr Opin Biotechnol. 2007, 18 (4): 365370. 10.1016/j.copbio.2007.07.009
 57.
Ciliberti S, Martin OC, Wagner A: Innovation and robustness in complex regulatory gene networks. Proc Natl Acad Sci USA. 2007, 104 (34): 1359113596. 10.1073/pnas.0705396104
 58.
Lay DC: Linear algebra and it's applications. 2006, Boston: Pearson/AddisonWesley, 3
 59.
W Chen CS, He W: The DQ solution of geometrically nonlinear bending of orthotropic rectangular plates by using Hadamard and SJT product. Computers & Structures. 2000, 74 (1): 6574. 10.1016/S00457949(98)003204.
 60.
Marquardt DW: An algorithm for leastsquares of nonlinear parameters. SIAM J Appl Math. 1963, 11: 431441. 10.1137/0111030.
 61.
Tikhonov A: On the stability of inverse problems. Dokl Akad Nauk SSSR. 1943, 39 (5): 195198.
Acknowledgements
The authors acknowledge partial support by project DynaMo (PTDC/EEAACR/69530/2006; S. Vinga, PI) from the Portuguese Science Foundation (FCT) and the INESCID. We also would like to thank the anonymous reviewers for the constructive suggestions. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the sponsoring institutions.
Author information
Additional information
Competing interests
The authors declare that they have no competing interests.
Authors' contributions
MV and MAGMM conceived the core of the optimization method. SV participated in the analysis and systematization of the statistical method used in this work. EOV initiated the field of network identification with Ssystems and supervised activities leading to this paper. JSA conceived the ideas of automating the identification of Ssystems and creating a model pipeline. All authors contributed to in the preparation of the manuscript.
Electronic supplementary material
Authors’ original submitted files for images
Rights and permissions
About this article
Received
Accepted
Published
DOI
Keywords
 Monte Carlo
 Local Optimal Point
 Nonlinear Optimization Algorithm
 Propose Optimization Algorithm
 Degradation Term