# Parameter optimization by using differential elimination: a general approach for introducing constraints into objective functions

- Masahiko Nakatsui
^{1}, - Katsuhisa Horimoto
^{1, 2}Email author, - Masahiro Okamoto
^{3}, - Yasuhito Tokumoto
^{4}and - Jun Miyake
^{4, 5}

**4(Suppl 2)**:S9

https://doi.org/10.1186/1752-0509-4-S2-S9

© Horimoto et al; licensee BioMed Central Ltd. 2010

**Published: **13 September 2010

## Abstract

### Background

The investigation of network dynamics is a major issue in systems and synthetic biology. One of the essential steps in a dynamics investigation is the parameter estimation in the model that expresses biological phenomena. Indeed, various techniques for parameter optimization have been devised and implemented in both free and commercial software. While the computational time for parameter estimation has been greatly reduced, due to improvements in calculation algorithms and the advent of high performance computers, the accuracy of parameter estimation has not been addressed.

### Results

We propose a new approach for parameter optimization by using differential elimination, to estimate kinetic parameter values with a high degree of accuracy. First, we utilize differential elimination, which is an algebraic approach for rewriting a system of differential equations into another equivalent system, to derive the constraints between kinetic parameters from differential equations. Second, we estimate the kinetic parameters introducing these constraints into an objective function, in addition to the error function of the square difference between the measured and estimated data, in the standard parameter optimization method. To evaluate the ability of our method, we performed a simulation study by using the objective function with and without the newly developed constraints: the parameters in two models of linear and non-linear equations, under the assumption that only one molecule in each model can be measured, were estimated by using a genetic algorithm (GA) and particle swarm optimization (PSO). As a result, the introduction of new constraints was dramatically effective: the GA and PSO with new constraints could successfully estimate the kinetic parameters in the simulated models, with a high degree of accuracy, while the conventional GA and PSO methods without them frequently failed.

### Conclusions

The introduction of new constraints in an objective function by using differential elimination resulted in the drastic improvement of the estimation accuracy in parameter optimization methods. The performance of our approach was illustrated by simulations of the parameter optimization for two models of linear and non-linear equations, which included unmeasured molecules, by two types of optimization techniques. As a result, our method is a promising development in parameter optimization.

## Background

The investigation of network dynamics is a major issue in systems and synthetic biology [1]. In general, a network model for describing the kinetics of constituent molecules is first constructed with reference to the biological knowledge, and then the model is mathematically expressed by differential equations, based on the chemical reactions underlying the kinetics. Finally, the kinetic parameters in the model are estimated by various parameter optimization techniques [2], from the time-series data measured for the constituent molecules. While the computational time for parameter estimation has been greatly reduced, due to the improvement in calculation algorithms and the advent of high performance computers, the accurate numerical estimation of parameter values for a given model remains a limiting step. Indeed, the parameter values estimated by various optimization techniques are frequently quite variable, due to the conditions for parameter estimation, such as the initial values. In particular, we cannot always obtain the data measured for all of the constituent molecules, due to limitations of measurement techniques and ethical constraints. In this case, one of the issues we should resolve is that the parameters are estimated from the data for only some of the constituent molecules. Unfortunately, it is quite difficult to estimate the parameters in such a network model including unmeasured variables.

Boulier and his colleagues developed differential elimination [3], derived from the Roselfeld-Gröbner base [4]. Differential elimination rewrites a system of original differential equations into an equivalent system. The rewriting feature was applied to solve the parameter optimization issue, especially in network dynamics including unmeasured variables [3, 5], and in the applications, the equations rewritten by differential elimination were utilized to estimate the initial values for the parameter optimization, by Newton-type numerical optimization.

Here, we propose a new method for optimizing the parameters, by using differential elimination [3]. Our method partially utilizes a technique from a previous study [3], regarding the introduction of differential elimination into parameter optimization in a network including unmeasured variables. Instead of using differential elimination for estimating the initial values for the following parameter optimization, the equations derived by differential elimination are directly introduced as the constraints into the objective function for the parameter optimization. To validate the effectiveness of the constraint introduction, we performed simulations in two models of linear and nonlinear differential equations, where we assumed that the data for only one molecule among them were measured, by using two kinds of evolutionary optimization techniques. The accuracy of the parameter values estimated by the objective functions with and without the new constraints was compared. Finally, we discussed merits and pitfalls of our method in terms of its extension to more realistic and complex models.

## Results

We first describe a perspective of our method, and then the two models are analyzed to illustrate its performance. The two models were chosen from representative kinetic models for biological phenomena at the molecular level: one model (Model 1) is composed of two variables, analogous to molecular binding and dissociation, such as affinity binding in an antibody cross-link, and the other model (Model 2) is composed of four variables, analogous to a molecular reaction cascade, such as phosphorylation in signal transduction. Notably, we assumed that only one variable is measured among the variables in the two models.

### Overview of present method

The key point of this study is the introduction of new constraints obtained by differential elimination into the objective function, to improve the parameter accuracy. Following an explanation of differential elimination, the method of introducing the constraints is briefly described.

Differential algebra aims at studying differential equations from a purely algebraic point of view [6, 7]. Differential elimination theory is a sub theory of differential algebra [3], based on Rosenfeld-Gröbner [4]. The differential elimination rewrites the inputted system of differential equations to another equivalent system according to ranking (order of terms). Here, we provide an example of differential elimination, as shown below, according to Boulier [3, 5].

*x*

_{1}and

*x*

_{2}, in Fig. 1, which is described by the following system of parametric ordinary differential equations,,(1)

*k*

_{12},

*k*

_{21},

*k*

_{e}and

*V*

_{e}are some constants. Here, two molecules are assumed to bind according to Michaelis-Menten kinetics. The differential elimination then produces the following two equations equivalent to the above system.(2)

When we define the left sides of the above system as *C*_{1,t} and *C*_{2,t}, *C*_{2,t} is composed of *x*_{1}, its derivatives, and the parameters obtained by eliminating *x*_{2}, and *C*_{1,t} is composed of *x*_{1}, its derivatives, the parameters and *x*_{2}. Note that *x*_{2} in *C*_{1,t} can be expressed by *x*_{1}, its derivatives and the parameters in *C*_{2,t}. Then, the values of *C*_{1,t} and *C*_{2,t} can be calculated, if we have time-series data of *x*_{1}, and they would be zero, if all parameters were exactly estimated. Thus, *C*_{1,t} and *C*_{2,t} can be regarded as a kind of error function that expresses the difference between the measured and estimated data.

*E*. The parameter set is then estimated when the total relative error falls below a given threshold. However, the immense searching space of parameter values frequently hinders correct parameter estimation. To overcome this problem, we introduce the constraint between the estimate obtained by differential elimination (DE constraints),

*C*, into the objective function, i.e.,,(3)

where *α* is a weighting factor, which is approximately estimated by Pareto optimal solutions for *E* and *C*, and then is manually modified (see details in Methods).

### Model 1

*x*

_{AB}, was generated (Figure 2B), and two optimization techniques, genetic algorithm (GA) and particle swarm optimization (PSO), were applied to it to evaluate the effect of the introduction of differential elimination constraints (DE constraints) (see details in Additional File 1) into the objective function.

*k*

_{ p }and km were correctly estimated with the introduction, while the estimation of

*k*

_{ p }failed without the introduction. Indeed, the most frequent values estimated with the introduction (right side of Figure 3A) were found in the bins corresponding to the range between 0.045 and 0.055 for

*k*

_{ p }and between 0.45 and 0.55 for

*k*

_{ m }. In contrast, the most frequent values estimated without the introduction (left side of Figure 3A) were found in the range between 0.065 and 0.075 for

*k*

_{ p }, while those for

*k*

_{ m }were correctly estimated. By using PSO (Figure 3B),

*k*

_{ m }was correctly estimated with the introduction, but

*k*

_{ p }failed, while the estimations of both parameters failed without the introduction. Furthermore, another difference between the estimations with and without the introduction is the distribution form of the estimated values, although the numbers of trial successes in the optimization were different with and without the introduction (see details in Methods). As seen in Figure 3A and 3A, the values with the introduction were sharply distributed, while those without the introduction were widely distributed. The introduction of DE constraints contracted the parameter space to facilitate the estimation of the correct values. As a result, the parameter accuracy was improved by the new objective function with the introduction of DE constraints in Model 1.

### Model 2

*x*

_{1}, was generated (Figure 5B), and GA and PSO were applied to it to evaluate the effect of the introduction of DE constraints (see details in Additional File 2) into the objective function.

*k*

_{31}and

*k*

_{41}failed without the introduction (left side). By using PSO (Figure 6B), all three parameters were also correctly estimated with the introduction (right side), while the estimations of

*k*

_{41}failed without the introduction (left side). Furthermore, the features of the distribution forms of the estimated values were similar to those in Model 1 (Figure 3). As seen in Figure 6A and 6B, the distribution of the estimated values with the introduction was sharp (right side), while that without the introduction was wide (left side). As a result, the parameter accuracy was also improved by the new objective function to contract the parameter space with the introduction of DE constraints in Model 2.

## Discussion

The introduction of DE constraints into the objective function clearly improved the parameter accuracy. Indeed, the parameter value sets were correctly estimated by the introduction of DE constraint into the objective function, while they were falsely estimated without the introduction. Furthermore, the parameter sets with the introduction were sharply distributed near the correct values in all cases, in contrast to the wide distribution without the introduction. In general, the derivatives included the information on the curve form of the measured time-series data, such as slope, extremal point and inflection point. This indicates that the new objective function estimates the difference of not only the values but also the forms between the measured and estimated data, while the standard objective function estimates only the value difference. Note that the DE constraint is rationally reduced from the original system of differential equations for a given model in a mathematical sense. Thus, our approach is expected to be a general approach in parameter optimization for improving the parameter accuracy.

As expected, the new objective function requires more computational time, in comparison with an objective function with only a standard error function, due to the increase of the functions in DE constraints. Indeed, the computational time of our method was larger than that of the standard method in Models 1 and 2; the computational times for the standard method and our method were 0.4 and 2.3 hours in Model 1, and 0.03 and 0.22 hours in Model 2 (32 CPU’s of Intel(R) Xeon(R) X5550 2.67GHz). In addition to the computational time, a pitfall of our method is the equation size of DE constraints. In the equivalent systems, the number of terms frequently increases (see Additional file 3), and this may result in the difficulty of the application of our method to a complex or large model. Although we do not still reach a clear conclusion to overcome the difficulty, two ways can be considered. One way is an approximation method and the other is a mathematical manipulation method. As for the former method, in the DE constraints, the terms with a higher order of derivatives in the differential equations appeared frequently in the equivalent system (see Additional files 2 and 3). The magnitude of the estimated values of the higher order derivatives was relatively smaller than those of the lower order derivatives. If the estimation of terms with higher order derivatives can be neglected, then the computational time will be reduced. As for the latter method, we can use some equation-simplification methods by symbolic computation (personal communication from Drs. A. Sedoglavic, F. Lemaire and F. Boulier of Lille University). Indeed, the size of DE constraints for the negative feedback model with oscillation was reduced from 7.4MB obtained by the pure differential elimination in present procedure to 0.1MB after the equation simplification by symbolic computation (data not shown). Further studies will be needed to shorten the computational time by the combination of the approximation and the simplification of the DE constraints.

One possible use of our method is its application to network inference without known structure. Since the present method is designed with the assumption of a known network structure, the application range of our method to network inference is naturally restricted. However, our method can select the most possible network structure among the networks with similar structures. Indeed, we designed a similar procedure for evaluating the network structures with measured data [10]. In our previous approach, we adopted the transformation of a system of differential equations into the equivalent system of algebraic equations by Laplace transformation. In this case, the system must be linear, due to the Laplace transformation. Furthermore, the numeric optimization in the previous approach frequently faces difficulties, due to the existence of the pole in the Laplace domain. In contrast, these pitfalls are overcome in the present method, by introducing the constraints by differential elimination. This supports the application of the present method to the model selection issue.

Various models for describing biological phenomena are available [11]. In particular, several feedback models are important for describing the biological phenomena [12, 13]. Although the performance of our approach for the two representative models in biological phenomena was tested in this study, further tests for the performance of the DE constraint introduction remain for the models that are important in systems and synthetic biology. In the near future, we will report the evaluation of our approach in the cases of various models, in addition to the reduction of computational time and the trials of model selection.

## Conclusions

The introduction of the constraints by using differential elimination was effectively improved the parameter accuracy in two models of linear and nonlinear equations, especially when we assumed that unmeasured variables were included, by two optimization techniques. This clearly indicates that the ability of our method for estimating the parameter values was far superior to that of various methods with the standard error function. Although the present study focused on two simple models, our method is a feasible approach for parameter estimation in network dynamics.

## Methods

### Analyzed models

We assume that the model expresses the binding and dissociation between two molecules, and that only one complex, *x*_{
AB
} , can be measured.

We assume that the molecules, *x*_{2}, *x*_{3}, and *x*_{4}, activate *x*_{1} with linear relationships, and that only one molecule, *x*_{1}, can be measured.

*X*

_{e}(

*t*), is expressed as follows:,

where *X*(*t*) is a value at time *t* in original curve of Figures 2 and 5, *Rn* is random variable according to the standard normal distribution, and *c* was set to 0.666.

### Optimization techniques

Two well-known parameter optimization techniques, the genetic algorithm (GA) [15–19] and the particle swarm optimization (PSO) [20, 21], were used. In the parameter optimization, two thresholds were set to stop the optimization: the average value of the error function over time points, *E*/*T*, and the number of generations per optimization. In this study, we performed the optimization 200 times in both techniques, and the thresholds of *E*/*T* were set to 0.01 for Model 1 and 0.001 for Model 2, and the threshold for generation number was set to 2000. As a result, the numbers of successes by 200 trials were 200 without DE constraints and 51 by GA and 11 by PSO with DE constraints, for Model 1, and 200 for all cases for Model 2.

### Introduction of the new constraints into the objective function

*x*

^{ c }

_{ i,t }is the time-course data at time

*t*of

*x*

_{ i }calculated by using the estimated parameter values, and

*x*

^{ m }

_{ i,t }represents the measured data at time

*t*. The sum of the absolute value of the relative error between

*x*

^{ c }

_{ i,t }and

*x*

^{ m }

_{ i,t }gives the total relative error,

*E*, as a standard error function, i.e.,,(6)

where *N* and *T* are the number of variables and the time points, respectively: *N* was 2 for Model 1 and 4 for Model 2, and *T* was 100.

*C*

_{DE}, as the linear combination of the equations in the equivalent system, as follows:,(7)

where *L* and *T* are the numbers of equivalent equations and time points, respectively: *L* was 2 for Model 1 and 5 for Model 2.

where *α* the a weight of two functions, which is approximately estimated by a Pareto optimal solutions for *E* and *C* and then is manually modified. In the present study, *α* was set to 0.1 in Model 1 and 0.9999999 in Model 2. As a result, our computational task is to determine a set of parameter values that minimize to *OF*.

### Implementation of differential elimination

All of the symbolic computations for the differential elimination were performed using the *diffalg* package of MAPLE 10. In the performance of differential elimination, the ranking of variables was: *x*_{A} ≻ *x*_{B} ≻ *x*_{AB} in Model 1 and P(Pool) ≻ *x*_{4} ≻ *x*_{3} ≻ *x*_{2} ≻ *x*_{1} in Model 2. Subsequently, we converted the form of the polynomial equations derived by differential elimination to the Java code by using the *CodeGeneration* feature in Maple 10.

## Declarations

### Acknowledgements

This work was partly supported by a project grant, ‘Development of Analysis Technology for Gene Functions with Cell Arrays’, from The New Energy and Industrial Technology Development Organization (NEDO). KH was partly supported by a Grant-in-Aid for Scientific Research on Priority Areas "Systems Genomics" (grant 20016028) and for Scientific Research (A) (grant 19201039) from the Ministry of Education, Culture, Sports, Science and Technology of Japan. In particular, the authors would like to express their gratitude to Drs. Alexander Sedoglavic, Francois Lemaire, and Francois Boulier of Lille University, for valuable discussions during the course of this work.

This article has been published as part of *BMC Systems Biology* Volume 4 Supplement 2, 2010: Selected articles from the Third International Symposium on Optimization and Systems Biology. The full contents of the supplement are available online at http://www.biomedcentral.com/1752-0509/4?issue=S2

## Authors’ Affiliations

## References

- Kitano H: System Biology: A Brief Overview. Science. 2002, 1662-1664.Google Scholar
- Nocedal J, Wright SJ: Numerical Optimization. 1999, Springer-Verlag, NYView ArticleGoogle Scholar
- Boulier F: Differential Elimination and Biological Modelling. Johann Radon Institute for Computational and Applied Mathematics (RICAM) Book Series. 2007, 2: 111-139.Google Scholar
- Boulier F, Lazard D, Ollivier F, Petitot M: Representation for the radical of a finitely generated differential ideal. Proceedings of ISSAC. 1995, 158-166.Google Scholar
- Boulier F: Differential algebra and system modelling in cellular biology. Proceedings of Algebraic Biology. 2008, 22-39.View ArticleGoogle Scholar
- Ritt JF: Differential Algebra. 1950, Dover Publications Inc., NYGoogle Scholar
- Kolchin EE: Differential Algebra and Algebraic Groups. 1973, Academic Press, NYGoogle Scholar
- Powell MJD: An efficient method for finding the minimum of a function of several variables without calculating derivatives. Computer Journal, . 1954, 7: 142-162.Google Scholar
- Powell MJD: On the calculation of orthogonal vectors. Computer Journal, . 1968, 11: 302-304. 10.1093/comjnl/11.3.302.View ArticleGoogle Scholar
- Yoshida H, Nakagawa K, Anai H, Horimoto K: An Algebraic-Numeric Algorithm for the Model Selection in Kinetic Networks. Proceedings of the 10th CASC. 2007, 4770: 433-447. , LNCSGoogle Scholar
- Tyson Jhon, Katherine Chen, Novak Bela: Sniffers, buzzers, toggles and blinkers: dynamics of regulatory and signalling pathways in the cell. Current opinion in cell biology. 2003, 14 (2): 221-231.View ArticleGoogle Scholar
- Yung-Keun Kwon, Kwang-Hyun Cho: Quantitative analysis of robustness and fragility in biological networks based on feedback dynamics. Bioinformatics. 2008, 24 (7): 987-994. 10.1093/bioinformatics/btn060View ArticleGoogle Scholar
- Novák B, Tyson Jhon: Design principles of biochemical oscillators. Nat. Rev. Mol. Cell Biol. 2008, 9 (12): 981-991. 10.1038/nrm2530PubMedPubMed CentralView ArticleGoogle Scholar
- Box GEP, Muller ME: A Note on the Generation of Random Normal Deviates. The Annals of Mathematical Statistics. 1958, 29 (2): 610-611. 10.1214/aoms/1177706645.View ArticleGoogle Scholar
- Holland JH: Adaptation in Natural and Artificial System. 1975, The University of Michigan Press, Ann Arbor, MIGoogle Scholar
- Goldberg DD: Genetic Algorithms in Search, Optimization and Machine Learning. 1989, Addison-Wesley Longman Publishing Co., Inc., Boston, MAGoogle Scholar
- Jonikow CZ, Michalewicz Z: An Experimental Comparison of Binary and Floating Point Representations in Genetic Algorithms. Proceedings of the Fourth International Conference on Genetic Algorithms. 1991, 31-36.Google Scholar
- Ono I, Kobayashi S: A real-coded genetic algorithm for function optimization using unimodal distribution crossover. Proceedings of 7th ICGA. 1997, 249-253.Google Scholar
- Satoh H, Ono I, Kobayashi S: A new generation alternation model of genetic algorithm and its assessment. J. of Japanese Society for Artificial Intelligence. 1997, 15 (2): 743-744.Google Scholar
- Eberhart R, Kennedy J: A New Optimizer Using Particle Swarm Theory. Proc. of Sixth International Symposium on Micro Machine and Human Science (Nagoya Japan), IEEE Service Center, Piscataway, NJ. 1995, 39-43.View ArticleGoogle Scholar
- Kennedy J, Eberhart R: Particle swarm optimization. Proc. IEEE International Conference on Neural Networks (Perth, Australia) IEEE Service Center, Piscataway, NJ. 1995, IV: 1942-1948.Google Scholar

## Copyright

This article is published under license to BioMed Central Ltd. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.