Open Access

Information theoretical quantification of cooperativity in signalling complexes

BMC Systems Biology20093:9

https://doi.org/10.1186/1752-0509-3-9

Received: 21 May 2008

Accepted: 16 January 2009

Published: 16 January 2009

Abstract

Background

Intra-cellular information exchange, propelled by cascades of interacting signalling proteins, is essential for the proper functioning and survival of cells. Now that the interactome of several organisms is being mapped and several structural mechanisms of cooperativity at the molecular level in proteins have been elucidated, the formalization of this fundamental quantity, i.e. information, in these very diverse biological contexts becomes feasible.

Results

We show here that Shannon's mutual information quantifies information in biological system and more specifically the cooperativity inherent to the assembly of macromolecular complexes. We show how protein complexes can be considered as particular instances of noisy communication channels. Further we show, using a portion of the p27 regulatory pathway, how classical equilibrium thermodynamic quantities such as binding affinities and chemical potentials can be used to quantify information exchange but also to determine engineering properties such as channel noise and channel capacity. As such, this information measure identifies and quantifies those protein concentrations that render the biochemical system most effective in switching between the active and inactive state of the intracellular process.

Conclusion

The proposed framework provides a new and original approach to analyse the effects of cooperativity in the assembly of macromolecular complexes. It shows the conditions, provided by the protein concentrations, for which a particular system acts most effectively, i.e. exchanges the most information. As such this framework opens the possibility of grasping biological qualities such as system sensitivity, robustness or plasticity directly in terms of their effect on information exchange. Although these parameters might also be derived using classical thermodynamic parameters, a recasting of biological signalling in terms of information exchange offers an alternative framework for visualising network cooperativity that might in some cases be more intuitive.

Background

A cellular pathway, whether enzymatic or signal transducing, can in a simplistic manner be described as a causal relationship between an environmental signal (such as nutrients, osmolytes or hormones) and a cellular response (generally through gene regulation). Cellular signals are mediated through a series of successive protein-protein interactions that bridge spatial and topological boundaries (between the plasma membrane and cell nucleus for example) and that allow for crosstalk between different pathways [1, 2]. This protein-based modular strategy achieves integrated cellular responses that are both specific and at the same time tuned to global environmental and cellular requirements. This specificity is organized through the cooperativity between the members of the complex and the introduction of temporal and spatial constraints on the expression levels of the different members of the signalling pathways. Over- or under-expression, for instance, of the signalling components may have disastrous effects on the cellular phenotype, e.g. the development of cancer.

Cooperativity is a thermodynamic concept that is used in different biochemical contexts [36]. Here this notion refers to the formation of multi-protein complexes with non-additive free-energies of assembly, i.e. complexes for which the stability of the final assembly is higher than the sum of all individual binary association [6]. A classic way to study cooperativity is by the analysis of a thermodynamic cycle [7]. Consider an assembly process that involves three proteins A, B and C that together form a ternary complex ABC, where B acts as an adaptor protein providing a separate binding surface for each of the two other molecules (see left panel Figure 1). Two alternative routes can then create the complex ABC: Either the binary complex AB is formed first and C binds afterwards or first the proteins B and C are joined before A is added. Every pair-wise reaction between the proteins in isolation and between individual proteins and partially formed complexes is annotated by a rate of dissociation (Kd), specifying the likelihood of that particular assembly/disassembly step. Here there are four dissociation constants: KdA-B, KdB-C, KdAB-Cand KdA-BC, where the latter dissociation constants refer to the dissociation of C from ABC and A from ABC respectively. Since it is known that the overall thermodynamics and free energies for both routes to construct ABC is the same around the cycle (KdA-BKdAB-C= KdB-CKdA-BC), one can simply determine the cooperativity of the system by comparing the energy changes when one of the proteins is already bound or not [6]. For instance, when the likelihood of binding of C is enhanced when A is already bound to B, then there is positive cooperativity, resulting in bigger gain in energy when comparing it with the same reaction without the presence of A. When the binding of C to B is inhibited by the presence of A, then there is negative or anticooperativity, producing in turn a decrease in the free energy when comparing it to the process without A. When the presence of A in the complex does not influence the binding affinity of C, then there is no cooperativity or independent binding.
Figure 1

Abstract ternary protein complex. (A) The protein B in this abstract complex acts as a communication pathway between the two other proteins A and C. Binding protein A sends information over the pathway λ to the binding site of protein C, facilitating the binding of protein C. (B) Like a communication channel, the ternary complex can be described by a number of conditional probabilities. The conditional probabilities P(C = 0|A = 0) and P(C = 1|A = 1), describe the accuracy of the communication channel, i.e. the likelihood that a given output signal corresponds to the appropriate input signal. A second set of probabilities, P(C = 1|A = 0) and P(C = 0|A = 1), describes the intrinsic noise of the communication channel, i.e. the likelihood that a given input signal is not correctly conveyed.

Even though this scheme identifies the presence and type of cooperation in the assembly process, it does not shed light on the molecular concentrations, possibly reflecting the intracellular conditions, required for efficient regulation or coordination between a pathway's active (ABC) and inactive (B) state. Here we provide an information theoretical method that, in the same spirit as the Hill and Scatchard plots [8], identifies and quantifies cooperativity in macromolecular assemblies and visualizes for a spectrum of concentrations when optimal coordination is obtained for the given experimental data. Different from those established methods, our approach goes beyond multiple bindings of the same ligand to a homogeneous oligomer (as in the binding of oxygen to haemoglobin [9]): We consider here the construction of heterogeneous protein assemblies mediated by multiple binding surfaces on adaptor proteins. As such, and as far as we are aware, this method provides an original and novel approach for the analysis of the cooperativity in macromolecular complexes that are part of some signalling cascade.

Results and discussion

General description of the approach

In analogy with cellular pathways, each protein in a cellular network can be considered as an element receiving an input signal (from upstream ligands) and generating an output signal (towards downstream effectors). Hence, we can reinterpret the ternary protein complex ABC as an instance of communication over a noisy channel [10, 11], where protein B provides the communication channel through which information is exchanged between upstream ligand A and downstream effector C (See left panel in Figure 1). Like all noisy communication channels [10, 11] the protein complex ABC is determined by a set of conditional probabilities relating input to output (See right panel in Figure 1). A first set of probabilities, P(C = 0|A = 0) and P(C = 1|A = 1), describe the accuracy of the communication channel, i.e. the likelihood that a given output signal corresponds to the appropriate input signal. In other words, if no input signal is given, meaning A is not bound to B (A = 0) then C should also be not bound to B (C = 0), and vice versa. A second set of probabilities, P(C = 1|A = 0) and P(C = 0|A = 1), describes the intrinsic noise of the communication channel, i.e. the likelihood that a given input signal is not correctly conveyed. Thus inappropriate transmission occurs when either C is bound to B (C = 1) even though A was not bound before (A = 0) or when A is bound (A = 1) and C is not (C = 0). These probabilities describe a partition of the initial protein concentrations [A]0, [B]0 and [C]0 over all possible association states of the system including A, B and C in isolation, the binary complexes AB and BC and the ternary complex ABC. Next to the total concentrations [A]0, [B]0 and [C]0, the steady state that is defined by this partitioning depends on the different dissociation constants including KdA-B, KdB-C, KdAB-Cand KdA-BC. Given the steady state concentrations, the individual and conditional probabilities may be obtained (see Equation (1) in Methods). These probabilities are necessary to derive Shannon's entropy (H(A), H(C), ...) and subsequently the mutual information I(A;C) exchanged between A and C over channel B (see Equation (3) in Methods).

We quantify the degree of cooperativity of the system by the amount of information that is exchanged between the elements of the complex. In terms of the protein concentrations, this mutual information expresses how well the ratio of the steady-state concentrations of the ternary complex ABC and the free adaptor protein B are balanced while at the same time requiring the concentration of both binary complexes (AB and BC) to be as low as possible. So, on the one hand, low information exchange corresponds to an equilibrium situation where the protein (B) and complex concentrations (ABC) are out of balance or where to many intermediate complexes are present making it hard for the biological system to perform its function. On the other hand, high amounts of information exchange correspond to an optimized system where all members achieve the required coordination to switch efficiently between active and inactive states of the cellular process. Note, that the approach described here for a ternary protein complex can be further generalised to describe communication channels having multiple inputs or outputs (i.e. to study signal integrators or differentiators). In that case the mutual information between the different components needs to be deduced by a multivariate approach (see Methods) [11, 12]. It is also important to note that the mutual information does not change from swapping the input with the output components, i.e. I(A;C) = I(C;A).

Biophysical model system

To clarify the biophysical meaning and illustrate our method we here describe the information exchange over a part of the p27 regulatory pathway. The p27 pathway controls the degradation of the cyclin-dependent kinase 2 (Cdk2) inhibitor p27 [1316] thereby playing an important role in cell cycle progression [17, 18]. In particular, phosphorylation of p27 triggers Cks1-mediated binding of p27 to Skp2. As Skp2 is part of the SCFSkp2 ubiquitin ligase this results in p27 degradation and cell cycle progression. In a recent study [19], the assembling mechanism for part of the SCFSkp2 multiprotein complex has been analyzed in order to understand 1) how and in which order the different units assemble and 2) how the specific order of this process influences the mutual affinities between the components and intermediately formed complexes. Seeliger et al. [19] showed that the Skp2-Cks1 complex increases the affinity of Cks1 for the Cdk2 inhibitor p27 a 100-fold. Additional inclusion of Cdk2 increases the affinity for p27 even more. Through mutational analysis the authors also showed long-range coupling between distant functional sites in Cks1, making it a principal example how adaptor proteins can play a central role in tightly controlling the assembly of a critical complex. As a consequence, it forms a biophysically meaningful case to investigate the communication between the different binding sites of the Cks1 structure in terms of Shannon's information theory (see Methods). Note here that Shannon's information theory can also be used to derive the communication pathway in Cks1. We recently demonstrated this lower-level analysis for the SH2 domain of Fyn [20]. Given the appropriate structural data, the same analysis could occur which should reveal the communication between the three binding sites [21].

The biophysical data obtained in [19], i.e. the dissociation constants, is used to perform the current analysis, (see also Table 1 for the data). The thermodynamic cycle including the adaptor protein Cks1 (acting as component B) [21], the proteins Skp2 (acting as component C) and p27 (acting as component A) produced from this data shows that both paths around the cycle are cooperative: Having Skp2 bound to Cks1 makes it easier for p27 to bind and vice versa. In a first step, we focus on the thermodynamic cycle for the formation of this ternary complex p27-Cks1-Skp2 (see Methods). Since in vivo p27 is bound to Cdk2, we will in a second step consider the quaternary complex Cdk2-p27-Cks1-Skp2. In that case two signals (Cdk2 and p27) are integrated and conveyed over the communication channel Cks1. As the mutual binding affinities of this system, i.e. KdSkp 2-Cks 1, KdCks 1-p 27, KdSkp 2-Cks 1p 27and KdSkp 2Cks 1-p 27, have been determined experimentally (see [19] and Table 1), we can quantify the information exchange between the input and output components of the system and study the transmission efficiency, meaning under which conditions we observe the highest degree of cooperativity, of the adaptor protein Cks1 under a wide range of chemical potentials (see Methods). Note that only one of the dissociation constants, KdSkp 2-Cks 1p 27or KdSkp 2Cks 1-p 27, is required for the derivation of the different steady-state concentrations (see Methods).
Table 1

Dissociation constants between Cks1 and the different subunits.

Protein or complex

Binding partner

Kd (μM)

Cks1

p27

70

Cks1-Skp2

p27

0.47

Cks1-Cdk2

p27

120

Cks1-Cdk2-Skp2

p27

0.15

Cks1

Skp2

0.014

Cks1-Cdk2

Skp2

0.07

Cks1-p27

Skp2

0.0001

Cks1-Cdk2-p27

Skp2

0.0001

Cks1

Cdk2

1.5

Cks1-Skp2

Cdk2

8

Cks1-p27

Cdk2

3

Cks1-Skp2-p27

Cdk2

2.6

Dissociation constants in bold were derived experimentally. The remaining values were obtained through the thermodynamic cycles [19].

How much information is exchanged in the p27-Cks1-Skp2 complex?

Figure 2 (left panel) shows the phase-space of the degree of cooperativity of the ternary p27-Cks1-Skp2 complex as it represents the Cks1-mediated information exchange between p27 and Skp2 over a range of concentrations varying between 0.0 μM and 50 μM for p27 and 0.0 μM and 0.2 μM for Skp2, whereas the concentration of Cks1 has been kept constant at 0.1 μM. For each concentration distribution {[p27], [Skp2]} (Both ranges for Skp2 and phosphorylated p27 were discretized into 100 values each) the mutual information is calculated (see Methods), producing a matrix of information values. As argued earlier, mutual information expresses to what extent the proteins in the assembly properly coordinate their actions to achieve efficient switching between active and inactive states. Concretely when information about the association between Skp2 and Cks1 proteins is independent of the association of p27 and Cks1 proteins, then mutual information I(Skp2, p27) equals 0 bits (no cooperativity), turning the matrices in Figure 2 completely blue when this is the case. Conversely, when the association of all Skp2 proteins with Cks1 uniquely defines the associations of phosphorylated p27 then mutual information I(Skp2, p27) equals 1 bit (full cooperativity). This would mean that no binary assemblies are present, which is for most biochemical equilibria unlikely due to the underlying kinetics. Thus the intracellular process needs to find a good balance between the degree of cooperativity and the effectiveness of the switching mechanism.
Figure 2

Phase-space of cooperativity for the Cks1 adaptor protein. Both contour diagrams show the mutual information for different concentrations of Cks1, Skp2 and phosphorylated p27 (left panel) and for different concentrations of Cks1, Skp2, Cdk2 and phosphorylated p27 (right panel). In both panels, the concentration of Cks1 is kept fixed (0.1 μM) and the concentration of Skp2 and p27 vary between 0.0–0.2 μM and 0.0–50 μM respectively. In the right panel, Cdk2 varies together with p27, meaning that we assumed [p27] = [Cdk2] for all combinations. As can be observed, the signal between Skp2 and phosphorylated p27 is clearly constrained by the input concentrations. Moreover, when adding Cdk2, as shown in the right plot, the signal is reinforced.

As can be seen in Figure 2, only a small part of the phase space displays some or significant degree of cooperativity. In the cooperative regime the assembly of Cks1 to p27 will be conditional on the concentration of Skp2 and vice versa. In other words, binding of p27 to Cks1 and recruitment to the SCFSkp2 ubiquitin ligase machinery will be mutually dependent events. Outside of this regime the assembly of these elements is still possible, but as under these conditions the fraction of bound protein is no longer influenced by changes in the concentration of the other, information exchange becomes very noisy. Moreover, from the difference in the extent for which each protein, Skp2 and p27, shows information exchange one can derive that it is Skp2 that forms the natural input signal for this regulatory process: It controls the switching in an effective manner.

How does channel concentration affect robustness of the system?

Interestingly, although the area of maximum cooperativity of p27-Cks1-Skp2 represents only a minor part of the phase space, it displays a relatively slow decline for increasing p27 concentration (see also Figure 3. left panel). This relatively broad, although suboptimal response curve, gives a measure for the robustness of the system to extrinsic noise due to fluctuations in input or output ligand concentrations. This is not the case for Skp2. As can be seen Figure 3 (right panel), the responsive area for Skp2 is very tightly defined (given 0.1 μM concentration for Cks1). Hence variations in Skp2 strongly influence the responsiveness of the biochemical system.
Figure 3

Analysis of information exchange at Cks1 capacity. In both plots, the mutual information (green line) is shown. In the left panel, it is visualized for optimal [Cks1]* and [Skp2]* and varying [p27]. In the right panel, the same information is shown for optimal [Cks1]* and [p27]* and varying [Skp2]. In both plots, the blue striped line marks the concentrations of p27 (left panel) and of Skp2 (right panel) where the channel's capacity is obtained. In the left panel, the error probabilities f and g are added and are shown to be equal when the optimal value of mutual information is achieved. Both error probabilities intersect around f = g = 0.075 for [p27]5.79 μM. In the right panel, the probabilities that Skp2 is not bound to Cks1 and p27 is bound to Cks1 were added.

The robustness towards the Skp2 concentration increases as the concentration of Cks1 increases, as is shown in Figure 4. Increasing the concentration of Cks1 from 0.01 μM and 0.3 μM results in an increase of the cooperative area, particularly for the concentration of Skp2. This effect of Cks1 on the concentration of Skp2 makes sense biochemically since it is know that the expression of Cks1 alternates in parallel with the concentration of Skp2 when passing through the cell cycle [21]. Even though the area for Skp2 increases, the concentrations of p27 for which cooperativity is high remains the same. Moreover, even though the actual concentration for Skp2 changes, the maximum mutual information, or the capacity, remains always the same for this ternary system (see Figure 4).
Figure 4

Cooperativity phase-space for other concentrations of Cks1. This figure extends the results shown in Figure 2, left panel. We show here that increasing the concentration of Cks1 (from 0.01 μM to 0.3 μM) results in an increase of cooperativity of the system, specifically for the concentrations of Skp2. Yet this increase has no effect on the maximum amount of information that is exchanged.

How much information is exchanged in the Cdk2-p27-Cks1-Skp2 complex?

The right panel of Figure 2 shows the cooperativity profile of the quaternary complex Cdk2-p27-Cks1-Skp2. In this case both Cdk2 and p27, which are associated with a 1:1 stochiometry can be considered as two input signals, which are combined over the Cks1 adaptor to elicit Skp2 binding. The multivariate approach to mutual information [11, 12] (see Methods) makes it possible to analyse the different components that define the information flow in this quaternary complex. As can be expected, the integration of the two signals gives rise to an increased cooperativity of the system, but also to a broader maximum indicating a more robust system response (see Figure 2, right panel). In addition, we can deconstruct the transmission within the complex Cdk2-p27-Cks1-Skp2 into the mutual information between Skp2 and p27 (Figure 5, top left panel), Skp2 and Cdk2 (Figure 5, top right panel) and the effect of integrating both signals, called the interaction information A(Skp2, Cdk2, p27) [12] (Figure 5, bottom left panel). Even though the majority of the communication occurs between Skp2 and p27, the integration of these two signals, A (see Methods), shows how the concentration of Cdk2 affects the communication between Skp2 and p27: First for low initial concentrations of both proteins, knowledge about Cdk2 inhibits slightly the information exchange over the cooperative channel. As the concentration of Skp2 increases, the transmission is amplified resulting in the higher capacity that is shown in the right panel of the Figure 2. Consequently, the contour plot of A(Skp2, Cdk2, p27) gives a quantitative interpretation of how these three proteins affect the communication over the channel Cks1.
Figure 5

Deconstruction of the information exchange in the quaternary complex Cdk2-p27-Cks1-Skp2. This figure shows the mutual information between Skp2 and p27 (top left panel), Skp2 and Cdk2 (top right panel) and the effect of integrating both signals called mutual information (bottom left panel). The majority of the transmission occurs between the proteins Skp2 and p27. There is little interaction between Skp2 and Cdk2. Yet the stochiometric relation between p27 and Cdk2 modulates the signal in such a way that more information can be exchanged.

Channel capacity and noise of the p27-Cks1-Skp2 complex

The maxima of these contour diagrams represent the capacity of the system, i.e. the maximum amount of information that can be transmitted over the channel with an arbitrary small probability of error [22]. As can be seen in Figure 2 (left panel) channel capacity is achieved for the optimal input distribution {[p27], [Skp2]}* = {5.8 μM, 0.0512 μM} (relative to the channel concentration [Cks1]* = 0.1 μM) for the ternary complex p27-Cks1-Skp2. At these concentrations ~0.61 bits of information is received as output for every bit of input. As previously argued, this capacity remains the same even when the concentration of Cks1 changes. In terms of cooperativity, this means that even though the system is not fully cooperative, the balance between the assembled degradation system and independent adaptor protein is rather efficient while at the same time the intermediate complexes are few. Hence the capacity of the system is limited by the molecular concentrations that can be attained at steady state.

Figure 3 (left panel), shows further that optimal capacity is achieved when the error probabilities f and g are approximately equal, making the channel symmetric. The error probabilities intersect at f = g = 0.075 for [p27] = 5.8 μM. This fact follows from our previous argumentation that the highest mutual information is attained when both the ratio of complete assembly and individual adaptor protein is well balanced and the concentrations of binary complexes are as low as possible. This small error value shows that the cooperative channel within Cks1 is a very efficient channel. Even more, for increasing values of [Skp2], the capacity of the channel will never go beyond this point. This result is shown in Figure 6. When following the sequence of plots from the top left to the bottom right, one can observe that, while the concentration of Skp2 increases, magnitude of the mutual information (green line) increases until it reaches a maximum (centre plot in Figure 6). Afterwards the magnitude decreases again. Additionally we visualized the two error-probabilities f (black line) and g (red line) representing the error of reading a signal when no input was given (when p27 binds to Cks1 without prior binding of Skp2) and the error of giving a signal and not reading the output (when Skp2 binds to Cks1 without posterior binding of p27). As can be seen the concentration of Skp2 has no (or very little) influence on these probabilities since there is almost no change in both lines when comparing the different plots. The concentration of p27, on the other hand, has a strong influence on these values. When [p27] = 0 μM, f becomes 0 and g becomes 1 (see Methods). For very high values of [p27] (like 50 μM) the value of f reaches almost 50%, which indicates that the channel is completely random, and g becomes almost zero. This extreme situation is due to an over-representation of the p27 protein. As a consequence, there is more than enough p27 to bind to Cks1 without having any input signal (Skp2).
Figure 6

Skp2 regulates the mutual information and noise in the communication. We analysed the relationship between the mutual information for fixed concentrations Cks1. The concentration of Skp2 varies between plots from 0 μM and 0.2 μM. The concentration of p27 varies in each plot between 0 μM and 20 μM. Following the plots from top left to bottom right, the amount of Skp2 increases. As can be observed, the amount of mutual information (green line) also increases until some maximum is reached (centre plot). At this same point the error probabilities f and g intersect.

We added in each plot a blue line that marks the concentration of p27 where the maximum information transmission is found ([p27]* = 5.79 μM and [Skp2]* = 0.0512 μM) for all combinations of [Skp2] and [p27] (see also Figure 2 for this maximum). We immediately notice here that this vertical line lies at the intersection of the two error probabilities (left panel Figure 3) as well as the 50% mark of the probability of having p27 bound to Cks1 (right panel Figure 3). Consequently, when the channel is symmetric the SCFSkp2model system studied here reaches maximal information exchange.

By adding Cdk2 in the system, the capacity becomes ~0.75 bits (see Figure 2 right panel), confirming the effect of including Cdk2 as discussed in ref. 10. Moreover, the area of optimal response increases (see red area in Figure 2 right panel) allowing a wider range of useful Skp2 and p27 concentrations.

Conclusion

All these results show that, given the binding affinities at equilibrium and the overall concentrations of the different components, mutual information quantifies for which protein concentrations the systems' cooperativity, or more specific its coordination, is optimal. Our analysis clearly shows (see Figure 2) that both Skp2 and Cks1 concentrations are crucial parameters to obtain proper regulatory behaviour in the p27 degradation system. If binding is independent then no information exchange would be observed. Moreover, this narrow range for Skp2 remains operational for a wide range of p27 concentrations, making the SCFSkp2 ubiquitination system robust to variations in p27 abundance. When moving outside the boundaries defined for Skp2 the coordination between the complex members is lost, leading possibly to continuous activation without proper regulation. This observation seems to corroborate experimental studies on the over expression of Skp2 (and Cks1) in relation to the development cancer [2326]. Even though, the current analysis is performed on the steady-state information of this biochemical system, the underlying idea is that all the proteins of the macromolecular complex become expressed so that their mutual concentrations fall into the highly cooperative area (see Figure 2). Further analysis is off course required to verify this hypothesis. To conclude, the present result show that Shannon's information theory quantifies the cooperativity of biochemical systems, making it an important tool for the current attempts to understand cooperativity in a systems perspective.

Methods

Defining Cks1 as an asymmetric noisy channel

In information theory communication occurs through noisy channels[10], where the noise is the result of an error in the transmission. Different kinds of channels exist, but here the focus is on discrete and memoryless channels. Concretely, a noisy channel is defined by an input alphabet A X , output alphabet A Y , a set of conditional probability distributions P(y|x). The conditional probability distributions provide for every input signal (x A X ) the probability that a particular output signal (y A Y ) is produced. When the alphabets contain only two symbols and the probability of having a miscommunication is the same for both input symbols, the channel is also referred to as a binary symmetric noisy channel. Given this description, the cooperative pathway within Cks1 can now be defined as a noisy channel where the input and output alphabets both consist of the symbols 0 and 1, referring respectively to the unbound and bound states of both binding sites of Cks1. To keep things simple, the channel description of Cks1 uses only two of the three proteins that bind to Cks1 in the SCFSkp2 model system: Skp2 and phosphorylated p27. Since the symbols refer to bound and unbound state of either Skp2 or p27 to Cks1, there are four probabilities relevant here: The probability that both Skp2 and p27 are bound, that both are unbound and that Skp2 is bound (unbound) to Cks1 and p27 is unbound (bound). These probabilities are visualized in Figure 1 (right panel). Since the information transmission corresponds to one of the four output complexes in the right panel of Figure 1, the associated probabilities can be derived from the concentrations of these complexes at equilibrium. These probabilities can be organized in a transmission matrix Q:
( 1 f g f 1 g ) MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWaaeWaaeaafaqabeGacaaabaGaeGymaeJaeyOeI0IaemOzaygabaGaem4zaCgabaGaemOzaygabaGaeGymaeJaeyOeI0Iaem4zaCgaaaGaayjkaiaawMcaaaaa@36CE@
(1)
where f and g are defined as
f = P ( p 27 = 1 | S k p 2 = 0 ) = [ C k s 1 p 27 ] [ C k s 1 ] + [ C k s 1 p 27 ] g = P ( p 27 = 0 | S k p 2 = 1 ) = [ S k p 2 C k s 1 ] [ S k p 2 C k s 1 ] + [ S k p 2 C k s 1 p 27 ] MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaqbaeaabqWaaaaabaGaemOzaygabaGaeyypa0dabaGaemiuaa1aaeWaaeaacqWGWbaCcqaIYaGmcqaI3aWncqGH9aqpcqaIXaqmcqGG8baFcqWGtbWucqWGRbWAcqWGWbaCcqaIYaGmcqGH9aqpcqaIWaamaiaawIcacaGLPaaaaeaaaeaacqGH9aqpaKqbagaadaWcaaqaaiabcUfaBjabdoeadjabdUgaRjabdohaZjabigdaXiabgkHiTiabdchaWjabikdaYiabiEda3iabc2faDbqaaiabcUfaBjabdoeadjabdUgaRjabdohaZjabigdaXiabc2faDjabgUcaRiabcUfaBjabdoeadjabdUgaRjabdohaZjabigdaXiabgkHiTiabdchaWjabikdaYiabiEda3iabc2faDbaaaOqaaiabdEgaNbqaaiabg2da9aqaaiabdcfaqnaabmaabaGaemiCaaNaeGOmaiJaeG4naCJaeyypa0JaeGimaaJaeiiFaWNaem4uamLaem4AaSMaemiCaaNaeGOmaiJaeyypa0JaeGymaedacaGLOaGaayzkaaaabaaabaGaeyypa0dajuaGbaWaaSaaaeaacqGGBbWwcqWGtbWucqWGRbWAcqWGWbaCcqaIYaGmcqGHsislcqWGdbWqcqWGRbWAcqWGZbWCcqaIXaqmcqGGDbqxaeaacqGGBbWwcqWGtbWucqWGRbWAcqWGWbaCcqaIYaGmcqGHsislcqWGdbWqcqWGRbWAcqWGZbWCcqaIXaqmcqGGDbqxcqGHRaWkcqGGBbWwcqWGtbWucqWGRbWAcqWGWbaCcqaIYaGmcqGHsislcqWGdbWqcqWGRbWAcqWGZbWCcqaIXaqmcqGHsislcqWGWbaCcqaIYaGmcqaI3aWncqGGDbqxaaaaaaaa@A0DF@

This relation between concentrations implies that the errors, and later also the information exchange, depends on the concentration of the proteins that may be produced by the system.

Determining the equilibrium concentrations

We determine the concentrations of the different proteins and protein complexes using the dissociation constants determined by Seeliger et al. [19]. Using these binding affinities a system of equations is derived, which is numerically solved by determining the roots of these equations. For the simplified model system, which only incorporates Cks1, Skp2 and phosphorylated p27, this system of six equations is the following:
1. [ C k s 1 ] 0 = [ C k s 1 ] + [ C k s 1 p 27 ] + [ S k p 2 C k s 1 ] + [ S k p 2 C k s 1 p 27 ] 2. [ S k p 2 ] 0 = [ S k p 2 ] + [ S k p 2 C k s 1 ] + [ S k p 2 C k s 1 p 27 ] 3. [ p 27 ] 0 = [ p 27 ] + [ C k s 1 p 27 ] + [ S k p 2 C k s 1 p 27 ] 4. K d [ S k p 2 C k s 1 ] = [ C k s 1 ] * [ S k p 2 ] [ S k p 2 C k s 1 ] 5. K d [ C k s 1 p 27 ] = [ C k s 1 ] * [ p 27 ] [ C k s 1 p 27 ] 6. K d [ S k p 2 C k s 1 p 27 ] = [ S k p 2 C k s 1 ] * [ p 27 ] [ S k p 2 C k s 1 p 27 ] MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaqbaeaabKGaaaaaaeaacqaIXaqmcqGGUaGlaeaacqGGBbWwcqWGdbWqcqWGRbWAcqWGZbWCcqaIXaqmcqGGDbqxdaWgaaWcbaGaeGimaadabeaakiabg2da9iabcUfaBjabdoeadjabdUgaRjabdohaZjabigdaXiabc2faDjabgUcaRiabcUfaBjabdoeadjabdUgaRjabdohaZjabigdaXiabgkHiTiabdchaWjabikdaYiabiEda3iabc2faDjabgUcaRaqaaaqaaiabcUfaBjabdofatjabdUgaRjabdchaWjabikdaYiabgkHiTiabdoeadjabdUgaRjabdohaZjabigdaXiabc2faDjabgUcaRiabcUfaBjabdofatjabdUgaRjabdchaWjabikdaYiabgkHiTiabdoeadjabdUgaRjabdohaZjabigdaXiabgkHiTiabdchaWjabikdaYiabiEda3iabc2faDbqaaiabikdaYiabc6caUaqaaiabcUfaBjabdofatjabdUgaRjabdchaWjabikdaYiabc2faDnaaBaaaleaacqaIWaamaeqaaOGaeyypa0Jaei4waSLaem4uamLaem4AaSMaemiCaaNaeGOmaiJaeiyxa0Laey4kaSIaei4waSLaem4uamLaem4AaSMaemiCaaNaeGOmaiJaeyOeI0Iaem4qamKaem4AaSMaem4CamNaeGymaeJaeiyxa0Laey4kaScabaaabaGaei4waSLaem4uamLaem4AaSMaemiCaaNaeGOmaiJaeyOeI0Iaem4qamKaem4AaSMaem4CamNaeGymaeJaeyOeI0IaemiCaaNaeGOmaiJaeG4naCJaeiyxa0fabaGaeG4mamJaeiOla4cabaGaei4waSLaemiCaaNaeGOmaiJaeG4naCJaeiyxa01aaSbaaSqaaiabicdaWaqabaGccqGH9aqpcqGGBbWwcqWGWbaCcqaIYaGmcqaI3aWncqGGDbqxcqGHRaWkcqGGBbWwcqWGdbWqcqWGRbWAcqWGZbWCcqaIXaqmcqGHsislcqWGWbaCcqaIYaGmcqaI3aWncqGGDbqxcqGHRaWkaeaaaeaacqGGBbWwcqWGtbWucqWGRbWAcqWGWbaCcqaIYaGmcqGHsislcqWGdbWqcqWGRbWAcqWGZbWCcqaIXaqmcqGHsislcqWGWbaCcqaIYaGmcqaI3aWncqGGDbqxaeaacqaI0aancqGGUaGlaeaacqWGlbWscqWGKbazdaWgaaWcbaGaei4waSLaem4uamLaem4AaSMaemiCaaNaeGOmaiJaeyOeI0Iaem4qamKaem4AaSMaem4CamNaeGymaeJaeiyxa0fabeaakiabg2da9KqbaoaalaaabaGaei4waSLaem4qamKaem4AaSMaem4CamNaeGymaeJaeiyxa0LaeiOkaOIaei4waSLaem4uamLaem4AaSMaemiCaaNaeGOmaiJaeiyxa0fabaGaei4waSLaem4uamLaem4AaSMaemiCaaNaeGOmaiJaeyOeI0Iaem4qamKaem4AaSMaem4CamNaeGymaeJaeiyxa0faaaGcbaGaeGynauJaeiOla4cabaGaem4saSKaemizaq2aaSbaaSqaaiabcUfaBjabdoeadjabdUgaRjabdohaZjabigdaXiabgkHiTiabdchaWjabikdaYiabiEda3iabc2faDbqabaGccqGH9aqpjuaGdaWcaaqaaiabcUfaBjabdoeadjabdUgaRjabdohaZjabigdaXiabc2faDjabcQcaQiabcUfaBjabdchaWjabikdaYiabiEda3iabc2faDbqaaiabcUfaBjabdoeadjabdUgaRjabdohaZjabigdaXiabgkHiTiabdchaWjabikdaYiabiEda3iabc2faDbaaaOqaaiabiAda2iabc6caUaqaaiabdUealjabdsgaKnaaBaaaleaacqGGBbWwcqWGtbWucqWGRbWAcqWGWbaCcqaIYaGmcqWGdbWqcqWGRbWAcqWGZbWCcqaIXaqmcqGHsislcqWGWbaCcqaIYaGmcqaI3aWncqGGDbqxaeqaaOGaeyypa0tcfa4aaSaaaeaacqGGBbWwcqWGtbWucqWGRbWAcqWGWbaCcqaIYaGmcqGHsislcqWGdbWqcqWGRbWAcqWGZbWCcqaIXaqmcqGGDbqxcqGGQaGkcqGGBbWwcqWGWbaCcqaIYaGmcqaI3aWncqGGDbqxaeaacqGGBbWwcqWGtbWucqWGRbWAcqWGWbaCcqaIYaGmcqGHsislcqWGdbWqcqWGRbWAcqWGZbWCcqaIXaqmcqGHsislcqWGWbaCcqaIYaGmcqaI3aWncqGGDbqxaaaaaaaa@6463@
(2)

The system contains six parameters, namely [Cks1]0, [Skp2]0, [p27]0, Kd[Skp 2-Cks 1], Kd[Cks 1-p 27]and Kd[Skp 2Cks 1-p 27]. The first three parameters correspond to the total concentrations of the proteins in the model system both in isolation and in complexes. The latter three parameters are the three dissociation constants specific to the SCFSkp2 model system. Note that the dissociation constant Kd[Skp 2Cks 1-p 27]refers here to the dissociation of p27 from the complex Skp2-Cks1-p27. The results remain the same if the alternative dissociation constant, dissociating Skp2 from Skp2-Cks1-p27, is used. When the values for these parameters are inserted from ref. 19, a root finding algorithm is applied to determine the equilibrium concentrations of all the members of this system: [Cks1], [Skp2], [p27], [Skp2-Cks1], [Cks1-p27] and [Skp2-Cks1-p27]. Once these concentrations are obtained, the probabilities in the matrix Q can be determined.

Calculating mutual information

Mutual information expresses the amount of information that the output conveys about the input (and vice versa). It is formally expressed in terms of entropy:
I ( X ; Y ) = I ( Y ; X ) = H ( X ) + H ( Y ) H ( X , Y ) = H ( X ) H ( X | Y ) MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGceaabbeaacqWGjbqscqGGOaakcqWGybawcqGG7aWocqWGzbqwcqGGPaqkcqGH9aqpcqWGjbqscqGGOaakcqWGzbqwcqGG7aWocqWGybawcqGGPaqkaeaacqGH9aqpcqWGibascqGGOaakcqWGybawcqGGPaqkcqGHRaWkcqWGibascqGGOaakcqWGzbqwcqGGPaqkcqGHsislcqWGibascqGGOaakcqWGybawcqGGSaalcqWGzbqwcqGGPaqkaeaacqGH9aqpcqWGibascqGGOaakcqWGybawcqGGPaqkcqGHsislcqWGibascqGGOaakcqWGybawcqGG8baFcqWGzbqwcqGGPaqkaaaa@5766@
(3)
where the entropies are calculated as:
H ( X ) = x X p ( x ) log p ( x ) H ( X , Y ) = x , y X , Y p ( x , y ) log p ( x , y ) H ( X | Y ) = y Y p ( y ) ( x X p ( x | y ) log p ( x | y ) ) MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGceaqabeaacqWGibascqGGOaakcqWGybawcqGGPaqkcqGH9aqpcqGHsisldaaeqbqaaiabdchaWjabcIcaOiabdIha4jabcMcaPiGbcYgaSjabc+gaVjabcEgaNjabdchaWjabcIcaOiabdIha4jabcMcaPaWcbaGaemiEaGNaeyicI4SaemiwaGfabeqdcqGHris5aaGcbaGaemisaGKaeiikaGIaemiwaGLaeiilaWIaemywaKLaeiykaKIaeyypa0JaeyOeI0YaaabuaeaacqWGWbaCcqGGOaakcqWG4baEcqGGSaalcqWG5bqEcqGGPaqkcyGGSbaBcqGGVbWBcqGGNbWzcqWGWbaCcqGGOaakcqWG4baEcqGGSaalcqWG5bqEcqGGPaqkaSqaaiabdIha4jabcYcaSiabdMha5jabgIGiolabdIfayjabcYcaSiabdMfazbqab0GaeyyeIuoaaOqaaiabdIeaijabcIcaOiabdIfayjabcYha8jabdMfazjabcMcaPiabg2da9iabgkHiTmaaqafabaGaemiCaaNaeiikaGIaemyEaKNaeiykaKYaaeWaaeaadaaeqbqaaiabdchaWjabcIcaOiabdIha4jabcYha8jabdMha5jabcMcaPiGbcYgaSjabc+gaVjabcEgaNjabdchaWjabcIcaOiabdIha4jabcYha8jabdMha5jabcMcaPaWcbaGaemiEaGNaeyicI4SaemiwaGfabeqdcqGHris5aaGccaGLOaGaayzkaaaaleaacqWG5bqEcqGHiiIZcqWGzbqwaeqaniabggHiLdaaaaa@997F@

The base of the logarithm determines the units in which mutual information is expressed. Usually it is either a natural (ln x) or a binary logarithm (log2 x), making the units either nats (natural digits) or bits (binary digits). Here, a binary logarithm is used. So mutual information (see Equation 3) expresses how much we learn about the output (or input) of a channel when we receive information about the input (or output). This is calculated by subtracting the entropy (uncertainty) on the state of the output (or input) from the entropy (uncertainty) of the output (or input) when we know the input (or output). So all entropy scores are related to the state of the channel (here Cks1) and not the state of the input and output proteins, respectively Skp2 and p27.

Concretely, all entropy values can be easily derived from the probabilities related to the input and output state of Cks1. For instance, if X corresponds to Skp2, then P(Skp2 = 0) and P(Skp2 = 1) correspond to the probabilities that Skp2 is bound or not bound to Cks1. This leads to the following formulation of the entropy for Skp2:
H ( S k p 2 ) = i { 0 , 1 } P ( S k p 2 = i ) log ( P ( S k p 2 = i ) ) MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGceaqabeaacqWGibascqGGOaakcqWGtbWucqWGRbWAcqWGWbaCcqaIYaGmcqGGPaqkcqGH9aqpaeaacqGHsisldaaeqbqaaiabdcfaqjabcIcaOiabdofatjabdUgaRjabdchaWjabikdaYiabg2da9iabdMgaPjabcMcaPiGbcYgaSjabc+gaVjabcEgaNnaabmaabaGaemiuaaLaeiikaGIaem4uamLaem4AaSMaemiCaaNaeGOmaiJaeyypa0JaemyAaKMaeiykaKcacaGLOaGaayzkaaaaleaacqWGPbqAcqGHiiIZcqGG7bWEcqaIWaamcqGGSaalcqaIXaqmcqGG9bqFaeqaniabggHiLdaaaaa@5A82@
(4)
where
P ( S k p 2 = 0 ) = [ C k s 1 ] + [ C k s 1 p 27 ] Θ P ( S k p 2 = 1 ) = [ S k p 2 C k s 1 ] + [ S k p 2 C k s 1 p 27 ] Θ Θ = [ C k s 1 ] + [ S k p 2 C k s 1 ] + [ C k s 1 p 27 ] + [ S k p 2 C k s 1 p 27 ] MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGceaqabeaacqWGqbaudaqadaqaaiabdofatjabdUgaRjabdchaWjabikdaYiabg2da9iabicdaWaGaayjkaiaawMcaaiabg2da9KqbaoaalaaabaGaei4waSLaem4qamKaem4AaSMaem4CamNaeGymaeJaeiyxa0Laey4kaSIaei4waSLaem4qamKaem4AaSMaem4CamNaeGymaeJaeyOeI0IaemiCaaNaeGOmaiJaeG4naCJaeiyxa0fabaGaeuiMdefaaaGcbaGaemiuaa1aaeWaaeaacqWGtbWucqWGRbWAcqWGWbaCcqaIYaGmcqGH9aqpcqaIXaqmaiaawIcacaGLPaaacqGH9aqpjuaGdaWcaaqaaiabcUfaBjabdofatjabdUgaRjabdchaWjabikdaYiabgkHiTiabdoeadjabdUgaRjabdohaZjabigdaXiabc2faDjabgUcaRiabcUfaBjabdofatjabdUgaRjabdchaWjabikdaYiabgkHiTiabdoeadjabdUgaRjabdohaZjabigdaXiabgkHiTiabdchaWjabikdaYiabiEda3iabc2faDbqaaiabfI5arbaaaOqaaiabfI5arjabg2da9iabcUfaBjabdoeadjabdUgaRjabdohaZjabigdaXiabc2faDjabgUcaRiabcUfaBjabdofatjabdUgaRjabdchaWjabikdaYiabgkHiTiabdoeadjabdUgaRjabdohaZjabigdaXiabc2faDjabgUcaRiabcUfaBjabdoeadjabdUgaRjabdohaZjabigdaXiabgkHiTiabdchaWjabikdaYiabiEda3iabc2faDbqaaiabgUcaRiabcUfaBjabdofatjabdUgaRjabdchaWjabikdaYiabgkHiTiabdoeadjabdUgaRjabdohaZjabigdaXiabgkHiTiabdchaWjabikdaYiabiEda3iabc2faDbaaaa@AF4D@
The entropy H(p27) is derived in the same way. The joint entropy is
H ( S k p 2 , p 27 ) = i , j { 0 , 1 } P ( S k p 2 = i , p 27 = j ) log ( P ( S k p 2 = i , p 27 = j ) ) MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGceaqabeaacqWGibascqGGOaakcqWGtbWucqWGRbWAcqWGWbaCcqaIYaGmcqGGSaalcqWGWbaCcqaIYaGmcqaI3aWncqGGPaqkcqGH9aqpaeaacqGHsisldaaeqbqaaiabdcfaqjabcIcaOiabdofatjabdUgaRjabdchaWjabikdaYiabg2da9iabdMgaPjabcYcaSiabdchaWjabikdaYiabiEda3iabg2da9iabdQgaQjabcMcaPiGbcYgaSjabc+gaVjabcEgaNnaabmaabaGaemiuaaLaeiikaGIaem4uamLaem4AaSMaemiCaaNaeGOmaiJaeyypa0JaemyAaKMaeiilaWIaemiCaaNaeGOmaiJaeG4naCJaeyypa0JaemOAaOMaeiykaKcacaGLOaGaayzkaaaaleaacqWGPbqAcqGGSaalcqWGQbGAcqGHiiIZcqGG7bWEcqaIWaamcqGGSaalcqaIXaqmcqGG9bqFaeqaniabggHiLdaaaaa@6E2A@
(5)
where
P ( S k p 2 = 0 , p 27 = 0 ) = [ C k s 1 ] Θ P ( S k p 2 = 1 , p 27 = 0 ) = [ S k p 2 C k s 1 ] Θ P ( S k p 2 = 0 , p 27 = 1 ) = [ C k s 1 p 27 ] Θ P ( S k p 2 = 1 , p 27 = 1 ) = [ S k p 2 C k s 1 p 27 ] Θ MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGceaqabeaacqWGqbaudaqadaqaaiabdofatjabdUgaRjabdchaWjabikdaYiabg2da9iabicdaWiabcYcaSiabdchaWjabikdaYiabiEda3iabg2da9iabicdaWaGaayjkaiaawMcaaiabg2da9KqbaoaalaaabaGaei4waSLaem4qamKaem4AaSMaem4CamNaeGymaeJaeiyxa0fabaGaeuiMdefaaaGcbaGaemiuaa1aaeWaaeaacqWGtbWucqWGRbWAcqWGWbaCcqaIYaGmcqGH9aqpcqaIXaqmcqGGSaalcqWGWbaCcqaIYaGmcqaI3aWncqGH9aqpcqaIWaamaiaawIcacaGLPaaacqGH9aqpjuaGdaWcaaqaaiabcUfaBjabdofatjabdUgaRjabdchaWjabikdaYiabgkHiTiabdoeadjabdUgaRjabdohaZjabigdaXiabc2faDbqaaiabfI5arbaaaOqaaiabdcfaqnaabmaabaGaem4uamLaem4AaSMaemiCaaNaeGOmaiJaeyypa0JaeGimaaJaeiilaWIaemiCaaNaeGOmaiJaeG4naCJaeyypa0JaeGymaedacaGLOaGaayzkaaGaeyypa0tcfa4aaSaaaeaacqGGBbWwcqWGdbWqcqWGRbWAcqWGZbWCcqaIXaqmcqGHsislcqWGWbaCcqaIYaGmcqaI3aWncqGGDbqxaeaacqqHyoquaaaakeaacqWGqbaudaqadaqaaiabdofatjabdUgaRjabdchaWjabikdaYiabg2da9iabigdaXiabcYcaSiabdchaWjabikdaYiabiEda3iabg2da9iabigdaXaGaayjkaiaawMcaaiabg2da9KqbaoaalaaabaGaei4waSLaem4uamLaem4AaSMaemiCaaNaeGOmaiJaeyOeI0Iaem4qamKaem4AaSMaem4CamNaeGymaeJaeyOeI0IaemiCaaNaeGOmaiJaeG4naCJaeiyxa0fabaGaeuiMdefaaaaaaa@A90E@

Multivariate mutual information

To derive the information exchange between three or more proteins a multivariate approach needs to be followed[11, 12]. This approach allows the analysis of the signal between two input proteins and an output protein. As in the previous formulation, the mutual information is determined using entropy:
I ( X ; Y , Z ) = I ( Y , Z ; X ) = H ( X ) + H ( Y , Z ) H ( X , Y , Z ) MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGceaabbeaacqWGjbqscqGGOaakcqWGybawcqGG7aWocqWGzbqwcqGGSaalcqWGAbGwcqGGPaqkcqGH9aqpcqWGjbqscqGGOaakcqWGzbqwcqGGSaalcqWGAbGwcqGG7aWocqWGybawcqGGPaqkaeaacqGH9aqpcqWGibascqGGOaakcqWGybawcqGGPaqkcqGHRaWkcqWGibascqGGOaakcqWGzbqwcqGGSaalcqWGAbGwcqGGPaqkcqGHsislcqWGibascqGGOaakcqWGybawcqGGSaalcqWGzbqwcqGGSaalcqWGAbGwcqGGPaqkaaaa@5323@
(6)

where X represents the output signal and Y and Z represent two input signals or visa versa. In addition, the effect of either one of the components on the two other ones can be analysed by eliminating this component. For instance if one wants to determine the effect of Cdk2 on the communication between Skp2 and phosphorylated p27, the mutual information I(Skp2;p27) and the averaged transmitted information ICdk 2(Skp2;p27) need to be determined (see ref. 9 for the details). If I(Skp2;p27) is not equal to ICdk 2(Skp2;p27) then Cdk2 has an effect on the transmission between the two other proteins. This difference, called the interaction information A(Skp2, Cdk2, p27), is the gain (or loss) in the sample information transmitted between any two of the proteins, caused by the additional knowledge of the third one. Combining the interaction information with I(Skp2;p27) and I(Skp2;Cdk2) produces again the multivariate mutual information I(Skp2; Cdk2, p27). As a consequence, A(Skp2, Cdk2, p27) expresses how the two signals are modulated, which can be either in a negative of positive way.

Declarations

Acknowledgements

Discussions with Dr. Sebastian Maurer-Stroh are gratefully acknowledged. We also acknowledge valuable input from Prof. James R. Williamson and Dr. Mark Isalan to improve the clarity of the manuscript.

Authors’ Affiliations

(1)
SWITCH, VIB
(2)
Vrije Universiteit Brussel
(3)
Ørsted.DTU, Technical Uiversity of Denmark, Kgs

References

  1. Spirin V, Mirny LA: Protein complexes and functional modules in molecular networks. Proc Natl Acad Sci USA. 2003, 100: 12123-12128.PubMed CentralView ArticlePubMedGoogle Scholar
  2. Rousseau F, Schymkowitz J: A systems biology perspective on protein structural dynamics and signal transduction. Curr Opin Struct Biol. 2005, 15: 23-30.View ArticlePubMedGoogle Scholar
  3. Senear DF, Ross JB, Laue TM: Analysis of protein and DNA-mediated contributions to cooperative assembly of protein-DNA complexes. Methods. 1998, 16: 3-20.View ArticlePubMedGoogle Scholar
  4. Luque I, Leavitt SA, Freire E: The linkage between protein folding and functional cooperativity: two sides of the same coin?. Annu Rev Biophys Biomol Struct. 2002, 31: 235-256.View ArticlePubMedGoogle Scholar
  5. Goh CS, Milburn D, Gerstein M: Conformational changes associated with protein-protein interactions. Curr Opin Struct Biol. 2004, 14: 104-109.View ArticlePubMedGoogle Scholar
  6. Williamson JR: Cooperativity in macromolecular assembly. Nat Chem Biol. 2008, 4: 458-465.View ArticlePubMedGoogle Scholar
  7. Horovitz A, Fersht AR: Strategy for analysing the co-operativity of intramolecular interactions in peptides and proteins. J Mol Biol. 1990, 214: 613-617.View ArticlePubMedGoogle Scholar
  8. Dahlquist FW: The meaning of Scatchard and Hill plots. Methods Enzymol. 1978, 48: 270-299.View ArticlePubMedGoogle Scholar
  9. Imai K: Hemoglobin Chesapeake (92 alpha, arginine – leucine). Precise measurements and analyses of oxygen equilibrium. J Biol Chem. 1974, 249: 7607-7612.PubMedGoogle Scholar
  10. Shannon CE: A mathematical theory of communications. Bell Syst Tech J. 1948, 27: 379-423.View ArticleGoogle Scholar
  11. Fano RM: Transmission of Information: A statistical theory of communications. 1961, MIT Press & John Wiley and Sonds, IncGoogle Scholar
  12. McGill WJ: Multivariate mutual information. Psychometrika. 1954, 19: 97-116.View ArticleGoogle Scholar
  13. Bartek J, Lukas J: p27 destruction: Cks1 pulls the trigger. Nat Cell Biol. 2001, 3: E95-98.View ArticlePubMedGoogle Scholar
  14. Ganoth D, Bornstein G, Ko TK, Larsen B, Tyers M, Pagano M, Hershko A: The cell-cycle regulatory protein Cks1 is required for SCF(Skp2)-mediated ubiquitinylation of p27. Nat Cell Biol. 2001, 3: 321-324.View ArticlePubMedGoogle Scholar
  15. Harper JW: Protein destruction: adapting roles for Cks proteins. Curr Biol. 2001, 11: R431-435.View ArticlePubMedGoogle Scholar
  16. Spruck C, Strohmaier H, Watson M, Smith AP, Ryan A, Krek TW, Reed SI: A CDK-independent function of mammalian Cks1: targeting of SCF(Skp2) to the CDK inhibitor p27Kip1. Mol Cell. 2001, 7: 639-650.View ArticlePubMedGoogle Scholar
  17. Slingerland J, Pagano M: Regulation of the cdk inhibitor p27 and its deregulation in cancer. J Cell Physiol. 2000, 183: 10-17.View ArticlePubMedGoogle Scholar
  18. Pagano M, Benmaamar R: When protein destruction runs amok, malignancy is on the loose. Cancer Cell. 2003, 4: 251-256.View ArticlePubMedGoogle Scholar
  19. Seeliger MA, Breward SE, Friedler A, Schon O, Itzhaki LS: Cooperative organization in a macromolecular complex. Nature Structural Biology. 2003, 10: 718-724.View ArticlePubMedGoogle Scholar
  20. Lenaerts T, Ferkinghoff-Borg J, Stricher F, Serrano L, Schymkowitz JW, Rousseau F: Quantifying information transfer by protein domains: analysis of the Fyn SH2 domain structure. BMC Struct Biol. 2008, 8: 43-PubMed CentralView ArticlePubMedGoogle Scholar
  21. Sitry D, Seeliger MA, Ko TK, Ganoth D, Breward SE, Itzhaki LS, Pagano M, Hershko A: Three different binding sites of Cks1 are required for p27-ubiquitin ligation. J Biol Chem. 2002, 277: 42233-42240.View ArticlePubMedGoogle Scholar
  22. MacKay DJC: Information Theory, Inference and Learning Algorithms. 2002, Cambridge University Press, CambridgeGoogle Scholar
  23. Gstaiger M, Jordan R, Lim M, Catzavelos C, Mestan J, Slingerland J, Krek W: Skp2 is oncogenic and overexpressed in human cancers. Proc Natl Acad Sci USA. 2001, 98: 5043-5048.PubMed CentralView ArticlePubMedGoogle Scholar
  24. Slotky M, Shapira M, Ben-Izhak O, Linn S, Futerman B, Tsalic M, Hershko DD: The expression of the ubiquitin ligase subunit Cks1 in human breast cancer. Breast Cancer Res. 2005, 7: R737-744.PubMed CentralView ArticlePubMedGoogle Scholar
  25. Kitajima S, Kudo Y, Ogawa I, Bashir T, Kitagawa M, Miyauchi M, Pagano M, Takata T: Role of Cks1 overexpression in oral squamous cell carcinomas: cooperation with Skp2 in promoting p27 degradation. Am J Pathol. 2004, 165: 2147-2155.PubMed CentralView ArticlePubMedGoogle Scholar
  26. Huang HY, Kang HY, Li CF, Eng HL, Chou SC, Lin CN, Hsiung CY: Skp2 overexpression is highly representative of intrinsic biological aggressiveness and independently associated with poor prognosis in primary localized myxofibrosarcomas. Clin Cancer Res. 2006, 12: 487-498.View ArticlePubMedGoogle Scholar

Copyright

© Lenaerts et al; licensee BioMed Central Ltd. 2009

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.