Académique Documents
Professionnel Documents
Culture Documents
2013 Changes
Version 0.02: Minor changes to the section on op amps to clarify some explanations. Incorrect figure referenced in Assignment 2. Incorrect caption in inverting
op amp figure.
Version 0.03: Added new figure to on variables to Chapter 1, added chapters on
Intro to Feedback, Inputs to a System, Linear Systems and Response of Linear
Systems
Version 0.04: Added exercises to chapter on intro to feedback
Version 0.05: Corrected some typos in signals chapter
Version 0.06: Rearranged chapters, filled in Chapter 2, added chapters on steady
state, response to linear systems and Laplace transforms. Fixed error in section on
impulses identified by RL.
Version 0.07: Updated steady state chapter and linear response chapter
Version 0.08: Clarified and fixed some typos in Chapter ??: Response of Linear
Systems.
Version 0.09: Fixed typo in Jacobian (A) indexing
Version 0.10: Clarified section on time invariance in Chapter 7.
Version 0.11: Split Chapter ?? in to two separate chapter.
Version 0.12: Math typo in Chapter 7 identified by TDM. Added section to Chapter
4 on elasticities
Version 0.13: Added more examples to Chapter 9 and fixed a math error in Chapter
8, page 200, identified by KT.
Version 0.14: Modified section on Time Shifting in Chapter 9.
Version 0.15: Fixed three typos on page 206 (k1 denominator should be k2), identified by DL. Update Chapter 11
Version 0.16: Fixed inconsistent use of h in Q10 of Chapter 4, identified by DL.
Updated text in Chap 11 and 12
Version 0.17: Fixed another typo in Q10 of Chapter 4, identified by AT
Version 0.18: Updated Chapter 10 with a Block diagram example. Updated Chapter
12 and 13.
ii
Version 0.19: Significant updates to Chapters 14, 15, 16, 17 and rearranged Appendices.
Version 0.20: Fixed typos in Chapter 14 and 15, identified by DL.
Version 0.21: Additions to the feedback chapter 15.
2014 Changes
Version 0.25: New version for 2014 Class (January 5, 2014)
Version 0.26: Fixed typos, updated text and exercises in Chap 1 (January 9, 2014)
Version 0.27: Added text on low impedance (January 9, 2014)
Version 0.28: Fixed Figure 1.13 and 1.14 plus typos in exercises for Chap 1 (January 13, 2014)
Version 0.29: Fixed typos in Chapter 3 (January 14, 2014)
Version 0.31: Updated exercises in Chapter 4 (January 16, 2014)
Version 0.32: Fixed error in dS3=dt, example 4.8 thanks to KM (January 19, 2014)
Version 0.33: Updated Chapter 5 (January 20, 2014)
Version 0.34: Updated Exercises in Chapter 5 (January 21, 2014)
Version 0.36: Updated Chapter 6 (January 23, 2014)
Version 0.37: Updated Chapter 10, fixed typo in Chapter 6, thanks to AT (January
23, 2014)
Version 0.38: Further updates to Chapter 10 (February 2, 2014)
Version 0.39: Fix to electrical model state space representation in Chapter 8, thanks
to CC (February 3, 2014)
Version 0.40: Fix to some typos and updates to Chapter 11 and 12 (February 4,
2014)
Version 0.41: Updated Chapter 16 (February 5, 2014)
Version 0.42: Clarified section on characteristic equation in Chapter 11, thanks to
JW (February 9, 2014)
Version 0.43: Updated various typos thanks to AT (February 11, 2014)
Version 0.44 Minor correction to eqn 12.3 and the next equation in Chapter 12
(February 11, 2014)
Version 0.45 Updates to Chapter 12, 13, 14, and 17 (February 25, 2014)
Version 0.46 Updates to Chapter 11, 12 and 17 (February 26, 2014)
Version 0.47 Correction to equation 6.2 (December 8, 2014)
Version 0.50 Major update across all sections (December 31, 2014)
iii
Herbert M. Sauro
University of Washington
Seattle, WA
Ambrosius Publishing
Contents
Introduction to Modeling
1.1
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . .
1.2
1.3
1.4
1.5
Classification of Models . . . . . . . . . . . . . . . . . . . . .
15
1.6
Electrical Models . . . . . . . . . . . . . . . . . . . . . . . . .
17
1.7
Block Diagrams . . . . . . . . . . . . . . . . . . . . . . . . . .
25
1.8
27
1.9
29
1.10
Approximations . . . . . . . . . . . . . . . . . . . . . . . . . .
31
1.11
Behavior . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
32
Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
33
Simulation Software
45
2.1
45
2.2
49
2.3
50
Further Reading . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
50
51
3.1
51
3.2
Historical Perspective . . . . . . . . . . . . . . . . . . . . . . .
52
3.3
57
vi
CONTENTS
3.4
61
Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
65
Biological Networks
67
4.1
Gene Regulation . . . . . . . . . . . . . . . . . . . . . . . . . .
67
4.2
Metabolic Pathways . . . . . . . . . . . . . . . . . . . . . . . .
69
4.3
Protein Networks . . . . . . . . . . . . . . . . . . . . . . . . .
70
4.4
Stoichiometric Networks . . . . . . . . . . . . . . . . . . . . .
72
4.5
Reaction Kinetics . . . . . . . . . . . . . . . . . . . . . . . . .
77
4.6
Elasticity Coefficients . . . . . . . . . . . . . . . . . . . . . . .
81
4.7
87
4.8
Mass-Balance Equations . . . . . . . . . . . . . . . . . . . . .
90
4.9
Stoichiometry Matrix . . . . . . . . . . . . . . . . . . . . . . .
95
4.10
99
4.11
100
Further Reading . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
102
Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
103
111
5.1
System States . . . . . . . . . . . . . . . . . . . . . . . . . . .
111
5.2
Equilibrium . . . . . . . . . . . . . . . . . . . . . . . . . . . .
111
5.3
Transients . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
114
5.4
Steady State . . . . . . . . . . . . . . . . . . . . . . . . . . . .
114
5.5
120
5.6
128
5.7
131
5.8
Introduction to Stability . . . . . . . . . . . . . . . . . . . . . .
131
5.9
Sensitivity Analysis . . . . . . . . . . . . . . . . . . . . . . . .
134
Further Reading . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
136
Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
136
Tellurium Scripts . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
138
CONTENTS
vii
Inputs to a System
143
6.1
144
6.2
145
6.3
Addition of Signals . . . . . . . . . . . . . . . . . . . . . . . .
148
6.4
Ramp Signal . . . . . . . . . . . . . . . . . . . . . . . . . . . .
150
6.5
Impulse Signal . . . . . . . . . . . . . . . . . . . . . . . . . .
152
6.6
Sinusoidal . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
156
6.7
Exponential . . . . . . . . . . . . . . . . . . . . . . . . . . . .
160
Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
164
6.8
165
Appendix . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Linear Systems
167
7.1
Linear Systems . . . . . . . . . . . . . . . . . . . . . . . . . .
168
7.2
Time-Invariance . . . . . . . . . . . . . . . . . . . . . . . . . .
171
7.3
Linearization . . . . . . . . . . . . . . . . . . . . . . . . . . .
173
Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
181
183
8.1
183
8.2
188
Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
194
197
9.1
197
9.2
General Solutions . . . . . . . . . . . . . . . . . . . . . . . . .
201
Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
206
10 Laplace Transforms
209
10.1
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . .
209
10.2
Laplace Transform . . . . . . . . . . . . . . . . . . . . . . . .
210
10.3
Inverse Transforms . . . . . . . . . . . . . . . . . . . . . . . .
216
10.4
220
10.5
223
viii
CONTENTS
10.6
Additional Examples . . . . . . . . . . . . . . . . . . . . . . .
224
10.7
Block Diagrams . . . . . . . . . . . . . . . . . . . . . . . . . .
226
Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
228
11 Zero-State Response
231
11.1
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . .
231
11.2
231
11.3
Convolution . . . . . . . . . . . . . . . . . . . . . . . . . . . .
233
11.4
State Equation . . . . . . . . . . . . . . . . . . . . . . . . . . .
237
11.5
Transfer Function . . . . . . . . . . . . . . . . . . . . . . . . .
243
11.6
245
11.7
First-Order Systems . . . . . . . . . . . . . . . . . . . . . . . .
247
Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
248
12 Cellular Networks
251
12.1
Systems Equation . . . . . . . . . . . . . . . . . . . . . . . . .
251
12.2
Output Equations . . . . . . . . . . . . . . . . . . . . . . . . .
258
12.3
261
13 Frequency Response
263
13.1
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . .
263
13.2
. . . . . . . . . . . . . . . . .
264
13.3
Frequency Response . . . . . . . . . . . . . . . . . . . . . . .
265
13.4
Sinusoidal Input . . . . . . . . . . . . . . . . . . . . . . . . . .
267
13.5
268
13.6
Bode Plots . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
270
13.7
276
Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
282
283
14.1
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . .
283
14.2
284
ix
CONTENTS
15 Structural Constraints
289
15.1
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . .
289
15.2
Structural Constraints . . . . . . . . . . . . . . . . . . . . . . .
290
16 First-Order Systems
295
16.1
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . .
295
16.2
Free Response . . . . . . . . . . . . . . . . . . . . . . . . . . .
297
16.3
Impulse Response . . . . . . . . . . . . . . . . . . . . . . . . .
298
16.4
Unit Response . . . . . . . . . . . . . . . . . . . . . . . . . . .
298
16.5
Ramp Response . . . . . . . . . . . . . . . . . . . . . . . . . .
300
16.6
Frequency Response . . . . . . . . . . . . . . . . . . . . . . .
300
Further Reading . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
302
Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
302
17 Stability
303
17.1
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . .
303
17.2
Internal Stability
. . . . . . . . . . . . . . . . . . . . . . . . .
303
17.3
External Stability . . . . . . . . . . . . . . . . . . . . . . . . .
308
17.4
Phase Plots . . . . . . . . . . . . . . . . . . . . . . . . . . . .
308
17.5
Bifurcation Plots . . . . . . . . . . . . . . . . . . . . . . . . .
315
17.6
Computing Stability . . . . . . . . . . . . . . . . . . . . . . . .
330
Further Reading . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
332
Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
333
18 Negative Feedback
337
18.1
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . .
337
18.2
Types of Feedback . . . . . . . . . . . . . . . . . . . . . . . .
338
18.3
Transient Response . . . . . . . . . . . . . . . . . . . . . . . .
340
18.4
Stability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
342
18.5
Accuracy . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
351
18.6
Linearization . . . . . . . . . . . . . . . . . . . . . . . . . . .
352
18.7
Sensitivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
354
CONTENTS
18.8
355
18.9
356
Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
358
19 Positive Feedback
19.1
359
Basic Properties . . . . . . . . . . . . . . . . . . . . . . . . . .
20 Oscillators
20.1
359
365
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . .
365
365
20.2
366
20.3
Relaxation Oscillators . . . . . . . . . . . . . . . . . . . . . . .
369
20.4
Oscillator Classification . . . . . . . . . . . . . . . . . . . . . .
372
Appendix A Symbols
A.1
377
List Of Symbols . . . . . . . . . . . . . . . . . . . . . . . . . .
377
379
B.1
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . .
379
B.2
Numerical Solutions
. . . . . . . . . . . . . . . . . . . . . . .
381
382
386
B.2.3 Runge-Kutta . . . . . . . . . . . . . . . . . . . . . . .
388
389
391
B.3
Matlab Solvers . . . . . . . . . . . . . . . . . . . . . . . . . .
392
B.4
Other Software . . . . . . . . . . . . . . . . . . . . . . . . . .
394
Further Reading . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
394
Exercises . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
395
397
C.1
Michaelis-Menten Kinetics . . . . . . . . . . . . . . . . . . . .
397
C.2
398
CONTENTS
xi
C.3
399
C.4
Haldane Relationship . . . . . . . . . . . . . . . . . . . . . . .
399
C.5
Competitive Inhibition . . . . . . . . . . . . . . . . . . . . . .
400
C.6
Cooperativity . . . . . . . . . . . . . . . . . . . . . . . . . . .
401
C.7
Allostery . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
404
Further Reading . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
405
407
D.1
407
D.2
NullClines . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
407
D.3
Bifurcation Plots . . . . . . . . . . . . . . . . . . . . . . . . .
408
411
E.1
Polynomials . . . . . . . . . . . . . . . . . . . . . . . . . . . .
411
E.2
Absolute Values . . . . . . . . . . . . . . . . . . . . . . . . . .
411
E.3
Common Integrals . . . . . . . . . . . . . . . . . . . . . . . . .
412
413
413
419
421
References
425
History
431
Index
433
xii
CONTENTS
Introduction to Modeling
You take the blue pill, the story ends, you wake up in your bed and
believe whatever you want to believe. You take the red pill, you stay in
Wonderland, and I show you how deep the rabbit hole goes.
The Matrix, 1999
1.1 Introduction
In todays society we are surrounded by machines of all kinds, including rockets,
aeroplanes, tablets, the internet and the list goes on. We generally take for granted
that our machines are on the whole remarkably reliable and predicable. Our mobile
phone always connects and space probes always reach their destination. The same
can be said for biological systems; one average we breath 672,768,000 times in a
lifetime, without fail; we can stand, walk and run usually without any incident. Is
it a striking that a complex system such as a human, can remain fully functional
for almost 100 years. Part of the reason for this lies in the sophisticated systems
that exist to regulate, control and repair our biological cells. The reliability of our
machines can also be traced to engineering sophisticated control mechanisms to
ensure that our machines behave in a predicable way. We can understand manmade and natural systems achieve these feats using the science of control. Control
is concerned with understanding how systems behave predicably, that is how future
behavior unfolds as a result of regulatory mechanisms. One of the simplest examples of a regulated system is the temperature thermostat. A thermostat is able to
regulate a heat source such that the temperature at a particular location, for example
a room, is held at a steady level. The temperature in the room is therefore highly
predicable.
Engineers have for many years had a well developed set of modeling and theoretical
tools to help them understand and design predicable systems. This is particularly
so in the electrical engineering community where computer design tools are widely
used in developing highly complex circuits. In more recent years control theory
has seen applied to understanding control systems found in living organisms, for
example approaches to help us understand the effect of genetic or environmental
perturbations on a given phenotype. There are many books on control theory that
deal with mechanical and electrical systems, far fewer on biological systems. In
this book we will discuss the application of control theory to understanding living
systems.
Students who are well versed in modeling and building computational models may
omit the first chapter. We will start by considering the building and simulation of
physical models of systems.
so that:
dh
Q1 Q2
D
dt
A
Using Torriellis Law we know that the rate of water flowing out of a given tank, i ,
is equal to:
p
Qi D Ki h
Where Ki is a constant related to the resistance of the output pipe. Therefore for
the first tank we have:
p
Q1 K h
dh
D
dt
A
Q1
h1
Q4
K1
Q2
h2
K2
Q3
A simplified or idealized description or conception of a particular system, situation, or process, often in mathematical terms, that is put forward as a basis for theoretical or empirical understanding, or for calculations, predictions, etc.
This definition embodies a number of critical features that defines a model, the most
important is that a model represents an idealized description, a simplification, of
a real world process. This may at first appear to be a weakness but simplification
is usually done on purpose. Simplification is important because it allows us to
comprehend the essential features of a complex process without being burdened
and overwhelmed by unnecessary detail. In the water tank model we assumed for
example that as the water flowed from each tank, any resulting water turbulence
had only a negligible effect on the system dynamics.
Models come in various forms including verbal, written text, visual, mathematical
and others. Biology, particular molecular biology, has a long tradition of using visual models to represent cellular structure and function; one need only look through
a modern textbook to see instances of visual models on every page. Visual models
have been immensely useful at describing complicated biological processes but are
limited in their scope. A more interesting way to describe models is to use mathematics, a language designed for logical reasoning. Mathematical models are useful
in biology for a number of reasons, but the three most important are increased precision, prediction, and the capacity for analysis. Analysis can be carried out either
by simulation or by analytic means. Although visual models can be used to make
predictions, the kinds of predictions that can be made are limited. The use of mathematical models opens up whole new vistas of study which visual models simply
can not match.
One word that has been used but has been undefined is the word System. Here
we define what we mean by the system:
System: The part of the Universe we are interested in studying with a defined
boundary and set of processes within.
Isolated System
Closed System
Energy
Open System
Energy
Mass
Class Question 1
For each of the following systems, decide whether the system is isolated, closed or
open. Comment on the nature of the surroundings.
i) A system represented by a mechanical clock slowly winds down in a room controlled by a thermostat.
ii) A car engine running idle in the open air.
iii) A bacterial culture is grown in batch and kept in a sealed and insulated chamber.
The theory of common descent is falsifiable. If an entire population of fossilized rabbits were
found in a precambrian formation, the theory would be in serious trouble.
sent everything in a given problem, a model must be selective and represent those
things most relevant to the task at hand.
Electrical Circuits
Purely Mathematical
Algebraic Equations
Linear differential equations
Linear difference equations
Model Variables
One of the first decisions to make when developing a model is to identify the external (the causes) and internal variables (effects).
The internal variables of the system are often termed the state variables or dynamic variables and will change in time as the system evolves. The values of the
state variables are determined by the model or system. In an experiment involving
living cells the state variables will often be concentrations of molecules and ions
and voltages across membranes. In the control theory literature, state variables are
often denoted by the symbol x.
The state of a system at time t is described by a set of state variables:
x.t/
They are the smallest set of variables that define the state of the system.
Initial Conditions
An important part in defining a model is to assigned initial values to all the state
variables. In the dynamics literature the initial values are called the initial conditions. For a state variable x.t/, a common notation to indicate an initial condition
is:
x.0/ D 1:2345
That is at time zero before a simulation starts, we set the value for the state variables
at zero time.
Output Variables
A second closely related set of variables are the output variables. These are sometimes equivalent to the state variables so that measurement of an output from a
system is a direct measure of a particular state variable. However, some state variable may not be easily measurable (this is particularly the case in cell biology) or
the measured output is a function of one or more state variables. This means that the
number of output variables will be less than the number of state variables. If a state
variable is measurable via an output variable we will often try to make sure that the
output variable is proportional to the state variable. For example, the movement on
the needle on an old style voltage meter isnt a direct measure of the voltage but the
location of the needle is proportional to the voltage. We thus use the movement of
the needle as a proxy for the voltage. In the control literature the output variables
are often denoted by the symbol y.
The output of a system at time t is described by a set of output variables:
y.t /
If possible we try to make sure that an output variable is directly proportional to one of the state variables.
In contrast to the state variables, the external variables are fixed and in principle
are under the complete control of the experimenter. Often the external variables are
clamped to some fixed values (cf. voltage clamp). The external variables are also
called boundary variables because they are considered to be at the boundary of
the system and the rest of the universe. In an experiment involving living cells the
10
external variables will often be concentrations of nutrients in the medium, the level
of oxygen or the concentrations of externally applied drugs or growth factors.
Model Inputs
The inputs to a biological system will often be drug infusions, levels of nutrients,
voltages or depending on the system, gene expression levels. Inputs are generally under the control of the experimenter and are often clamped by some kind
of buffering mechanism. The buffering mechanism can simply be a large external
reservoir so that any exchange of mass between the system and the external reservoir has a negligible effect on the external concentration. Alternatively there may
be active mechanisms maintaining an external concentration. A classic example of
active maintenance of an external variable is the voltage clamp used in electrophysiology. Finally, external concentrations may simply be slow moving (for example
gene expression) compared to the timescale of the model. Over the study period,
the external concentrations change very little. A typical example of the latter is the
study of a metabolic response over a timescale that is shorter than gene expression.
This permits a modeler to study a metabolic pathway without considering the effect of changes in gene expression. Inputs that are concentrations are often called
boundary variables in systems and synthetic biology.
The inputs to a system at time t is described by a set of input variables:
u.t/
Sometimes inputs are also represented using the symbol p.t/.
The separation between the state and boundary variables depends ultimately on
practical considerations and the particular questions that the experimenter and modeler wish to answer. However, once the choice is made, the separation is strictly
adhered to during the course of a study. This means for example that the environment surrounding the physical system will, by definition, be unaffected by the
behavior of the system. If for some reason parts of the environment do change as
a result of changes inside the system then these parts should be considered part of
the system. In the control literature, the inputs are usually denoted by the symbol u
but sometimes also using the symbol p.
11
Model Parameters
As described previously, the state variables are quantities that evolve in time and
the boundary variables are concentrations or voltages that are clamped by the observer or under the control of a time dependent function. There is a third set of
quantities which are often called the system parameters. These include the various kinetic constants and enzyme activity factors whose values are determined by
a combination of thermodynamic, physical or genotypic properties. Some system
parameters will appear in the reaction rate laws as kinetic constants, others will be
related to physical characteristics such as the volume of the system (assuming the
volume is constant during the study period). Although many parameters can be
changed by the experimenter, in any particular study a set of parameters are designated fixed during the experiment. The fixed quantities are called parameters. One
can imagine for example changing kinetic constants via site directed mutagenesis
or changing enzyme activities by altering promoter efficiencies. The actual choice
of system parameters in any particular model will depend on physical limitations
and the question being considered by the researcher. If a particular system parameter can be altered by the experimenter then it should be considered an input to the
system.
The unchanging properties of a system are described by a set of parameters:
p
Forcing Functions
As described earlier, it is common to make sure that the surroundings do not change
during the duration of the study. For example we might make sure that the pH remains constant by using a buffer solution. The key point is that the experimenter
has complete control over the experiment. In some cases it is useful for an experimenter to change the surrounding conditions in a controlled fashion. For example,
he/she might slowly increase the concentration of an administered drug or make a
step change in a variable such as the concentration of an enzyme. In systems theory
such controlled changes are often called forcing functions and are denoted by the
symbol u.t / to indicate that the inputs can be functions of time.
12
Quantities
Absolute
constants
Quantiaties
which can vary
Space and
Time
Variables which
vary in space
and time
Independent
Variables
Not Allowed
to Vary
Parameters p
Inputs u(t)
Outputs y(t)
Taken together if we choose to describe our system using a set of ordinary differential equations, then we can write the system equation in the following way:
u1(t)
x1, x2 , .... xn
u2(t)
State Variables
p 1, p 2 ,.... p q
ua(t)
Parameters
System Inputs
13
y 1(t)
y 2(t)
ym(t)
Observed Variables
dx
D f .x.t/; u.t//
dt
y D g.x.t /; u.t //
Class Question 2
For the water tank model described in section 1.2, identify the state variables, outputs, inputs and parameters of the system.
Variables: Heights, h1 and h2 in tank one and two.
14
15
We can classify a model as a combination of the above attributes. The water tank
model uses a deterministic, continuous approach. The lion model might use a discrete and stochastic approach. In this book we will be primarily concerned with
deterministic and continuous models. Table 1.2 shows the four combinations and
examples where each combination might be appropriately used.
Type
Example
Continuous/Deterministic
Continuous/Stochastic
Discrete/Deterministic
Discrete/Stochastic
Projectile motion
Brownian motion
Large population dynamics
Small population dynamics
A game of chess
Two teams playing American Football
A forest fire
Gene regulation in a bacterium
The beating of a healthy heart
A tornado
Growth of a tumor
Upper atmosphere chemistry
A digital computer
Electron flow through a single transistor
A population of cells from which emerges a cancerous cell
Metabolism
The swimming of a single E. coli through water
16
1.
2.
3.
4.
Linear or Non-Linear
Dynamic or Static
Time invariant or time dependent
Lumped or distributed parameter models
17
v1
v2
R1
i1
i2
R1
R2
i3
vo ?
18
i = Current R = Resistance
q = Charge
Ohms Law: v D Ri
1
q
C
In addition, the charge on a capacitor is given by:
Z
q D idt
Capacitor Law: q D C v or v D
dv
dt
Resistors in series/parallel:
P
In series: R D Ri
P
In parallel: 1=R D 1=Ri
Capacitors in series/parallel:
X
In series: 1=C D
1=Ci
X
In parallel: C D
Ci
Kirchhoffs Laws - KCL and KVL
X
X
P
KCL: At a node
ii n D
iout or ik D 0
X
KVL: Around a cycle
vk D 0
10V
19
i1
R2 = 2.5K
i3
5K
vo
R1 = 2.5K
i2
20
v1
b)
R1
vo
C1
v1
vo
C1
R1
Integrator
Differentiator
the workhorse of many analog circuits, they can be purchased very cheaply with
varying characteristics to suit a myriad of different applications.
74
741
offset
null
no
connect
-input 2
+input 3
-15V 4 -Vcc
+Vcc 7 +15V
6 output
offset
null
Figure 1.8 Common 741 Op Amp. Image from Jameco Electronics, cost $0.25
each.
In its basic form, an op amp has five terminals. Two of these terminals are used
to supply power to the circuit and are not normally shown when drawing op amp
circuits. For the most common op amps, the positive supply voltage is often +15
volts and the negative supply voltage -15 volts. The remaining three terminals
represent two input and one output terminal. The two input terminals are designated
the inverting and non-inverting terminals respectively. In diagrams the two input
terminals are often indicated with a plus and a minus sign (Figure 1.9)
Op amps are readily available from electronic stores and the most popular, the 741,
is priced at around 25 cents per unit.
21
Vs+
V+
V-
Vout
Vs-
Figure 1.9 Symbol for an Op Amp. Vs+ and Vs- are the positive and negative
supply voltages. V+ and V- are the inverting and non-inverting inputs.
R1
i
v2
R2
where v2 is the output voltage and R2 a load on the system. The two voltage
drops across the circuit can be described by:
v1
v2 D iR1
v2 D iR2
v1
R1
1C
R2
That is when R1 R2 :
v2 w v1
In other word, provided the output resistance R1 is low, the output voltage at v2
will be unaffected by the load R2 .
22
Saturation
v+ - v -
Example 1.1
An op amp has a gain of 100,000. In each of the following situations indicate the output
voltage:
a) V C D 4V; V
Vout D 100000 .4
b) V C D 2V; V
Vout D 100000 .2
D 2V
2/ D 0:2V
D 6V
6/ D
0:4V
23
D 4V
4/ D 0V
Although the basic op amp is essentially a very high-gain amplifier, it has certain
other properties which make it extremely useful. These properties are listed below:
1. The resistance (or impedance) into the input terminals is extremely high.
2. The resistance (or impedance) of the output terminal is extremely low.
The high impedance on the input terminals means than an op amp will impose
hardly any load on the circuit it is connected to. An op amp connected to a signal
source will have little effect on the signal voltage so that the full strength of the
signal can be fed input the op amp. The input resistance will often be of the order
of a least a mega ohm, often more.
A low impedance on the output terminal of the order of 70 ohms or less, means
that attaching loads to the op amp will have hardly any effect on the output voltage.
As a result of these properties, op amps are ideas components to use as buffering
circuits for the computational circuits described in the previous section.
Summary of Op Amp Properties:
High Gain
High Input Impedance
Low Output Impedance
To illustrate a simple example, consider the non-inverting amplifier shown in Figure 1.11. From the circuit we can determine the currents, i1 and i2 from the voltage
drops across, R1 and R2 , that is:
i1 D
vF 0
R1
i2 D
vO
vF
R2
Note that the ground voltage is 0. In an ideal op amp, the amount of current flowing
into the positive and negative inputs is zero owning to the high input impedance.
This means that the two currents, i1 and i2 must be equal, therefore it must be true
that:
vF
vO vF
vF
vF vO
D
)
C
D0
(1.1)
R1
R2
R1
R2
24
Another property of an op amp is that the output voltage is equal to the gain, A,
times the difference in the input voltage:
vO D A.vI
vF /
vO
A
Inserting vF into the current equation (1.1) gives:
(1.2)
vF D vI
vI
vO
vI
A C
R1
vO
A
R2
vO
D0
Since the gain, A, of an op am is very large, ideally infinite, therefore the above
equation simplifies to:
vI
vI vO
C
D0
R1
R2
This can now be rearranged to solve for vO :
R2
vO D vI 1 C
R1
The remarkable thing about this result is that the gain of the op amp, A, has vanished. This is an important effect because it means that the performance of the
circuit is only dependent on the two resistors, R1 and R2 . Any variation in the
op amp itself is eliminated. So long as the resistors are high quality this circuit
is immune to variation in the characteristics of the op amp and to environmental
variation such as temperature changes.
Computationally the circuit takes an input voltage, vI , and multiplies it by a scaling
factor. The only possible disadvantage to the non-inverting circuit is that the scaling
will be greater than one, it is no possible to scale vI by a factor < 1. The circuit
does however make an excellent signal amplifier.
More importantly, given the high input and low output impedance, the circuit doesnt
suffer the loading issues of a simple resistive circuit. Combining op amps with resistive/capacitive circuits is a common way to make analog computational modules
that can be easily connected together without the modules interfering with each
other.
Equation 1.2 is of interest because it helps to understand the operation of the circuit
intuitively. Given that the gain, A, of the op amp is huge we can approximate
equation 1.2 to:
vI D vF
25
vI
vF
R1
vo
i2
R2
i1
0v
The relationship tells us that what ever the input voltage, vI is, vF will match it. In
order for this to happen, as the input voltage, vI , changes the op amp must be going
through a dynamic change so that vF will match any change in vI . The explanation
is simple but crucial to an initiative understanding of the operation of the circuit.
First we must recall that an op amp is a difference amplifier. Imagine we increase vI
above vF , this means the difference .vI vF / increases which in turn is amplified
resulting in an increase in vO . Given the high gain of the op amp the difference
is likely to cause vO to increase quite significantly. As vO increases part of it
is passed back via resister R2 to vF so that vF increases. If vF increases, the
difference .vI vF / decreases, which in turn reduces the rise in vO . The feedback
cycle continues until vF converges to vI at which point the circuit reaches a steady
state, that is vO stops changing and has settled to a voltage that maintains vI D vF .
Figure 1.12 shows three computational circuits that use an op amp, a summer, an
integrator and a differentiator.
kx1
26
v3
v2
v1
R3
R2
R1
R4
vo
Summer
v1
v1
R1
C
R1
vo
Integrator
vo
Differentiator
Figure 1.12 Three different computational circuits made from op amps. Note the
similarity with the non-op amp circuits in Figure 1.7.
consider it the integral block. The input to this block will be the derivative, dx1 =dt
and the output will be x1 . However x1 has a negative term in the differential equation, kx1 , therefore we will also invert x1 at the same time. Once we have x1 we
only need to add a multiply by the constant k to obtain the value of dx1 =dt . This
final operation closes the loop. Figure 1.14 illustrates the complete block diagram.
From the block diagram we could construct this device out op amps such that the
electrical circuit will solve the differential equation. Before the advent of the digital
computer, such analog computers were common place and were use to solve a
variety of problems in science and engineering. One of the advantages of an analog
computer is that they are very fast and can be operated in a mode such that it
is possible to interact with the model in real time and the analog computer will
automatically adjust the solution.
With the advent of fast and particularly cheap digital computers the need for analog
computer has declined although there is increasing interest to use some of the ideas
in a new field called Neuromorhpic electronics [4, 49] where the aim is to recreate
brain like computational systems from electrical analog circuitry.
y1
y2
y3
Summation
27
y1+ y2 + y3
yo
yo
y
Mult by Constant y
Integration
Multiply Variables y1
y2
-y
ky
y1 y2
v2
x1 ! x2 ! x3
We assume that each chemical reaction is governed by simple mass action kinetics
so that the reaction rates, v1 and v2 are given by:
v1 D k1 x1
v2 D k2 x2
We can write down a simple mathematical model that describes the time evolution
28
dx1
dt
x1
dx1
D
dt
kx1
k1 x1
dx2
D k1 x1
dt
k2 x2
dx3
D k2 x2
dt
In this case the model is simple enough that we can derive the analytical solution for
this model. Assuming that at time zero, both x2 and x3 are at zero concentration,
then it can be shown that:
x1 .t/ D x1 .0/ e
x2 .t/ D
k1
k2
k1 t
x1 .0/ e
k2
x3 .t/ D x1 .0/ 1
k2 e
k1 t
k1 t
k2
e
k2 e
k1
k2 t
k2 t
In the majority of cases it is not possible to derive such an analytical solution and we
must therefore fall back to obtaining numerical solutions. To do this we must find a
suitable software application that can solve differential equations. The plot shown
in Figure 1.15 was obtained using Python (tellurium.analoglachine.org) but
similar plots could be obtained using Matlab.
29
Simulation Result
S2
S1
S3
10
9
Species Concentrations
8
7
6
5
4
3
2
1
0
0
10
time
Assignment 1
Obtain the same plot shown in Figure 1.15 using Matlab or Python.
30
this makes is easy to calculate the number of moles using the following relation
moles D
mass
molecular weight
Dimensional Analysis
Dimensional analysis is a simple but effective method for uncovering mistakes
when formulating kinetic models. This is particularly true for concrete models
where one is dealing with actual quantities and kinetic constants. Conceptual models are more forgiving and dont usually require the same level of attention because
they tend to be simpler.
Amounts of substance is usually expressed in moles and concentrations in moles
per unit volume (mol l 1 ). Reaction rates can be expressed either in concentrations
or amounts per unit time depending on the context (mol t 1 , mol l 1 t 1 ).
Rate constants are expressed in differing units depending on the form of the rate
law, the rate constants in simple first order kinetics are expressed in per unit time
(t 1 ), while in second order reactions the rate constant is expressed per concentration per unit time (mol 1 t 1 ).
In dimensional analysis, units on the left and right-hand sides of expressions must
have the same units (or dimensions). There are certain rules for combining units
when checking consistency in units. Only like units can be added or subtracted,
thus the expression S C k1 cannot be summed because the units of S are likely to
be mol l 1 and the units for k1 , t 1 . Even something as innocent looking as 1 C S
can be troublesome because S has units of concentration but the constant value 1
is unit-less. Quantities with different units can be multiplied or divided with the
units for the overall expression computed using the laws of exponents and treating
the unit symbols as variables.
Class Question 7
Determine the overall units for the expression, k1 S=Km where the units for each
variable are k1 .t 1 l), S.mol l 1 ) and Km .mol l 1 ).
In exponentials such as e x , the exponent term must be dimensionless, or at least
the expression should resolve to dimensionless, thus e k t is permissible but e k is not
1.10. APPROXIMATIONS
31
1.10 Approximations
By their very nature, models involve making assumptions and approximations. The
best modelers are those who can make the most shrewd and reasonable approximations without compromising a models usefulness. There are however some kinds
of approximations which are useful in most problems, these include:
Neglecting small effects.
Assuming that the system environment is unchanged by the system itself.
Replacing complex subsystems with lumped or aggregate laws.
Assuming simple linear cause-effect relationships where possible.
Assuming that the physical characteristics of the system do not change with
time.
Neglecting noise and uncertainty.
Neglecting small effects. This is the most common approximation to make. In
many studies there will always be parts of the system that have a negligible effect
on the properties of the system, at least during the period of study. For example,
the rotation of the earth, the cycle of the moon or the rising and setting of the sun
will most likely have negligible influence on our system. Assuming of course we
are not studying circadian rhythms.
Assuming that the system environment is unchanged by the system itself. This
is a basic assumption in any study. The minute a system starts to affect the environment we have effectively extended the system boundaries to include more of the
environment. It will often be the case that the interface between the environment
and the system will not be perfect so that there will be some effect that the system
has on the environment. So long as this effect is small we can assume that the
environment is not affected by the system.
Replacing complex subsystems with lumped or aggregate laws. Lumping subsystems is a commonly used technique in simplifying cellular models. The most
32
1.11 Behavior
We can classify the behavior of systems in to three broad categories, thermodynamic equilibrium, steady state or transient. The steady state can be further classified into stable or unstable steady states. For example, if we pick biochemical
systems as our example then:
Thermodynamic Equilibrium In this state the concentrations of reactants and
products show not net change over time; in addition, the rates of all forward
and reverse reactions are equal. This means that at thermodynamic equilibrium there is no net movement of mass from one part of the system to another
and not net dissipation of energy. In thermodynamics, equilibrium is the state
that maximizes the entropy of the system.
Steady State At steady state, the concentrations of reactants and products show
1.11. BEHAVIOR
33
not net change over time; however, unlike thermodynamic equilibrium there
is a net flow of mass or energy between the system and the environment.
At steady state the system will continually dissipate entropy to the external
environment while the entropy level of the system itself remains constant.
Transient State Under a transient state, a system will be moving from either one
steady state to another or from a steady state to thermodynamic equilibrium.
Systems
Thermodynamic
Equilibrium
Steady State
Stable
Transient State
Unstable
Exercises
1. Watch the YouTube videos: RC Integrator Circuit2 and RC Differentiator
Circuit3 (Figure 1.7).
2. Using the same approach to what was used in the non-inverting amplifier
(Figure 1.11), derive the relationship between the input and output voltage
for the inverting amplifier.
a) Show that the inverting amplifier behaves as a voltage scaler while at the
same time inverting the signal.
2
3
http://www.youtube.com/watch?v=JbpR5nGu0Gg
http://www.youtube.com/watch?v=qtrYd0uJzyA
34
Since the inputs to the op amp carry negligible current, we can assume that
i1 D i2 . i1 and i2 can be shown to equal:
i1 D
vI
vF
i2 D
R1
vF
v0
R2
vI
D
R1
or
R2
vI
R1
Equation v0 D R2 =R1 vI is negative and the term R2 =R1 scales the output
voltage. b) Show that the voltage at vF is zero. Using the explanation that
vI D vF for the non-inverting op amp, explain why vF is zero in an inverting
op amp.
v0 D
Since vF D
v0 =A and A 0, then vF D 0.
R2
vI
R1
i1
i2
vF -
vo
k1 x1
dx2
D k1 x1
dt
x2
1.11. BEHAVIOR
35
So1
dS1
dt
S1
k2
x
S 1 k2
x
k1
S 1 k2
+
dS2
dt
So2
S2
36
v1
R1
vo
C1
i2
Prey Predator
dPredator
D Prey Predator
dt
Predator
where is the growth rate of the prey, is the rate of predation,
is the death
rate of the predator and is the rate of growth in the predator population.
b) Explore how the parameters affect the evolution of the populations. The
parameter can range between 0.1 and 2.0. Set the initial population of prey
and predators to 1 each. Your grading will be determined by how well you
present your results.
4. Biochemical Models
Build a computer model of the following simple metabolic pathway:
v1
v2
v3
X 0 ! x1 ! x2 ! X 1
Assume that the metabolites, Xo and X1 are fixed. Assume also that v1 to v3
are governed by the rate laws:
v1 D E1 .k1 Xo
k2 Sx1/
v2 D E2 .k3 x1
k4 x2 /
v3 D E3 k5 x2
where Ei is the concentration of the i t h enzyme and ki are the kinetic rate
constants.
1.11. BEHAVIOR
37
a) Write out the state variables, suggested inputs, and parameters of the system.
b) Write out the differential equations for this system.
c) Assign arbitrary but realistic values to the various parameters in the model
and compute the steady state, that is when dxi =dt D 0. Report the steady
state concentrations of x1 and x2 .
d) What is the effect of doubling all Ei by a factor on the steady state
concentrations of x1 and x2 ?
5. Pharmokinetic Model
Concentration
v1
Exchange
Accumulation
Site of
Action
v6
Time
Blood
Stream
Db
Da
v5
v3
v2
Liver
Dl
v4
Degradation
Excretion
Figure 1.21 Simple model of a drug begin injected and distributed through the
body.
We will assume that movement of the drug between compartments obeys
simple first-order kinetics. For example, the rate, v2 , at which drug enters the
liver is given by k2 Db where Db is the concentration of drug in the blood
38
stream and k2 is the kinetic constant for the process. The complete list of
rates is given below:
v2
v3
v4
v5
v6
D k2 Db
D k3 Dl
D k4 Dl
D k5 Db
D k6 Db
where Db is the concentration of drug in the blood stream, Dl the concentration of drug in the liver, and Da the concentration of drug at the site of
action. Using the conservation of mass, write out three differential equations
that describe the rate of change of drug in each of the three compartments.
Previous pharmokinetic studies of the drug indicate that the rate constants for
the entry and exit of drug to and from the various compartments are given by:
k2 D 6:12; k3 D 0:2; k4 D 0:45; k5 D 1; k6 D 5.
All units in seconds. Using a simulation model answer the following questions.
Rate of Infusion
a) Assume that the nurse can inject 1 mM of drug per second into the patients
blood stream. That is, in one second the concentration of drug in the blood
stream increases by 1 mM. This is another way of saying that v1 D 1 mM
sec 1 . The nurse can start and stop the injection very quickly such that the
profile one observes in the rate at which drug enters the blood looks like a
pulse as shown in Figure 1.22.
Time (Secs)
1.11. BEHAVIOR
39
Answer: The injection must last at least 6 seconds. b) The model has a
number of assumptions, for example, that mixing of drug is instantaneous in
each compartment, that weve accounted for all the possible compartments
and that the rate constants are accurate. But there is one particularly bad
assumption which makes the model and its predictions highly suspect. What
might that be?
Answer: All compartment volumes are identical
40
equilibrium.
Class Question 2
For the water tank model described in section 1.2, write out the variables and parameters of the system. Identify the boundary parameters.
State Variables: Heights, h1 and h2 in tank one and two.
Outputs: Heights h1 and h2 .
Parameters: K1 , K2 and the cross-sectional area, A.
Boundary parameters (Inputs): Q1 and Q4 .
Independent variable: Time, t.
Class Question 3
How you would describe each of the following systems in terms of how you might
model them?
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
A game of chess
Two teams playing American Football
A forest fire
Gene regulation in a bacterium
The beating of a healthy heart
A tornado
Growth of a tumor
Upper atmosphere chemistry
A digital computer
Electron flow through a single transistor
Population of cells from which emerges a cancerous cell
Metabolism
The swimming of a single E. coli through water
Class Question 4
Given the resistor network shown in Figure 1.23, determine the voltage Vo .
When youve derived the relation for vo , assume R2 R1 and then infer what
calculation could be done with the circuit.
i1 C i2 D i3
vo D i3 R2 ;
i3 D
vj
R2
1.11. BEHAVIOR
41
v1
v2
R1
R2
i1
i2
i3
vo ?
R1
v1
vo
i2 D
v2
vo
R1
R1
Inserting i1 ; i2 and i3 into the current law (KCL) and solving for Vj yields:
vo D
v1 C v2
2 C R1 =R2
1
.v1 C v2 /
2
1X
vi
n
In other words, this resistive network computes the average of the input voltages.
This only works if R2 draws a negligible amount of current.
Class Question 5
Show that the resistive circuit in Figure 1.24 can be used to compute a voltage times
a constant.
We will assume there is no load on the circuit, that is i3 D 0. The voltage across the
potentiometer between 10 volt rail and the wiper location is therefore: i2 R2 . The
voltage vo is then given by the difference:
vo D 10
i2 R2
The current drop between ground and the wiper position is given by:
vo D i2 R1
i2 D
vo
R1
42
10V
i1
R2 = 2.5K
i3
5K
vo
R1 = 2.5K
i2
vo
R2
R1
10
1 C R2 =R1
(1.3)
Assuming R1 D R2 D 2:5K:
vo D 10
1
D 5 volts
2
That is, the circuit multiplies by a constant of 0.5. By changing the wiper position
different constants can be dialed in.
Class Question 6
What is the problem with all these circuits that could make them potentially poor
at computing?
Answer: Loading. A downstream circuit that draws too much current will affect the
output voltage and hence the answer to the computation. Consider the following
load on the multiply by a constant circuit (Figure 1.25).
Now assume a load R3 on the circuit, b). This time i3 0, and therefore i1 D
i2 C i3 . i1 is then given by:
vo
vo
i1 D
C
R1
R3
Taking a similar approach
vo D 10
vo
vo
C
R2
R1
R3
1.11. BEHAVIOR
43
a)
10V
b)
i1
R2 = 2.5K
5K
i3
i1
i3
vo
vo
R1 = 2.5K
i2
i2
R3
10K
10
1 C R2 =R1 C R2 =R3
If R3 is very large the equation reduces to the case when there is no loading. The
smaller R3 , i.e. the larger the loading, the smaller the output voltage vo and the
worse the calculation. For example, using the resistor values given in the diagram
yields:
1
D 4:44 volts
vo D 10
2:25
If the circuit applied the multiplication perfectly we expected a result of 5 volts,
instead we got 4.44 volts which is roughly a 11% error and would clearly be unacceptable in a real application. Often computing circuits will be connected to other
computing circuits so that the loading resistance will vary and compound from circuit to circuit resulting in considerable error accumulation across the network. For
this reason, using passive resistor circuits to carry out computations is not always
practical.
Class Question 7
Determine the overall units for the expression, k1 S=Km where the units for each
variable are k1 .t 1 l), S.mol l 1 ) and Km .mol l 1 ).
We first write out the expression in terms of the individual units:
t
l mol=.mol l
by treating the symbols are algebraic variables we can see that the symbol mol will
cancel and using exponents rules we can bring the l 1 term up to the enominator to
yield:
t 1l 2
44
Simulation Software
Cogito ergo sum I think therefore I am
Ren Descartes, 1664
http://www.mathworks.com
http://www.wolfram.com
45
46
many researchers have either written their own solvers or use publicly available
software libraries.
There are free Matlab like tools such as Octave3 , SciLab4 , Euler5 and Yorick6 that
can be successfully used. There is also considerable support for solving ODEs
from the popular scripting language Python7 . In addition to these there are specialized tools for modeling specific types of physical systems. For example, Python
based tool Tellurium 8 has specific support for describing chemical or biochemical
networks.
%
%
%
%
http://www.gnu.org/software/octave/
http://www.scilab.org/
5
http://euler.sourceforge.net/
6
http://yorick.sourceforge.net/
7
http://www.python.org/
8
tellurium.analogmachine.org
4
47
Listing 2.1 A Matlab script for solving a simple set of differential equations and
plotting the results.
x C y
dy
D x
dt
dz
D
dt
yz
z C xy
The first example illustrates how Matlab can be used to solve the differential equations. To solve differential equations we must first declare a function that computes
the right-hand sides of the differential equations and return the derivatives to the
caller. This is shown below:
function xprime = lorenz(t,x);
% Computes the derivatives of
% the differential equations
sig=10;
beta=8/3;
rho=28;
48
xprime=[-sig*x(1) + sig*x(2);
rho*x(1) - x(2) - x(1)*x(3);
-beta*x(3) + x(1)*x(2)];
To solve the Lorenz system, one of the Matlab provided ODE solver routines can
be used. In the script below ode45 is used which implements an adaptive step-size
Runge-Kutta method (See Appendix B).
>>x0 = [-8 8 27];
>>tspan = [0, 20];
>>[t,x] = ode45 (@lorenz, tspan, x0)
>>plot (t, x)
The first line sets up the initial conditions for the three variables, the second up the
duration of the simulation, and the third line carries out the actual simulation. The
last line plots the data as a graph.
The next example shows how we can use Python to solve the Lorenz system. This
relies on the SciPy and NumPy packages The SciPy package includes a wide variety
of numerical methods for solving different kinds of problem. The first part of the
script imports the required libraries:
import scipy
from scipy.integrate import odeint
import numpy
import matplotlib.pyplot as plt
The second part of the script defines the system of differential equations:
def lorenz(x, t):
sigma = 10
rho = 28
beta = 8.0/3
return [ sigma * (x[1] - x[0]),
x[0] * (rho -x[2]) - x[1],
x[0] * x[1] - beta* x[2]]
Note the indentation of the lines, this is important for Python. Finally we invoke
the ODE solver and plot the resulting simulation data.
xInitial = [0, 1, 1.05]
49
p = defn cell
$Xo -> S1; k1*Xo - k2*S1;
S1 -> S2; k3*S1/(k4 + S1);
S2 -> $X1; k5*S2^2;
end;
// Initialize parameters
p.Xo
p.k1
p.k3
p.k4
=
=
=
=
10;
0.6; p.k2 = 0.4;
3.4; p.k4 = 0.1;
1.2;
50
Further Reading
1. Scilab web site www.scilab.org
2. Jarnac web site http://sbw-_app.org/jarnac/
51
52
53
54
Figure 3.1 A typical governor from J. Farley, A Treatise on the Steam Engine:
Historical, Practical, and Descriptive (London: Longman, Rees, Orme, Brown, and
Green, 1827, p436
engine speed.
During this period devices for automatic control were designed through trial and
error and little theory existed to understand the limits and behavior of feedback
control systems. One of the difficulties with feedback control is the potential for
instability. As the governor became more widespread, improvements were made
in manufacturing mechanical devices which reduced friction. As a result engineers
began to notice a phenomena they termed hunting. This was where, after a change
in engine load, the governor would begin to hunt in an oscillatory fashion for
the new stream rate that would satisfy the load. This effect caused considerable
problems with maintaining a stable engine speed and resulted in James Maxwell
and independently Ivan Vyshnegradskii, undertaking the first theoretical analysis
of a negative feedback system.
Until the 20th century, feedback control was generally used as a means to achieve
automatic control, that is to ensure that a variable, such as a temperature or a pressure was maintained at some set value. However, entirely new applications for
feedback control emerged early in the 20th century, these included artillery tracking on ships and communication. In naval warfare a major issue is being able to
accurately fire a weapon from a moving ship to a moving target. In the early part
of the 20th century mechanical analog computers called rangekeepers were developed that could continuously compute a target bearing, predict the future target
position and make adjustments to the weapon to account for other factors such as
wind. Tracking uses negative feedback and when employed in this mode it is called
55
56
input. We can see both applications at work in the eye. On the one hand there is the
need to control the level of light entering the pupil where the diameter of the pupil
is controlled by two antagonistic muscles. If the external light intensity increases,
the muscles respond by reducing the pupil diameter, whereas the muscles increase
the pupil diameter if the light intensity falls. The pupil reflex serves as an example
of negative feedback used in a regulator mode. On the other hand the eye also need
to track objects that involves maintaining the eyeball fixed on the object. In this
mode the eye functions as a servomechanism.
Both regulator and servomechanism can be implemented using the same operational mechanism. Figure 3.2 shows a generic negative feedback circuit. On the
left of the figure can be found the input, sometimes termed the desired value or
more often the set point. If the circuit is used as a servomechanism then the output
tracks the set point. As the set point changes the output follows. If the circuit is
used as a regulator or homeostatic device then the set point is held constant and the
output is maintained at or near the set point even in the face of disturbances.
The central mechanism in the feedback circuit is the generation of the error signal,
that is the difference between the desired output (set point) and the actual output.
The error is fed into a controller (often something that simply amplifies the error)
which is used to increase or decrease the process. For example, if a disturbance on
the process block reduces the output, then the feedback operates by generating a
positive error, this in turn increases the process and restores the original drop in the
output.
Disturbances
Set Point
Error
Controller
Process
Output
Feedback
57
logical system. In addition, biological system are invariably more complex with
multiple nested feedback loops and multiple inputs and outputs. If it remarkable,
that even after 50 or 60 years of research, the role of many of the feedback systems
in biochemical networks is still highly speculative.
yi
yo
K
Figure 3.3 Generic structure of a negative feedback system.
We will assume some very simple rules that govern the flow of information in
this feedback system. For example, the output signal, yo , will be the value of A
multiplied by the error, e. The feedback signal will be assumed to be proportional
to yo , that is Kyo . Finally, the error signal, e will be given by the difference between
the set point, yi and the feedback signal, Kyo (Figure 3.4).
yi
e = yi -Kyo
A
yo = Ae
Kyo
Ayi
1 C AK
where:
GD
A
1 C AK
(3.1)
58
G is called the gain of the feedback loop, often called the closed loop gain. Gain
is a term that is commonly used in control theory and refers to the scalar change
between an input and output. Thus a gain of 2 simply means that a given output
will be twice the input.
The gain is a measure of the change that occurs between a signal output
and its input. A gain of two means that the output will change two times
in magnitude compared to a change in the input.
In addition to the close loop gain, engineers also define two other gain factors, the
open loop gain and the loop gain. The open loop gain is simply the gain from the
process, A. It is the gain one would achieve if the feedback loop were absent. The
loop gain is the gain from the feedback and process A combined, AK. The loop
gain is an important quantity when discussing the stability of feedback circuits.
Figure 3.4 illustrates the different types of gain in a feedback circuit.
Closed Loop Gain
A
K
Loop Gain
59
e D yi
Ky:
Eliminating e, we find
Ayi d
:
1 C AK
The sensitivity of the output to the disturbance is then
@y
1
D
:
@d
1 C AK
The sensitivity decreases as the loop gain AK is increased. In practical terms,
this means that the imposition of a load on the output, for example a current drain
in an electronic circuit, protein sequestration on a signaling network or increased
demand for an amino acid will have less of an effect on the circuit as the feedback
strength increases. In electronics this property essentially separates the network
into functional modules.
yD
Last but not least, feedback also improves the fidelity of the response. That is, for
a given change in the input, a system with feedback is more likely to faithfully
reproduce the input at the output than a circuit without feedback. An ability to
faithfully reproduce signals is critical in electronics communications and in fact it
was this need that was the inspiration for the development of negative feedback in
the early electronics industry.
60
Linearization Properties Consider now the case where the amplifier A is nonlinear. For example a protein cascade pathway exhibiting a sigmoid response. Ten
the behavior of the system G (now also nonlinear) is described by
G.yi / D yo D A.e/
e D yi
Kyo D yi
KG.yi /:
Differentiating we find
G 0 .yi / D A0 .yi /
Eliminating
de
,
dyi
de
dyi
de
D1
dyi
KG 0 .yi /:
we find
G 0 .yi / D
A0 .yi /
:
1 C A0 .yi /K
One objection to the implementation of feedback as described above is that the system sensitivity is not actually reduced, but rather is shifted so that the response is
more sensitive to the feedback K and less sensitive to the amplifier A. However, in
each of the cases described above, we see that it is the nature of the loop gain AK
(and not just the feedback K) which determines the extent to which the feedback
affects the nature of the system. This suggests an obvious strategy. By designing
a system which has a small clean feedback gain and a large sloppy amplifier,
one ensures that the loop gain is large and the behavior of the system is satisfactory.
Engineers employ precisely this strategy in the design of electrical feedback amplifiers, regularly making use of amplifiers with gains several orders of magnitude
larger than the feedback gain (and the gain of the resulting system).
61
1. Amplification of signal.
2. Robustness to internal component variation.
3. High fidelity of signal transfer.
4. Low output impedance so that the load does not affect the performance of
the circuit.
These are the main advantages of negative feedback but as we will see in a later
chapter on frequency response, feedback can confer additional useful properties.
62
Objective: Regulate the steam inlet so that the steam engine output is constant.
eo
R1
R2
Figure 3.8 Negative feedback used in servo mode, tracking and amplifying an
input signal.
4. Amino acid biosynthesis and glycolytic pathway, Figure 3.9.
Objective: Ensure that as demand for amino acids or ATP varies, the supply remains constant.
63
Figure 3.10 A governor controlling the steam inlet valve as a function of the speed
of the engine. Illustration from An Elementary Treatise on Stream and the Steamengine by Clark and Sewell published in 1892.
64
Class Question 2
Figure 3.2 shows a generic feedback circuit. Identify the set point in a voltage
follower circuit.
65
Exercises
1. Define negative feedback.
Two points
Negative feedback is where part of the output of a system is used to reduce
the magnitude of the system input, or something similar.
2. Define each of the following terms:
(a) Open Loop Gain
(b) Closed Loop Gain
(c) Loop Gain
Two points for each
a) The open loop gain refers to the gain generated by the amplifier A, independently of the any feedback mechanism.
b) The closed loop gain refers to the gain between the output and input of a
system that has an operational negative feedback loop
c) The loop gain refers to the gain net gain generated by the amplifier, A and
feedback strength K, often abbreviated to Ak.
3. List four benefits of negative feedback
Two points for each
a) Amplification of signal.
b) Robustness to internal component variation.
c) High fidelity of signal transfer.
d) Low output impedance so that the load does not affect the performance of
the circuit.
4. If the process block A is some nonlinear function A.e/, show that negative
feedback (Figure 3.2) will still maintain linearity between y0 and yi so long
as the gain of the process block is sufficiently high.
Five points
See section Linerization Properties section on page: 3.3
5. Investigate the physiology and molecular biology literature and describe two
other systems where negative feedback is used. Suggest a possible evolutionary advantage for such feedback in your selected systems.
66
6. Build a computer simulation of a simple water tank model that uses negative
feedback to control the water level.
Biological Networks
In a system with many interactions many properties arise which cannot
be assigned to any one isolable entity...
Henrik Kacser, 1963
67
68
could be its own gene) or via other transcription factors integrate multiple signals
on to another gene. Additionally, some transcription factors only become active
when phosphorylated or unphosphorylated by protein kinases and phosphatases.
Like protein signaling and control networks, gene regulatory networks can be elaborate, structurally and computationally.
Gene Activation
Gene Repression
Multiple Control
Gene Cascade
Auto-Regulation
69
Figure 4.2 Example of a mixed network involving gene regulatory and protein
phosphorylation networks in Caulobacter. Blunt ends to regulatory arcs indicate
inhibition while arrow ends indicate activation. Image from BioMed Central [6].
Xo
S1
v1
v2
S2
v3
v4
Figure 4.3 A cascade of two genes. v1 represent the expression rate for protein S1 .
v2 represents the degradation rate of S1 . v3 is the expression rate of protein S2 . v4
is the degradation rate of S2 . Xo is the input inducer to the cascade.
70
71
include control networks that are concerned with monitoring and coordinating internal changes, the most well known of these includes the cell cycle control network. Many external signals act by binding to cell-surface receptor proteins such
as the large family of receptor tyrosine kinases and G-protein coupled receptors.
Once a signal is internalized through the cell-surface receptors, other proteins, including protein kinases and phosphatases continue to process the signal often in
coordination with other signaling networks. Eventually the signalling pathway terminates on target proteins that leads to a change in the cells behavior. Such targets
can include a wide variety of processes such as metabolic pathways, ion channels,
cytoskeleton, motor proteins and gene regulatory proteins.
The molecular mechanisms employed by signalling and control pathways include
covalent modification, degradation and complex formation. Covalent modification,
in particular, is a common mechanism used in signaling networks and includes a variety of different modifications such as phosphorylation, acetylation, methylation,
ubiquitylation, and possibly others [5]. As a result the structure and computational
abilities [44] of such networks can be extremely elaborate. It has been estimated
from experimental studies that in E. coli, 79 proteins can be phosphorylated [32] on
serine, threonine and tyrosine side groups whereas in Yeast, 4000 phosphorylation
events involving 1,325 different proteins have been recorded [39].
The cell cycle control network is an excellent example of a sophisticated protein
control network that coordinates the replication of a biological cell. The cell cycle
includes a number of common molecular mechanisms that are found in many other
protein networks. These mechanisms can be grouped into at three broad types, they
include, phosphorylation, degradation and complex formation. Phosphorylation is
a common mechanism for changing the state of a protein and involves phosphorylation on a number of sites on the protein surface including serine/threonine and
tyrosine. In prokaryotes, histidine, arginine or lysine can also be phosphorylated.
Phosphorylation is mediated by kinases of which the Human genome may have
over 500 kinase encoding genes [33]. The effect of phosphorylation is varied but
generally it causes the protein undergoing phosphorylation to change catalytic activity, to change the proteins visibility to other proteins or to mark the protein for
degradation. For example, src is a tyrosine kinase protein involved in cell growth.
It has two states, active and inactive; when active it has the capacity to phosphorylate other proteins. Deactivation of src is achieved by phosphorylation of a tyrosine
group on the C-terminal end of the protein. Dephoshorylation of the tyrosine group
by tyrosine phosphatases results in the activation of the protein.
Phosphorylation can also be used to inactivate enzymes such as glycogen synthase
by the glycogen synthase kinase 3 protein. In the Yeast cell cycle, the protein Wee1
is phosphorylated and inactivated by the complex Cdc2-Cdc13. Active Wee1 in
72
turn (that is the unphosphorylated form) can inactivate Cdc2-Cdc13 by phosphorylating the Cdc2 subunit.
In addition to changing the activity of proteins, phosphorylation can also be used
to mark proteins for degradation. For example, the protein Rum1 that is part of
the Yeast cell cycle control network can be phosphorylated by Cdc2-Cdc13. Once
phosphorylated, Rum1 is degraded. Degradation itself is an important mechanism
used in protein networks and allows proteins to be rapidly removed from a network
according to the state of the cell. Degradation is usually mediated by ubiquitylation. For example, Cdc2-Cdc13, via Ste9 and APC is marked for degradation by
ubiquitylation (Rum1 is similarly processed once phosphorylated). Once marked
this way, such proteins can bind to the proteasome where they are degraded. Finally binding of one protein to another can change the target proteins activity or
visibility. An example of this is the inactivation of Cdc2-Cdc13 by Rum1. When
unphosphorylated Rum1 binds to Cdc2-Cdc13, the resulting complex is inactive.
Different combinations of these basic mechanisms are also employed. For example, phosphorylation of complexes can lead to the dissociation of the complex, or
the full activity of a protein requires multiple phosphorylation events. Although
signalling networks can appear highly complex and varied, most of them can be reduced to the three fundamental mechanisms, covalent modification, selective degradation and complex formation (Fig 4.4).
In higher eukaryotic cells, particulary human, around 2% of the protein-coding
part of the genome is devoted to encoding protein kinases, with perhaps 10% of
the coding region dedicated to proteins involved in signalling networks. It has also
been suggested that possibly as much as 30% of all cellular proteins in yeast and
human can be phosphorylated[10].
The actual size of the networks themselves is however even larger that these numbers suggest because of the significant number of covalent variants and binding
permutations. For example, p53, the tumor suppressor protein, has between 17
and 20 phosphorylation sites alone [48]. If every combination were phenotypically
significant, though unlikely, that amounts to at least 131,072 different states.
Ptacek and Snyder [40] have published a review on elucidating phosphorylation
networks where much more detailed information is given.
73
Complex Formation
Phosphorylation
Dephosphorylation
Degradation of a
complex
Degradation of a
phoshporylated form
The reactants appear on the left of the equation and the products on the right. Both
sides are separated by an arrow indicating the positive direction of the transformation. The simplest possible reaction is the conversion of a single reactant, A, into a
single product, B, as depicted in the following way:
A!B
Such a reaction can be studied by observing the change in concentration of A and/or
B in time. Experimentally there are a variety of ways to do this, for example by
observing the emission or absorption of light at a specific wavelength, the change
in pH, or the incorporation of a radioactive or heavy isotope into the product. An
example of an actual biochemical reaction is the familiar interconversion of the
adenine nucleotides:
2 ADP ! ATP C AMP
This describes two molecules of ADP being transformed into one molecule of ATP
and one molecule of AMP. Sometimes a double arrow is used to explicitly indicate
that a reaction is reversible, as in:
2 ADP
ATP C AMP
74
If a reaction is reversible (as almost all reactions are to some extent), then the
reaction rate can be positive or negative. By convention, a positive rate means that
the reaction progresses from left to right, whereas a negative rate indicates a right
to left reaction.
Example 4.1
What does the following reaction notation mean:
3A C 4B ! 3C C D
This notation means that during a reaction event, 3 molecules of A and 4 molecules of B
react to form 3 molecules of C and one molecule of D.
Stoichiometric Amounts
The stoichiometric amount is defined as the number of molecules of a particular
reactant or product taking part in a reaction. Stoichiometric amounts will always
be positive numbers. For example, in the reaction:
2 ADP
ATP C AMP
ADP has a stoichiometric amount of two, ATP a stoichiometric amount of one,
and AMP also with a stoichiometric amount of one. If the same species occurs on
the reactant and product side of a reaction then it must be treated separately. For
example, in the reaction:
2A C B C C ! 3A C D C 2B
The stoichiometric amounts on the reactant side include: A with two, B with one
and C with one. On the product side the stoichiometric amounts include: A with
three, D with one and B with two.
The stoichiometric amount is the number of molecules of a particular reactant or product taking part in a reaction.
75
Example 4.2
List the stoichiometric amounts in the following reaction:
2A C B ! A C C
On the reactant side the stoichiometric amount for A is two and for B is one. On the product
side, the stoichiometric amount for A is one and for C one.
Stoichiometric Coefficients
Stoichiometry deals with static information about the amounts of substances involved in a chemical transformation, whereas kinetics relates rates of change that
occur in these amounts. To paraphrase a statement made by Aris [1], stoichiometry
provides the framework within which chemical change takes place irrespective of
the forces that bring them about, and by kinetics the speed of chemical change. Aris
then went on to state, Just as the latter can only be built on a proper understanding
of the kinematics, so the analysis of stoichiometry must precede that of kinetics.
We will do the same here.
The stoichiometry coefficient refers to the relative amount of substance that is consumed and/or produced by a reaction. Given a reaction such as:
2A ! B
the stoichiometric amount of A is 2 and for B, 1. The species stoichiometry or
stoichiometric coefficient however, is the difference between the stoichiometric
amounts if a given species on the product side and the stoichiometric amount of
the same species on the reactant side. The definition below summarizes this more
clearly.
The stoichiometric coefficient, ci , for a molecular species Ai , is the difference between the stoichiometric amount of the species on the product side and
the stoichiometric amount of the same species on the reactant side, that is:
ci D Stoichiometric Amount of Product, Ai
Stoichiometric Amount of Reactant, Ai
In the reaction, 2A ! B, the stoichiometric amount of A on the product side is
zero while on the reactant size it is two. Therefore the stoichiometric coefficient
76
77
The stoichiometric amount of A on the reactant side is 1 and on the product side 12 , therefore
the stoichiometric coefficient for A is 1=2 1 D 1=2. The stoichiometric amount of B
on the reactant side is 0 and on the product side, 1, therefore the stoichiometric coefficient
for B is 1 0 D 1.
Example 4.3 (b) highlights another fact about stoichiometric coefficients. The coefficients can be fractional amounts, often represented as rational fractions.
Rates of Change
The rate of change can be defined as the rate of change in concentration or amount
(depending on units) of a designated species. If S is the species then the rate of
change is given by:
S
Rate D
t
Because rates change as reactants are consumed and products made, the rate of
change is better defined as the instantaneous change in concentration, or a derivative:
dS
Rate D
(4.1)
dt
If we were to plot the rate of product formation as a function of time, the rate of
reaction would be given by the slope of the curve (Fig. 4.5). If concentrations are
measured in moles per liter (L) and time in seconds (sec), then the rate of reaction
is expressed in mol L 1 sec 1 .
When reporting a rate of change, it is important to give the name of the species that
was used to make the measurement. For example, in the reaction 2A ! B, the rate
of change of A is twice the rate of change of B. In addition, the rate of change of
A is negative because it is consumed, whereas the rate of change of B is positive
because it is being made.
Concentration of Product, B
78
4
3
Slope = Rate of change of product
2
1
0
Time
Reaction Rates
In this section we will introduce the concept of a reaction rate, denoted by v.
The standard unit for the reaction rate is amount per volume per time. This is an
intensive property, which does not depend on the amount of substance, for example
mol L 1 sec 1 . In a previous section we introduced the rate of change. In practice
it is the rate of change that we measure experimentally. We also briefly mentioned
that in the reaction 2A ! B, A is consumed twice as fast as the production of
product, B. This means that the sign and magnitude of the rates of change will
vary depending on which species we choose to measure.
A simple way to avoid these differences is to divide each rate of change by the
species stoichiometric coefficient. In this case the stoichiometric coefficient of A
is 2 and for B is C1. If we do this we obtain:
1 dB
1 dA
D
Dv
2 dt
1 dt
In general, for a reaction of the form
n1 A C n2 B C : : : ! m1 P C m2 Q C : : :
79
where n1 ; n2 ; : : : and m1 ; m2 ; : : : represent the stoichiometric coefficients, the reaction rate is given by:
1 dB
1 dP
1 dQ
1 dA
D
::: D
D
:::
(4.2)
n1 dt
n2 dt
m1 dt
m2 dt
Defined this way, a reaction rate is independent of the species used to measure
it. The same applies if a given species appears on both sides of a reaction. For
example, in the reaction A ! 2A, the stoichiometric coefficient is C1 so that the
reaction rate, v, is:
Rate D v
vD
1 dA
C1 dt
To make the definition of the reaction rate more formal, let us introduce the extent
of reaction, indicated by the symbol, . We define a change from to Cd in time
dt to mean that c1 d moles of A1 , c2 d moles of A2 etc, react to form cn d moles
of An etc. By this definition we can state that for any component i, the following is
true for the time interval dt :
dni D ci d
(4.3)
or
d
dni
D ci
dt
dt
where ni equals the amount in moles of species i . From this relation we define the
extensive rate of reaction, vE , to be:
vE
In other words
d
dt
dni
D ci vE
dt
(4.4)
For the moment we will use vE and vI to distinguish the extensive and intensive
reaction rates. Note that has units of amount and vE has units of amount per
unit time and is therefore an extensive property, being dependent on the size of
the system. The advantage of introducing the extent of reaction is that it allows
us to formally define the rate of reaction independently of the species we use to
measure the rate. This convenient property can be expressed as:
vE
d
D
dt
1 dn1
D
c1 dt
1 dn2
1 dnn
1 dnnC1
::: D
D
:::
c2 dt
cn dt
cnC1 dt
80
Example 4.4
Express the rate of reaction and the rates of change for the following biochemical reaction:
2 ADP ! ATP C AMP The rate of reaction is given by
vD
d
dn.ATP/
dn.AMP/
D
D
dt
dt
dt
1 dn.ADP/
2
dt
If the volume, V , of the system is constant we can also express the rate in terms of
concentration, Ci D ni =V .
We can therefore rewrite the rate of reaction in the form:
vE
D
V
1 dC1
D :::
c1 dt
where vE has units of amount per unit time (mol s 1 ). The relation vE =V is
the intensive version of the rate, vI , with units of concentration per unit time
(mol L 1 s 1 ) and is the most commonly used form in biochemistry.
vI D
1 dCi
vE
D
V
ci dt
or
dCi
D ci vI
dt
(4.5)
(4.6)
81
form should be clear from the context. In this book, where we use v, we will
generally mean vI , the intensive form.
In situations involving multiple compartments of different volumes or where there
are specific mass conservation laws at work, the intensive rate is not appropriate.
This is because the intensive version is unable to keep track of the total number of
moles undergoing transformation. In these situations it is necessary to deal explicitly with the extensive rate of reaction, in other words:
dni
D V ci vI
dt
A Word on Notation
In many texts, the concentration (molarity) of a substance, X, is denoted using
square brackets, as in [X]. To avoid unnecessary clutter in the current text, the
use of square brackets to indicate molarity will be relaxed.
82
"vSi
D
@v Si
@Si v
D
Sj ;Sk ;:::
@ ln v
v%=Si %
@ ln Si
(4.7)
@v A
D0
@A v
2. v D kA
v
Elasticity: "A
D
Ak
@v A
D
D1
@A v
kA
3. v D kA2
v
Elasticity: "A
D
2kAA
@v A
D2
D
@A v
kA2
4. v D kAn
v
Elasticity: "A
D
@v A
nkAn 1 A
D
Dn
@A v
kAn
Example (4.5) shows that the elasticity corresponds to the expected kinetic order
for simple rate laws. The definition of the elasticity (4.7) also gives us a useful
operational interpretation.
Operational Definition: The elasticity is the fractional change in reaction rate
in response to a fractional change in a given reactant or product while keeping
all other reactants and products constant.
83
That is, the elasticity measures how responsive a reaction is to changes in its immediate environment. Since the elasticity is expressed in terms of fractional changes,
it is also possible to get an approximate value for the elasticity by considering
percentage changes. For example, if we increase the substrate concentration of a
particular reaction by 2% and the reaction rate increases by 1:5%, then the elasticity is given by 1:5=2 D 0:75. The elasticity is however only strictly defined (See
equation (4.7)) for infinitesimal changes and not finite percentage changes. So long
as the changes are small, the finite approximation is a good estimate for the true
elasticity.
For a given reaction, there will be as many elasticity coefficients as there are reactants, products and other effectors of the reaction. For species that cause reaction
rates to increase, the elasticity is positive, while for species that cause the reaction rate to decrease, the elasticity is negative. Therefore, reactants generally have
positive elasticities and products generally have negative elasticities.
For a species that increases the reaction rate: Elasticity is positive.
For a species that decreases the reaction rate: Elasticity is negative.
Log Form
The definition of the elasticity in equation (4.7) shows the elasticity expressed using
a log notation:
@ ln v
"vS D
@ ln S
This notation is frequently used in the literature. The right panel of Figure 4.6
shows one application of this notation, namely that a log-log plot of reaction rate
versus reactant concentration yields a curve where the elasticity can be read directly
from the slope. The origin of this notation will be explained here.
Consider a variable y to be some function f .x/, that is y D f .x/. If x increases
from x to .xCh/ then the change in the value of y will be given by f .xCh/ f .x/.
The proportional change however, is given by:
f .x C h/ f .x/
f .x/
The rate of proportional change at the point x is given by the above expression
divided by the step change in the x value, namely h:
84
B.
A.
1
0:8
0:6
ln v
@v
@S
0:4
"vS
@v S
@S v
20
@ ln v
@ ln S
2
4
0:2
0
"vS D
40
0 2
ln S
Figure 4.6 A. The slope of the reaction rate versus the reactant concentration
scaled by both the reactant concentration and reaction rate yields the elasticity, "vS .
B. If the log of the reaction rate and log of the reactant concentration are plotted,
the elasticity can be read directly from the slope of the curve. Curves are generated
by assuming v D S=.2 C S/.
Rate of proportional change =
1
f .x C h/
f .x C h/ f .x/
D
lim
h!0
hf .x/
f .x/ h!0
h
lim
f .x/
1 dy
y dx
From calculus we know that d ln y=dx D 1=ydy=dx, therefore the rate of proportional change equals:
d ln y
dx
and serves as a measure of the rate of proportional change of the function y. Just
as dy=dx measures the gradient of the curve, y D f .x/ plotted on a linear scale,
d ln y=dx measures the slope of the curve when plotted on a semi-logarithmic scale,
that is the rate of proportional change. For example, a value of 0.05 means that the
curve increases at 5% per unit x.
We can apply the same argument to the case when we plot a function on both x and
y logarithmic scales. In such a case, the following result is true:
d ln y
x dy
D
d ln x
y dx
85
Mass-action Kinetics
Computing the elasticities for mass-action kinetics is straight forward. For a reaction such as v D kS, we showed earlier (4.5) that "vS D 1. For a generalized
irreversible mass-action law such as:
Y n
vDk
Si i
the elasticity for species Si is ni . For simple mass-action kinetic reactions, the
kinetic order and elasticity are therefore identical and independent of species concentration.
For a simple reversible mass-action reaction rate law such as:
v D k1 S
(4.8)
k2 P
The elasticities for the substrate and product are given by:
"vS D
k1 S
vf
D
k1 S k2 P
v
"Pv D
k2 P
D
k1 S k2 P
(4.9)
vr
v
(4.10)
One very important point to note is that as the reaction approaches equilibrium, the
elasticities tend to plus infinity and negative infinity respectively.
Vm S
Km C S
86
A. Reaction velocity
Elasticity
B.
First-Order
0:8
@v
@S
0:6
@v S
@S v
0:4
Zero-Order
0:2
0
20
40
20
40
S
Figure 4.7 A. Left panel: the reaction velocity for an irreversible MichaelisMenten rate law as a function of substrate concentration. The curve is also marked
by the slope @v=@S. B. Right panel: the substrate elasticity is plotted as a function
of substrate concentration. Km D 4 and Vm D 1. Note the elasticity starts at one,
then decreases to zero as S increases.
"vS D
@v S
Km
D
@S v
Km C S
(4.11)
The substrate elasticity shows a range of values (Figure 4.7) from zero at high
substrate concentrations to one at low substrate concentrations. When the enzyme
is near saturation it is naturally unresponsive to further changes in substrate concentration, hence the elasticity is near zero. The reaction behaves as a zero-order
reaction at this point. When the elasticity is close to one at low S, the reaction
behaves with first-order kinetics. In addition, the reaction order changes depending
on the substrate concentration.
We can also express the elasticity in terms of the degree of saturation. Recall that
the degree of saturation is given by equation (??):
ES
S
D
E C ES
Km C S
A slight rearrangement of equation (4.11) allows us to write the elasticity in the
following form:
S
v
"vS D 1
D1
Km C S
Vm
87
so that:
"vS D 1
fractional saturation
(4.12)
"vE D 1
(4.13)
k2 P
88
1.
".a/ D 0
2.
".x/ D 1
3.
4.
".x a / D a
5.
6.
7.
f .x/
f .x/Cg.x/
".g.x//
g.x/
f .x/Cg.x/
".g.x//
k1 S
k1 S k2 P
"S .k2 P /
k2 P
k1 S k2 P
k1 S
k1 S k2 P
Vm S
Km C S
89
where Vm is the maximal velocity and Km the substrate concentration at half maximal velocity.
The elasticity for this equation can be derived by first using the quotient rule (rule
7) which gives:
"vS D ".Vm S/ ".Km C S/
The rules can now be applied to each of the sub-elasticity terms. For example we
can apply rule 6 to the first term, ".Vm S/, and rule 3 to the second term, ".Km CS/,
to yield:
S
Km
v
C ".S/
"S D .".Vm / C ".S//
".Km /
Km C S
Km C S
Applying rules 1 and 2 allows us to simplify (".Vm / D 0I ".Km/ D 0I ".S/ D 1)
the equation to:
S
v
"S D 1
Km C S
or
Km
"vS D
Km C S
Example 4.6
Determine the elasticity expression for the rate laws using log-log rules:
1. v D k.A C 1/ Begin with the product rule 6:
v
"A
D ".k/ C ".A C 1/ D ".A C 1/
v
"A
D ".A C 1/ D ".A/
".A C 1/ D
A
AC1
1
AC1
90
1
AC1
To make matters even simpler we can define the elasticity rules using an algebraic
manipulation tool such as Mathematica (http://www.wolfram.com/) to automatically derive the elasticities [68]. To do this we must first enter the rules in Table 4.1
into Mathematica. The script shown in Figure 4.8 shows the same rules (with a few
additional ones) in Mathematica format.
91
v1
v2
dSi =dt D
Si
P
Inflow
Outflows
Outflows
Figure 4.10 Mass Balance: The rate of change in species Si is equal to the difference between the sum of the inflows and the sum of the outflows
X
dSi
D
Inflows
dt
Outflows
(4.14)
92
For an even more general representation, we can write the mass-balance equations
using the stoichiometric coefficients. The rate at which a given reaction, vj contributes to change in a species, Si is given by the stoichiometric coefficient of the
species, Si with respect to the reaction, cij , multiplied by the reaction rate, vj . That
is, a reaction j contributes, cij vj rate to changes in species Si . For a species, Si
with multiple reactions producing and consuming Si , the mass-balance equation
(assuming constant volume conditions) is given by:
X
dSi
D
cij vj
dt
j
(4.15)
where cij is the stoichiometric coefficient for species i with respect to reaction,
j . For reactions that consume a species, the stoichiometric coefficient is often
negative. (See Enzyme Kinetics for Systems Biology, [43]). In considering the
simple example in Figure 4.9, the stoichiometric coefficient for S with respect to
v1 is C1 and for v2 is 1. That is
dS
D .C1/v1 C . 1/v2 D v1
dt
v2
The way in which the construction of the mass-balance equation is described may
seem overly formal, however the formality allows software to be written that can
automatically convert network diagrams into mass-balance differential equations.
Example 4.7
Consider a more complex linear chain of reactants from S1 to S5 shown in Figure 4.11.
Write out the mass-balance equations for this simple system.
S1
v1
S2
v2
S3
v3
S4
v4
S5
dS1
D
dt
93
v1
dS3
D v2
dt
v3
dS2
D v1
dt
v2
dS4
D v3
dt
v4
dS5
D v4
(4.16)
dt
Each species in the network is assigned a mass-balance equation which accounts for the
flows into and out of the species pool.
S1
v4
v3
S2
v5
v2
v3
v4
v5
Example 4.8
Write out the mass-balance equation for the more complex pathway:
ACX
X CY
Z
v1
! 2X
v2
! Z
v3
! Y CB
94
This example is more subtle because we must be careful to take into account the stoichiometry change between the reactant and product side in the first reaction (v1 ). In reaction v1 ,
the overall stoichiometry for X is C1 because two X molecules are made for every one
consumed. Taking this into account the rate of change of species X can be written as:
dX
D v1 C 2v1 v2
dt
v2 . The full set of mass-balance equations can therefore be written
or more simply as v1
as:
dA
D v1
dt
dY
D v3
dt
v2
dX
D v1
dt
v2
dZ
D v2
dt
v3
dB
D v3
dt
Example 4.9
Write out the mass-balance equation for pathway:
v1
S1 C S3 ! S2
v2
2S2 ! S3
v3
S3 ! 3S4
In this example we have non-unity stoichiometries in the second and third reaction steps.
The mass-balance equations are given by:
dS1
D
dt
v1
dS3
D
dt
v1 C v2
dS2
D v1
dt
v3
2v2
dS4
D 3v3
dt
It is therefore fairly straight forward to derive the balance equations from a visual
inspection of the network. Many software tools exist that will assist in this effort
by converting network diagrams, either represented visually on a computer screen
(for example, JDesigner) or provide a text file listing the reactions in the network
(for example via Jarnac).
95
cij
6 ::
N D Si 4 :
::
:
?
2
vj
3
::: :::
7
5
where cij is the stoichiometry coefficient for the i th species and j th reaction.
Example 4.10
Write out the stoichiometry matrix for the simple chain of reactions which has five molecular species and four reactions as shown below. The four reactions are labeled, v1 to v4 .
S1
v1
S2
v2
S3
v3
S4
v4
S5
96
v1
1
1
0
0
0
v2
0
1
1
0
0
v3
0
0
1
1
0
v4
0
0
0
1
1
S1
7 S2
7
7 S3
7
5 S4
S5
Example 4.11
Write out the stoichiometry matrix for the multibranched pathway shown in Figure 4.12
v1 v2
1
1
ND
0
0
v3
1
1
v4
0
1
v5
0
S1
1
S2
To illustrate that the stoichiometry matrix can be applied to other kinds of networks,
let us look at a simple signaling network and two simple gene regulatory networks.
Signaling Networks
Figure 4.13 illustrates a simple protein signaling network, comprising two double
phosphorylation cycles coupled by inhibition by protein C on the lower double
cycle (D; E and F ). In this model, all species are proteins and we can assume that
protein A and D are unphosphorylated, B and E singly phosphorylated and C and
F doubly phosphorylated. C acts as a kinase and phosphorylates D and E. The
reverse reactions, v2 ; v4 ; v7 and v8 are assumed to be catalyzed by phosphatases.
There is no specified stoichiometric mechanism for the inhibition on v5 and v6 .
Therefore the stoichiometric matrix will contain no information on this. The stoichiometric matrix for this system will look like:
A
B
C
ND
D
E
F
2
6
6
6
6
6
6
4
v1 v2 v3 v4 v5 v6 v7 v8
3
1
1
0
0
0
0
0
0
1
1
1
1
0
0
0
0 7
7
0
0
1
1
0
0
0
0 7
7
0
0
0
0
1
0
1
0 7
7
0
0
0
0
1
1
1
1 5
0
0
0
0
0
1
0
1
(4.17)
97
v1
v3
B
v2
C
v4
v5
v7
v6
v8
Figure 4.13 Simple Signaling Network. Protein C inhibits the activity of reactions
v5 and v6 .
The stoichiometric matrix can be seen to be composed of two separate blocks corresponding to the two cycle layers. It is important to note that whenever there are
regulatory interactions in a pathway diagram, these do not appear in the stoichiometry matrix. Instead, such information will reside in the rate law that describes
the regulation. If however the mechanism for the regulation is made explicit then
details of the regulation will appear in the stoichiometry matrix. Figure 4.14 will
shows an simple example of an inhibitor I regulating a reaction, S to P . On the left
is displayed the implicit regulatory interaction. All we see is a blunt ended arrow
indicating inhibition. In this case, details of the regulation will be found in the rate
law governing the conversion of S to P . On the right is displayed an explicit mechanism, a simple competitive inhibition. In this case all details of the mechanism will
find its way into the stoichiometry matrix. Figure 4.15 shows a comparison of the
implicit and explicit models in terms of the stoichiometry matrix.
98
I
S
k1
S+E
I
k5
Implicit Regulatory
Interaction
EI
k2
I
k4
ES
k3
E+P
Explicit Regulatory
Interaction
v
2 1 3
S
1
4
1 5
ND P
I
0
S
P
ND I
ES
EI
2
6
6
6
6
4
v1 v2 v3 v4 v5
3
1
1
0
1
1
0
0
1
0
0 7
7
0
0
0
1
1 7
7
1
1
1
0
0 5
0
0
0
1
1
Implicit
Explicit
Figure 4.15 Stoichiometry matrices corresponding to the two models in Figure 4.14
matrix is given below:
N D P1
v1
1
v2
1
(4.18)
The stoichiometry matrix has only one row indicating that there is only one species
in the model, P1 and there is no hint in the stoichiometry matrix that there is regulation.
Consider now that the interaction between P1 and v3 is made mechanistically explicit. The right hand network in Figure 4.16 shows one possible way in which to
represent the interaction of the transcription factor, P1 with gene v3 . In the explicit
model, the transcription factor, P1 is assumed to bind to a repressor site preventing
gene expression. In the explicit model there are two new species, designated active
gene and inactive gene. The stoichiometry matrix will therefore include two additional rows corresponding to these two species. The stoichiometry matrix for the
v1
99
Explicit or Mechanistic Model
v2
P1
v1
v3
v3
v2
P1
Active
v4f
v4r
P1
Inactive
Figure 4.16 Two simple gene regulatory networks involving gene repression. On
the left side is the implicit model where P1 represses v3 , on the right side is the
explicit model showing a more detailed mechanism for the regulation.
v2 v4r v4f
3
1
1
1
0
1
1 5
0
1
1
(4.19)
(4.20)
where N is the m n stoichiometry matrix and v is the n dimensional rate vector, whose i th component gives the rate of reaction i as a function of the species
concentrations. x is the m vector of species. The term p represents the vector of
parameters and inputs that can affect the reaction rate.
Looking again at the simple chain of reactions in Figure 4.11, the system equation
can be written down as:
2
3
2
3
1
0
0
0
v1
6 1
7
1
0
0 7 6
6
7
dx
7 6 v2 7
0
1
1
0
D Nv.x; p/ D 6
(4.21)
6
7
4 v3 5
dt
4 0
5
0
1
1
v4
0
0
0
1
100
If stoichiometry matrix is multiplied into the rate vector, the mass-balance equations show earlier (4.16) are recovered.
All stoichiometric interactions are placed in the stoichiometry matrix. The example
shown in Figure 4.13 and Figure 4.16 illustrated non-stoichiometric interactions,
namely two inhibition interactions from C to reactions v5 and v6 and repression on
v3 by P1 . As was noted, these interactions do no occur in the stoichiometry matrix.
Instead they will be found in the rate vector, x in the form of a particular rate law.
The stoichiometry matrix represents the mass transfer connectivity of the network
and contains information on the networks structural characteristics. These characteristics fall into two groups, relationships among the species and relationships
among the reaction rates. These relationships will be considered in a later chapter.
x
v2
v1
Figure 4.17 Single Gene Cassette with Gene Expression and Degradation.
Gene expression is induced by inducer I . Protein x is degraded at a rate v2 . This
simple unit is termed a gene cassette. We can write down the differential equation
that describes the rate of change of x as follows:
dx
D v1
dt
v2
In this equation we have not specified the specific rate laws that we will use for v1 or
v2 . These decisions will be based on specific biological knowledge. In this example
we can make the following assumptions. The first is that v2 is a first-order reaction
such that v2 D kx where k is a first-order rate constant. The second assumption is
that the rate of gene expression is a sigmoidal function of the inducer concentration.
The most commonly used sigmoid function in gene expression models is the Hill
101
equation. This is described in detail in the companion textbook Enzyme Kinetic for
Systems Biology [43]. For a positive inducer the Hill equation takes the form:
vD
Vm Indh
Kd C Indh
where Vm is the maximal expression rate, Ind the concentration of inducer and Kd
the dissociation constant for inducer binding to the operator site. h is called the Hill
coefficient and determines the sigmodicity of the response. Figure 4.18 shows how
the Hill equation behaves as a function of Inducer and different values for the Hill
coefficient.
1
hD8
Expression Rate
0:8
hD4
hD2
0:6
0:4
0:2
0
0:5
1
1:5
2
2:5
Inducer Concentration
Figure 4.18 Plot showing the response Hill equation with respect to inducer concentration and different Hill coefficients. Kd D 1.
As well as induction the Hill equation can also be used to model repression. This
requires only a simple modification where the induction term in the numerator is
removed:
Vm
vD
Kd C Indh
A more complex model is given in Figure 4.20 where we see a cascade of two gene
cassettes. This system has two proteins expressed, x1 and x2 . x1 expression is
induced by Xo and x2 expression induced by x1 . Both x1 and x2 are degraded. The
102
Expression Rate
0:8
0:6
0:4
0:2
hD2
hD4
hD8
0
0:5
1
1:5
2
2:5
Repressor Concentration
Figure 4.19 Plot showing the response Hill equation with respect to repressor
concentration and different Hill coefficients. Kd D 1.
differential equations for this model are given by:
dx1
D v1
dt
v2
dx2
D v3
dt
v4
If we assign first-order kinetics to the degradation steps and Hill equations to the
gene expression steps we obtain:
dx1
Vm1 Xoh
D
dt
Kd1 C Xoh
k1 x1
dx2
Vm2 x1h
D
dt
Kd2 C x1h
k2 x2
Further Reading
1. Bray D (2011) Wetware: A Computer in Every Living Cell. Yale University
Press. ISBN: 978-0300167849
Xo
x1
v1
v2
103
x2
v3
v4
Exercises
1. Explain the difference between the terms: Stoichiometric amount, Stoichiometric coefficient, rate of change (dX=dt) and reaction rate (vi ).
Two points for each one:
a) The stoichiometric amount is defined as the number of molecules of a particular reactant or product taking part in a reaction.
b) The species stoichiometry or stoichiometric coefficient however, is the difference between the stoichiometric amounts if a given species on the product
side and the stoichiometric amount of the same species on the reactant side.
c) The rate of change is defined as the rate of change in concentration or
amount of a designated molecular species
2. Determine the stoichiometric amount and stoichiometric coefficient for each
104
A
ACB
A
2A
3A C 4B
ACB
A C 2B
A ! 2B
v2
B ! 2C
v3
C !
Two points for each one:
dA
D
dt
v1
dB
D 2v1
dt
v2
dC
D 2v2
dt
v3
105
Derive the set of differential equations for the following model in terms of
the rate of reaction, v1 , v2 and v3 :
v1
A!B
v2
2B C C ! B C D
v3
D !C CA
Two points for each one:
dA
D v3
dt
v1
dB
D v1
dt
v2
dC
D v3
dt
v2
dD
D v2
dt
v3
4. Write out the stoichiometry matrix for the networks in question 3 and 4
Half point for each correct entry:
v1 v2 v3
3
A
1 0
0
1 0 5
ND B4 2
C
0
2
1
2
v1 v2 v3
3
A
1 0
1
B 6
1 0 7
6 1
7
ND
4
C
0
1 1 5
D
0
1
1
(4.22)
(4.23)
5. Derive the stoichiometry matrix for each of the following networks. In addition write out the mass-balance equations in each case.
(a)
106
v1
A
v2
v3
v1
v3
v4
B
v2
C
v4
(4.24)
(b)
v1
A
v2
v3
v1
v3
v4
B
v2
C
v4
3
0
1 5
1
(c)
v1
ACX !B CY
v3
B !C
DCY
X CW
v5
!X
v7
! 2Y
v2
B CX !Y
v4
C CX !DCY
v6
X !Y
2Y
v8
!X CW
(4.25)
107
A
X
B
ND Y
C
D
W
2
6
6
6
6
6
6
6
6
4
v1 v2 v3 v4 v5 v6 v7 v8
3
1 0
0
0
0
0
0
0
1
1 0
1 1
1
1 1 7
7
1
1
1 0
0
0
0
0 7
7
1
1
0
1
1 1
2
2 7
7
0
0
1
1 0
0
0
0 7
7
0
0
0
1
1 0
0
0 5
0
0
0
0
0
0
1 1
(4.26)
6. A gene G1 expresses a protein p1 at a rate v1 . p1 forms a tetramer (4 subunits), called p14 at a rate v2 . The tetramer negatively regulates a gene G2 . p1
degrades at a rate v3 . G2 expresses a protein, p2 at a rate v9 . p2 is cleaved
by an enzyme at a rate v4 to form two protein domains, p21 and p22 . p21 degrades at a rate v5 . Gene G3 expresses a protein, p3 at a rate v6 . p3 binds
to p22 forming an active complex, p4 at a rate v10 , which can activate G1 . p4
degrades at a rate v7 . Finally, p21 can form a complex, p5 , with p4 at a rate
v8 .
(a) Draw the network represented in the description given above.
One point for each correct interaction (14 in total):
v3
p1
v1
v6
4
p
v2
1
p5
v5
v8
p2
p1
2
p4
v7
v9
v11
p3
v10
v4
p 22
108
(b) Write out the differential equation for each protein species in the network in terms of v1 ; v2 ; : : :.
Two points for each correct entry:
dp1
dt
dp2
dt
dp22
dt
dp4
dt
D v1
v3
D v9
v4
D v4
v10
D v10
v8
dp14
dt
dp21
dt
dp3
dt
dp5
dt
4v2
v7
D v2
v6
D v4
v5
D v11
v8
v10
D v8
p1
p41 6
6
p2 6
6
p12 6
6
p22 6
6
p4 6
6
p5 4
p3
v1
1
0
0
0
0
0
0
0
v2
4
1
0
0
0
0
0
0
v3
1
0
0
0
0
0
0
0
v4
0
0
1
1
1
0
0
0
v5
0
0
0
1
0
0
0
0
v6
0
1
0
0
0
0
0
0
v7
0
0
0
0
0
1
0
0
v9
0
0
1
0
0
0
0
0
v8 v10 v11
3
0
0
0
0
0
0 7
7
0
0
0 7
7
1
0
0 7
7
0
1
0 7
7
1
1
0 7
7
1
0
0 5
0
1
1
7. Write out the differential equations for the system depicted in equation 4.21.
One point for each correct equation:
dS1
D v1
dt
v2
dS2
D v2
dt
v3
109
8. Build a differential equation computer model of the system shown in Figure 4.3. Assume the following rate laws for the four rates:
v1 D Vm1 Xoh =.K1 C Xoh /
v2 D K2 S1
v3 D Vm2 S1h =.K3 C S1h /
v4 D K4 S2
where h is the Hill coefficient, Vmx the maximal gene expression rate and Ki
various kinetic constants. Assume the following values:
K1 D 100I K2 D 0:6I K3 D 20I K4 D 2:5
Vm1 D 4I Vm2 D 15I h D 1
a) Plot the steady state values for S1 and S2 as a function of inducer molecule
Xo . Vary Xo between zero and 50.
Eight points for each correct graph:
b) Repeat question a) but set h D 8 and now vary Xo between zero and 3.0
Eight points for each correct graph:
110
5.2 Equilibrium
Thermodynamic equilibrium, or simply equilibrium, refers to the state of a system
when all energy gradients are dissipated. Thermodynamically it means that the
entropy of the system has reached its maximum and is constant. In chemistry,
111
112
thermodynamic equilibrium is when all forward and reverse rates are equal. This
also means that the concentration of chemical species is also unchanging and all
net flows are zero. Equilibrium is easily achieved in a closed system. For example,
consider the simple chemical isomerization reaction:
k1
(5.1)
S1
S2
k2
This system is closed because S1 and S2 cannot exchange mass with their surrounding. Let the net forward rate of the reaction, v, be equal to v D k1 S1 k2 S2 . The
rates of change of S1 and S2 are given by:
dS1
D
dt
dS2
Dv
dt
At equilibrium v D 0, therefore dS1 =dt and dS2 =dt both equal zero. The solution
to the differential equations can be derived as follows. Given that the system is
closed we know that the total mass in the system, S1 CS2 , is constant. This constant
is given by the sum of the initial concentrations of S1 and S2 which we will define
as S1o C S2o . We assume that the volume is constant and set to unit volume, this
allows us to state that the sum of the concentrations is conserved. The differential
equation for S1 is given by:
dS1
D k2 S2
dt
Let us replace S2 by the term S1o C S2o
dS1
D k2 S1o C k2 S2o
dt
k2 S1
k1 S1
S1 to yield:
k1 S1 D k2 .S1o C S2o /
S1 .k1 C k2 /
We now have an equation only in terms of S1 and a set of parameters. The easiest
way to solve this equation is to use Mathematica or Maxima. The Mathematica
command is
DSolve[S1'[t]==k2 (S1o + S2o) - S1[t] (k1 + k2), S1[0]==S1o, S1[t], t]
where S1[0] == S1o sets the initial condition for the concentration of S1 to be S1o .
By implication, the initial condition for S2o is .S1o C S2o / S1o D S2o . The result of
applying Mathematica yields the following solution:
S1 .t/ D
.k1 Ck2 /t
vinitial
k1 C k2
The first term in the solution is the equilibrium concentration of S1 . The second
term is an adjustment to the equilibrium concentration over time as a result of the
5.2. EQUILIBRIUM
113
non-equilibrium initial conditions and which tends to zero at infinite time. In later
chapters we will see this kind of structure appear again and again and is characteristic of linear systems.
Concentration, S1 or S2
At t D 0, the second term is equal to vinitial =.k1 C k2 / where vinitial is the net
reaction rate of the reversible reaction at the initial conditions, that is k1 S1o k2 S2o .
At t D 0, the value of the second term must be the difference between the initial
concentration S1o and the equilibrium concentration, S1eq . The second term also
has an exponential component which approaches zero as time goes to infinity. At
this point we are left with the first term which equals the concentration of S1 when
dS1 =dt D dS2 =dt D 0.
10
8
S2
6
4
S1
2
0
0:5
1:5
2
Time, t
2:5
k1
114
From this somewhat long winded analysis, it has been determined for the closed
reversible system, at infinite time, the concentrations of S1 and S2 reach some
constant values and that the net rate, v, is zero. The system is therefore at thermodynamic equilibrium.
In biochemical models it is often assumed that when the forward and reverse rates
for a particular reaction are very fast compared to the surrounding reactions that
the reaction is assumed to be in quasi-equilibrium. That is although the entire
system may be out of equilibrium there may be parts of the system that can be approximated as though they were in equilibrium. This is often done to simplify the
modeling process. Living organisms are not themselves at thermodynamic equilibrium, if they were then they would technically be dead. Living systems are open so
that there is a continual flow of mass and energy across the systems boundaries.
5.3 Transients
The second behavior that a system can show is a transient. A transient is usually
the change that occurs in the species concentrations as the system moves from one
state, often a steady state, to another. Equation 5.3 shows the solution to a simple
system (5.2) that describes the transient behavior of species S1 and S2 . Figure 5.2
illustrates the transient from an initial non-steady state condition to a steady state.
115
production of the system is a constant zero or even negative. Note that the entropy
of the surrounding will be positive such that the net production of entropy in the
system and environment is positive. This later requirement satisfies the second law.
Thermodynamic Equilibrium and Steady State
Thermodynamic equilibrium and the steady state are distinct states of a chemical system. In equilibrium, both the rate of change of species and the net flow
of mass through the system is zero. That is:
ds
D0
dt
for all i : vi D 0
where vi is the net reaction rate for the i th reaction step. At equilibrium there
is therefore no dissipation of gradients or energy fields. When a biological
system is at equilibrium, we say it is dead.
The steady state has some similarities with equilibrium except there is a net
flow through the system such that gradients and energy fields are continuously
being dissipated. This also means that one or more vi s must be non-zero
The steady state is defined when all dSi =dt are equal to zero while one or
more reaction rates are non-zero:
ds
D0
dt
vi 0
To convert the simple reversible model described in the last section into an open
system we need only add a source reaction and a sink reaction as shown in the
following scheme:
ko
k1
k3
Xo ! S1
S2 !
k2
(5.2)
In this case simple mass-action kinetics is assumed for all reactions. It is also assumed that the source reaction, with rate constant, ko , is irreversible and originates
from a boundary species, Xo , that is Xo is fixed. By this mean that although Xo
is being constantly consumed there is some outside mechanism that maintains Xo
constant. The rate of reaction of the source reaction (ko ) is therefore fixed at ko Xo
which we will call vo .
116
In addition it is assumed that the sink reaction, with rate constant, k3 is also irreversible. For the purpose of making it easier to derive the time course solution, the
reverse rate constant, k2 will be assumed to equal zero and we will set the initial
conditions for S1 and S2 to both equal zero. The mathematical solution for the
system under these conditions is given by equation 5.3:
S1 .t/ D vo
S2 .t/ D vo
e
k1
k1 1
k1 t
k3 t
C k3 e
k3 .k1 k3 /
k1 t
(5.3)
1
Note that vo D ko Xo . As t tends to infinity S1 .t/ tends to vo =k1 and S2 .t/ tends
to vo =k3 . The reaction rate through each of the three reaction steps is vo . This can
be confirmed by substituting the solutions for S1 and S2 into the reaction rate laws.
Given that vo is greater than zero and that S1 and S2 reach constant values given
sufficient time, we conclude that this system eventually settles to a steady state
rather than thermodynamic equilibrium. The system displays a continuous flow of
mass from the source to the sink. This can only continue undisturbed so long as the
source material, Xo never runs out and that reaction k3 is irreversible. Figure 5.2
shows a simulation of this system.
S1
S1 and S2
0:4
S2
0:2
2
Time
ko
k1
Figure 5.2 Time course for an open system reaching steady state. Xo ! S1 !
k3
117
Graphical Procedure
We can also illustrate the steady state using a graphical procedure. Consider the
simple model below:
Xo
v1
S1
v2
X1
where Xo and X1 are fixed boundary species and S1 is a species that can change
(also called a floating species). For illustration purposes we will assume some
very simple kinetics for each reaction, v1 and v2 . Let us assume that each reaction
is governed by simple first order mass-action kinetics,
v1 D k1 Xo
v2 D k2 S1
where k1 and k2 are both first-order reaction rate constants. In Figure 5.3 both
reaction rates have been plotted as a function of the floating species concentration,
S1 .
1:5
k2 D 0:3
1
v1 D v2
k2 D 0:1
0:5
k2 D 0:2
1
2
3
4
Substrate Concentration, S1
Figure 5.3 Plot of reaction rates versus concentration of S1 and different values
for k2 for the system Xo ! S1 ! X1 . The intersection of the two lines marks the
steady state point where v1 D v2 . Xo D 1; k1 D 0:4. Note that as k2 is decreased
the steady state level of S1 increases.
Note that the reaction rate for v1 is a horizontal line because it is unaffected by
changes in S1 (no product inhibition). The second reaction, v2 is shown as a straight
118
line with slope, k2 . Notice that the lines intersect. The intersection marks the point
when both rates, v1 and v2 are equal. This point marks the steady state concentration of S1 . By varying the value of k2 we can observe the effect it has on the steady
state. For example, Figure 5.3 shows that as we decrease k2 the concentration of S1
increases. This should not be difficult to understand, as k2 decreases, the activity
of reaction v2 also decreases. This causes S1 to build up in response.
In this simple model it is also straightforward to determine the steady state of S1
mathematically which amounts to finding a mathematical equation that represents
the intersection point of the two lines. We recall that the model for this system
comprises a single differential equation:
dS1
D k1 Xo
dt
k2 S1
At steady state, we set dS1 =dt D 0, from which we can solve for the steady state
concentration of S1 as:
k1 Xo
(5.4)
S1 D
k2
This solution tells us that the steady state concentration of S1 is a function of all
the parameters in the system. We can also determine the steady state rate, usually
called the pathway flux and denoted by J, by inserting the steady state value of S1
into one of the rate laws, for example into v2 :
J D k2
k1 Xo
D k1 Xo
k2
This answer is identical to v1 which is not surprising since in this model the pathway flux is completely determined by the first step and the second step has no
influence whatsoever on the flux. This simple example illustrates a rate limiting
step in the pathway, that is one step, and one step only, that has complete control
over the pathway flux.
Xo
v1
S1
v2
S2
v3
X1
119
where the rate law for the first step is now reversible and is given by:
v1 D k1 Xo
k2 S1
The remaining steps are governed by simple irreversible mass-action rate laws,
v2 D k3 S1 and v3 D k4 S2 . The differential equations for this system are:
dS1
D .k1 Xo k2 S1 /
dt
dS2
D k3 S1 k4 S2
dt
k3 S1
The steady state solution for S1 and S2 can be obtained by setting both differential
equations to zero to yield:
S1 D
k1 Xo
k2 C k3
S2 D
k3 k1 Xo
.k2 C k3 /k4
The steady state flux, J , can be determined by inserting one of the solutions into
the appropriate rate law, the easiest is to insert the steady state level of S2 into v3
to yield:
k3 k1 Xo
J D
k2 C k3
Once the first step is reversible we see that the steady state flux is a function of all
the parameters except k4 indicating that the first step is no longer the rate limiting
step. The equation shows us that the ability to influence the flux is shared between
the first and second steps. There is no rate limiting step in this pathway. Note that
if we set k2 D 0 then the solution reverts to the earlier simpler model.
We can also make all three steps reversible (kf Si
given by:
S1 D
Xo k1 .k4 C k5 / C X1 k4 k6
k3 k5 C k2 .k4 C k5 /
S2 D
X1 k6 .k2 C k3 / C Xo k1 k3
k3 k5 C k2 .k4 C k5 /
120
rate laws as used, such as Hill equations or Michaelis-Menten type rate laws, the
solutions become exceedingly difficult to derive. As a result steady states tend to
be computed numerically rather than analytically.
(5.5)
where x is the unknown and p one or more parameters in the equations. All numerical methods for computing solutions to equation 5.5 start with an initial estimate
for the solution, say xo . The methods are then applied iteratively until the estimate
converges to the solution. One of the most well known methods for solving equation 5.5 is called the Newton-Raphson method. It can be easily explained using
a geometric argument, Figure 5.4. Suppose x1 is the initial guess for the solution
to equation 5.5. The method begins by estimating the slope of equation 5.5 at the
value x1 , that is df =dx. A line is then drawn from the point (x1 ; f .x1 /), with slope
df =dx until it intersects the x axis. The intersection, x2 , becomes the next guess
for the method. This procedure is repeated until xi is sufficiently close to the solution. For brevity the parameter p is omitted from the following equations. From
the geometry shown in Figure 5.4 one can write down the slope of the line, @f =@x1
as:
f .x1 /
@f
D
@x1
x1 x2
This can be generalized to:
f .xk /
@f
D
@xk
xk xxC1
121
y=f(x)
Starting Point
Solution
df/dx
x3 x2
x1
x
Figure 5.4 The geometry of Newton-Raphsons method
or by rearrangement:
xkC1 D xk
f .xk /
@f =@xk
(5.6)
122
Iteration
Estimate
0
1
2
3
4
5
15
8.33333
5.666
5.0392
5.0001525
5.0
Table 5.1 Newton method used to compute the square root of 25, using equation 5.7 with a starting value of 15.
nature of the system being solved. In order to prevent the method from continuing
without end in the case when convergence fails if is often useful to halt the method
after a maximum of iterations, say one hundred iterations. In a case like this, a
new initial start is given and the method tried again. In biochemical models we an
always run a time course simulation for a short while and use the end point of that
as the starting point for the Newton method. This approach is much more reliable.
If the method does converge to a solution there are various ways to decide whether
convergence has been achieved. Two such tests include:
1. Difference between Successive Solutions Estimates. We can test for the difference between solution xi and the next estimate, xi C1 , if the absolute difference, jxi xi C1 j is below some threshold then we assume convergence
has been achieved. Alternatively we can check whether the relative error is
less than a certain threshold (say, 1%). The relative error is given by
D
xi C1 xi
100%
xi C1
The procedure can be made to stop at the i -th step if jf .xi /j < f for a given
f .
2. Difference between Successive dSi =dt Estimates. Here we estimate the
rates of change as the iteration proceeds and assume convergence has been
archived when the difference between two successive rates of change are below some threshold. If we are dealing with a model will more than one state
variable then we can construct the sums of squares of the rates of change:
X dSi 2
dt
123
The Newton method can be easily extended to systems of equations so that we can
write the Newton method in matrix form:
x kC1 D x k
@f .x/
@x
f .x k /
(5.8)
124
Jacobian Matrix
The Jacobian is a common matrix used in many fields especially control theory
and dynamical systems theory. We will frequently use it in this book. Given a
set of equations:
y1 D f1 .x1 ; : : : ; xn /
y2 D f2 .x1 ; : : : ; xn /
::
:
ym D fm .x1 ; : : : ; xn /
The Jacobian matrix is defined as the matrix of partial differentials:
2
3
@f1
@f1
6 @x1 : : : @xn 7
6 :
:: 7
::
:
J D6
:
:
: 7
6
7
4 @fm
@fm 5
:::
@x1
@xn
Although the Newton method is seductively simple, it requires the initial guess to be
sufficiently close to the solution in order for it to converge. In addition convergence
can be slow or not occur at all. A common problem is that the method can overshoot
the solution and will then being to rapidly diverge.
A further strategy that is frequently used to compute the steady state is to first use a
short time course simulation to bring the initial estimate closer to the steady state.
The assumption here is that the steady state is stable. The final point computed in
the time course is used to seed a Newton like method, if the Newton method fails to
converge then a second time course simulation is carried out. This can be repeated
as many times as desired. If there is a suspicion that the steady state is unstable,
one can also run a time course simulation backwards in time. In general there is
no sure way of computing the steady state automatically and sometimes human
intervention is required to supply good initial estimates.
As a result of these issues the unmodified Newton method is rarely used in practice
for computing the steady state of biochemical models. One common variant, called
the Damped Newton method is sometimes used. Both Gepasi and SCAMP use the
Damped Newton method for computing the steady state. This method controls the
derivative, df =dx by multiplying the derivative by a factor , 0 < < 1 and can
125
be used to prevent overshoot. There are many variants on the basic Newton method
and good simulation software will usually have reasonable methods for estimating
the steady state.
In the last ten years more refined Newton like methods have been devised and
one that is highly recommended is NLEQ2 (http://www.zib.de/en/numerik/
software/ant/nleq2.html. This is used by both Jarnac, libRoadRunner and
PySCeS for computing the steady state. The stiff solver suite sundials (https://
computation.llnl.gov/casc/sundials/main.html also incorporates an equation solver, however experience has shown that is it not quite as good as NLEQ2.
Xo
v1
S1
v2
S2
v3
X1
k2 S1 /
.k3 S1
k4 S2 /
(5.9)
k4 S2 /
.k5 S2
k6 X1 /
The values for the rate constants and the boundary conditions are given in Table 5.2.
This is a problem with more than one variable (S1 and S2 ) which means we must
use the Newton-Raphson matrix form (5.8) to estimate the steady state. To use this
we require two vectors, x k and f .x k / and one matrix, @f .x/=@x. The x k vector
is simply:
S
xk D 1
S2
The f .x k / vector is given by the values of the differential equations:
.k1 Xo k2 S1 / .k3 S1 k4 S2 /
f .x k / D
.k3 S1 k4 S2 / .k5 S2 k6 X1 /
126
Parameter
Value
k1
k2
k3
k4
k5
k6
Xo
X1
3.4
0.2
2.3
0.56
5.6
0.12
10
0
dS2
Notice that the elements of the Jacobian contain only rate constants. This is because the model we are using is linear. This also means we need only evaluate the
Jacobian and its inverse once. If we used nonlinear rate laws such as the MichaelisMenten rate law, the Jacobian matrix would also contain terms involving the species
concentrations and in this case the Jacobian would need to be reevaluated at each iteration because the value for the species concentration will change at each iteration.
For the current problem the Jacobian and its inverse is given by:
2:86 5:6
Jacobian D
0:56
11:2
Jacobian
D
0:3876
0:01938
0:1938
0:09898
Table 5.3 shows the progress of the iteration as we apply equation 5.8. What is
interesting is that convergence only takes one iteration. This is because the model
127
is linear. Nonlinear models may require more iterations. We can also see that after
the first iteration the rates of change have very small values, this is usually due to
very small numerical errors in the computer arithmetic but anything as small as
10 14 can be considered zero.
Iteration S1
S1
0
1
1
36.74
0.6589 2:8 10
1
13.18
dS1 =dt
dS2 =dt
14
-10.64
1:16 10
13
Table 5.3 Newton-Raphson applied to a Three Step Pathway with Linear Kinetics.
Starting values for S1 and S2 are both set at one. Convergence occurs within one
iteration. Note that the values for the rates of change are extremely small at the end
of the first iteration, indicating we have converged.
''')
Initialize value
= 10; X1 = 0;
= 3.4; k2 = 0.2;
= 2.3; k3 = 0.56;
= 5.6; k6 = 0.12;
128
Running the above script yields steady state concentrations of 13.1783 and 0.658915
for S1 and S2 respectively, which is the same if we compare these values to those in
Table 5.3. Other tools will have other ways to compute the steady state, for example graphical interfaces will generally have a button marked steady state that when
selected will compute the steady state for the currently loaded model.
When using Matlab the function fsolve can be use to solve systems of nonlinear
equation and in Mathematica one would use FindRoot.
Concentration of S1
0:6
0:4
S1 approaching steady state
0:2
10 12 14 16 18 20
Time
129
Concentration of S1
1
0:8
Perturbation in S1
0:6
0:4
Perturbation in S1
0:2
0
10
20
30
Time
40
50
60
Figure 5.6 Multiple Perturbations. The steady state concentration of the species
S1 is 0.5 and a perturbation is made to S1 by adding an additional 0.35 units of
S1 at time D 20 and removing 0.35 units at time D 40. In both cases the system
relaxes back. Python script: 5.5
130
Concentration of S1
1
0:8
0:6
k1 Restored to original value
0:4
Perturbation in k1
0:2
0
10
20
30
40 50
Time
60
70
80
Figure 5.7 Effect of Perturbing Model Parameters using the Python script ??.
131
Xo
S1
X1
Figure 5.8 illustrates the results from a simulation of a simple two step biochemical
pathway with one floating species, S1 . The Python script used to generate this
graph is given in Table ??. The initial concentrations of the model are set so that
it is at steady state, that is no transients are seen between t D 0 and t D 20. A
t D 20 a perturbation is made to the concentration of S1 by adding 0.25 units of
S1 . This could be accomplished by injecting 0.25 units of S1 into the volume where
the pathway resides. The system is now allowed to evolve further. If the system is
stable, the perturbation will relax back to the original steady state, as it does in the
simulation shown in Figure 5.8. This system therefore appears to be stable.
132
Concentration of S1
1
0:8
0:6
S1 Relaxes back
0:4
Perturbation in S1
0:2
0
10
20
30
40
50
Time
Figure 5.8 Stability of a simple biochemical pathway at steady state. The steady
state concentration of the species S1 is 0.5. A perturbation is made to S1 by adding
an additional 0.25 units of S1 at time D 20. The system is considered stable because
the perturbation relaxes back to the original steady state. See Table ?? for Python
listing.
The differential equation for the single floating species, S1 , is given by
dS1
D k1 Xo
dt
k2 S1
(5.10)
with a steady state solution S1 D k1 Xo=k2 . The question we wish to ask here is
whether the steady state is stable or not? We can show that the two step model
is stable be using the following mathematical argument. If the system is at steady
state, let us make a small perturbation to the steady state concentration of S1 , S1
and ask what is the rate of change of S1 C S1 as a result of this perturbation, that is
what is d.S1 C S1 /=dt? The new rate of change equation is rewritten as follows:
d.S1 C S1 /
D k1 Xo
dt
k2 .S1 C S1 /
If we insert the solution for S1 (equation 5.4) into the above equation we are left
with:
d S1
D k2 S1
(5.11)
dt
In other words the rate of change of the disturbance, S1 is negative, that is, the
system attempts to reduce the disturbance so that the system returns back to the
original steady state. Systems with this kind of behavior are called stable. If the
133
rate of change in S1 had been positive instead of negative however, the perturbation
would have continued to diverge away from the original steady state and the system
would them be considered unstable.
Dividing both sides of equation 5.11 by S1 and taking the limit S ! 0, we
find that @.dS1 =dt/=@S1 is equal to k2 . The stability of this simple system can
therefore be determined by inspecting the sign of @.dS1 =dt/=@S1 which can be
easily determined by taking the derivative of the differential equations with respect
to the species concentrations. For larger systems the stability of a system can be
determined by looking at all the terms @.dSi =dt/=@Si which are given collectively
by the expression:
d.d s=dt/
DJ
(5.12)
ds
where J is called the Jacobian matrix containing elements of the form @.dSi =dt/=@Si .
Equation 5.11 can also be written as
d.s/
D J s
dt
(5.13)
This is a set of linear differential equations that describes the rate of change in the
perturbation S. J is given by
2
3
@S1 =dt
@S1 =dt
6
7
6 @S1
7
@S
m
6
7
:
:
:
6
7
::
::
::
J D6
(5.14)
7
6
7
4 @Sm =dt
@Sm =dt 5
@S1
@Sm
Equation 5.13 is an example of an homogeneous linear differential equation and
has the general form:
dx
D Ax
dt
As we will see in a later chapter, solutions to such equations are well known and
take the form:
n
X
xj .t/ D
j k e k t
kD1
That solution involves the sum of exponentials, e k t . The exponents of the exponentials are given by the eigenvalues of the matrix, A, namely the Jacobian matrix 5.14. If the eigenvalues are negative then the exponents decay (stable) whereas
134
if they are positive the exponents grow (unstable). We can therefore determine the
stability properties of a given model by computing the eigenvalues of the Jacobian
matrix and looking for any positive eigenvalues. Note that the elements of the Jacobian matrix will often be a function of the species levels, it is therefore important
that the Jacobian be evaluated at the steady state of interest.
There are many software packages that will compute the eigenvalues of a matrix
and there are a small number packages that can compute the Jacobian directly from
the biochemical model. For example, the script below is taken from Python, it
defines a simple model, initializes the model values, computes the steady state and
then prints out the eigenvalues of the Jacobian matrix. For a simple one variable
model, the Jacobian matrix only has a single entry and the eigenvalue corresponds
to that entry. The output from running the script is given below showing that the
eigenvalue is 0:3. Sine we have a negative eigenvalue, the pathway must be stable
to perturbations in S1 .
import tellurium as te
rr = te.loada ('''
$Xo -> S1; k1*Xo;
S1 -> $X1; k2*S1;
// Set up the model initial conditions
Xo = 1; X1 = 0;
k1 = 0.2; k2 = 0.3;
''')
# Evaluation of the steady state
rr.getSteadyStateValues()
# print the eigenvalues of the Jacobian matrix
print print rr.getEigenvalues()
# Output follows:
[[-0.3 0. ]]
135
V
p
where V is the variable and p the parameter. This equation shows finite changes
to the parameter and variable. Unfortunately because most systems are nonlinear,
the value for the sensitivity will be a function of the size of the finite change. To
make the sensitivity independent of the size of the change, the sensitivity is usually
defined in terms of infinitesimal changes:
SD
dV
dp
Although absolute sensitivities are simple they have one drawback namely that the
value can be influenced by the units used to measure the variable and parameter. Often in making experimental measurements we wont be able to measure the
quantity using the most natural units, instead we may have measurements in terms
of fluorescence, colony counts, staining on a gel and so on. It is most likely that
the variable and parameter units will be quite different and each laboratory may
have its own way particular units is uses. Absolute sensitivities are therefore quite
difficult to compare.
136
To get round the problem of units, many people will use relative sensitivities, These
are simple scaled absolute sensitivities:
SD
dV p
dp V
The sensitivity is defined in terms of infinitesimal changes for the same reason
cited before. The reader may also recall that elasticities are measured in this way.
Relative sensitivities are immune from the units we use to measure quantities but
also relative sensitivities correspond more closely to how many measurements are
made, often in terms of relative or fold changes. In practice steady state relative
sensitivities should be measured by taking a measurement at the operating steady
state, making a perturbation (preferable a small one), waiting for the system to
reach a new steady state then measuring the system again. It is important to be
aware that the steady state sensitivities measure how a perturbation in a parameter
moves the system from one steady state to another.
Further Reading
1. Sauro HM (2011) Enzyme Kinetics for Systems Biology. ISBN: 978-0982477311
2. Kipp E, Herwig R, Kowald A, Wierling C and Lehrach H (2005) Systems
Biology in Practice, Wiley-VCH Verlag
3. Tellurium/Python web site http://tellurium.analogmachine.org/
Exercises
1. Consider the following simple branched network:
v2
v1
S1
v3
where v1 D vo ; v2 D k1 S1 and v3 D k2 S1 .
137
138
10. Construct a simple linear pathway with four enzymes as shown below:
v1
Xo
S1
v2
S2
v3
S3
v4
X1
Assume that the edge metabolites, Xo and X1 , are fixed. Assign reversible
Michaelis-Menten kinetics to each step and arbitrary values to the kinetics
constants. Assign a modest value to the boundary metabolite, Xo , of 10
mM. Compute the steady state for your pathway. If the software fails to
find a steady state, adjust the parameters. Once you have the steady state, use
the model to compute the sensitivity of the steady state flux with respect to
each of the enzyme maximal activities. You can compute each sensitivity by
perturbing each maximal activity and observing what this does to the steady
state flux.
How might you use the flux sensitivities in a practical application? Compute
the sum of the four sensitivities, what value do you get? Can you make a
statement about the sum?
Appendix
See http://tellurium.analogmachine.org for more details of Tellurium.
import tellurium as te
# Simulation of a simple closed system
rr = te.loada ('''
A -> B; k1 * A;
B -> A; k2 * B;
''')
A = 10; B = 0;
k1 = 1; k2 = 0.5;
import tellurium as te
# Simulation of an open system
rr = te.loada ('''
$Xo -> S1; vo;
S1 -> S2; k1*S1 - k2*S2;
S2 -> $X1; k3*S2;
''')
vo = 1
k1 = 2; k2 = 0; k3 = 3;
139
140
Xo
S1
k1
k2
''')
=
=
=
=
1;
0.5;
0.2;
0.4;
''')
Xo
S1
k1
k2
=
=
=
=
1;
0.0;
0.2;
0.4;
''')
Xo
S1
k1
k2
=
=
=
=
1;
0.5;
0.2;
0.4;
141
142
rr.model.k1 = 0.2
# Carry out final run of the simulation
m3 = rr.simulate(50, 80, 40, ["time", "S1"])
# Merge all data sets and plot
result = numpy.vstack((m1, m2, m3))
pylab.ylim([0,1])
te.plotWithLegend(rr, result)
''')
Xo
S1
k1
k2
=
=
=
=
1;
0.5;
0.2;
0.4;
Inputs to a System
Its designed to disrupt your input/output carrier signal.
The Matrix, 1999
In this chapter we are going to look at the different kinds of inputs that can be applied to a system. In control engineering such inputs are also called signals, forcing
functions or driving functions. Inputs to a system can be applied in a variety of different ways some of which are critical to a proper understanding of how the system
operates. In particular we are going to look at the step function, the impulse, the
ramp and the sinusoidal input. Along the way we will also introduce shifting or
time delays in signals. In the following text, t refers to time.
Figure 10.2 illustrates the main types of inputs that are commonly found in engineering with perhaps the step function being one of the most common. Increasing
the expression of a gene by changing the inducer concentration is an example of a
step function.
Example 6.1
What type of signal is represented in the following situations:
a) The concentration of glucose entering a chemostat is increased from 5 mM to 8 mM and
remains at 8 mM.
Step Input
b) A drug is very rapidly injected into a patients arm
143
144
u(t)
u(t)
0
Unit Step Function
u(t)
u(t)
a+1
u(t)
1
t
Ramp Function
u(t)
u(t)
Delayed Ramp
t
Unit Impulse
Sinusoidal Input
145
following way:
(
u.t/ D
0 t <0
1 t 0
Scaling
For steps other than unity, such as a step size of , we can write the step function
as:
uT .t/ D u.t/
u.t/
0
t
0
t
Figure 6.2 Unit Step Function, u.t/ left panel. Scaled Unit Step Function, 2u.t/
right panel.
Examples of step functions in biology are common. Increasing the expression of a
particular protein or increasing the concentration of a drug in a cell culture can all
be modeled as a step function.
146
u.t
a/
2
1:5
1
0:5
0
a
4
0
t
2/
a/u.t
a/
For example (See Figure 6.20), to shift the function, x 2 , 2 time units into the future,
we would write:
.x 2/2 u.t 2/
147
4
3
2
1
0
t
Figure 6.4 Shifted (delayed) function, x 2 using .x
2/2 u.t
2/
Example 6.2
Sketch the graphs for the following functions (t is the independent variable (x-axis)):
1. y D t C 1
2. y D u.t
3/
3. y D .t C 1/u.t
4. y D ..t
3/
3/ C 1/u.t
3/
1.
2.
3.
4.
6
4
2
0
4
t
6 0
4
t
6 0
4
t
6 0
Figure 6.5 Unit Step Function, u.t/ left panel. Scaled Unit Step Function, 2u.t/
right panel.
148
1/ C u.t
2/
To illustrate how addition works let use plot each signal separately (Figure 6.6, left
panel). The right panel in the same figure shows the effect of summing the two
individual terms. Signals can be added or subtracted where subtraction can be a
useful way to generate pulses.
y D u.t
3
u.t
1/
2/
1
u.t
1/ C u.t
2/
4
4
t
Figure 6.6 Individual terms left panel. Sum of terms right panel.
Pulses
If a drug is administered via an intravenous line for a set period we can often model
the input of the drug into the blood stream as a pulse with a given width. We can
define pulse signals by combining step functions and delays. For example consider
a pulse that starts at t D a and stops at a later time t D b. Such a situation can be
represented as follows:
pulse.a; b/ D u.t
a/
u.t
b/ where a < b
That is, at time a the unit step function rises to 1.0. At this point u.t b/ is still
zero. At t D b, u.t b/ goes to one but it subtracts this value from u.t a/ thereby
reducing the signal back to zero. Figure 6.7 illustrates such a response.
149
2
1:5
1
0:5
0
1/
u.t
t
Figure 6.7 Shifted Unit Pulse Signal, u.t
2/
Pulse Train
Once we can define a single pulse it is possible to describe a train of pulses. Figure 6.8 shows two pulses which are constructed by summing two separate pulses.
We can generalize this to n pulses using the equation:
f .t/ D
n h
X
. 1/
kC1
i
d k/
u.t
kD1
1/
u.t
2/ C u.t
t
Figure 6.8 Shifted Unit Pulse Signal, u.t
3/
u.t
4/
150
1:5
1:5
0:5
0:5
2
t
Figure 6.9 Left panel: Unit Ramp Signal: tu.t/. Right panel: shift unit ramp
signal: t u.t 1/
A shifted ramp, where the y value starts at y D 0 (not half way up as it did in
the last example), can be created with the following expression (Figure 6.10: Left
151
Panel):
k.tu.t
2/
2u.t
2//
The first term describes a ramp starting at t D 2, with slope k and a y value of 2 k.
In order for the ramp to start at y D 0, we must translate the ramp to the x-axis, To
do this we just subtract 2 k starting at t D 2.
To create a ramp that rises then stops (Figure 6.10: Right Panel) we can use the
expression:
k.tu.t/
.t
a// D k.t
a/u.t
2/u.t
2/
The first term describes a ramp starting at time, t D 0 with slope k. At time a,
the same slope is now subtracted from the original rising slope which gives us a
horizontal line. The term u.t a/ ensures that the negative slope, .t a/, is only
subtracted at time a. Think of the term .t a/ i.e . t C a/ as a line with negative
unit slope and a y intercept of a. Compare this to the generalized shifting function,
f .t a/u.t a/, we encountered earlier. In this case the function is simply kt, so
that the shifted function is k.t a/u.t a/.
3
2
t
Figure 6.10 Left Panel: Ramp Function with a slope of 1 that starts at time 2:
t u.t 2/ 2u.t 2/. Right Panel: Ramp Function with a slope of 1 that stops at
time 2: t u.t / .t 2/u.t 2/.
Class Question 1
Describe how a sawtooth signal might be represented.
152
The first expression states that the value of the impulse at all values other than zero
is zero. The second expression states that the area of an impulse is one. This arises
by first considering a pulse where the height of a pulse is 1=T and its width T . As
the width of the pulse is reduced, the area of the pulse remains one. The width can
be reduced to a value as small as we like (Figure 6.11), but still the area will remain
at one.
The description of an impulse can be further generalized to any time:, a:
a/ D 0 for
.t
Z
t a
.t
a/dt D 1
The strength of an impulse is defined by its area and we can change the area of an
impulse such that:
Z
1
K.t/dt D K
1
An impulse with a given strength is often indicated using the notation K.t/ where
K is the strength. Dont consider the K.t/ notation as a product, it is not. It is
only a convenient notation. Such an expression only makes sense under integration.
153
T=0.1
10
T=0.25
T=0.5
T=1
1/T
1
Figure 6.11 Pulses converging to an impulse while maintaining the area fixed at
1.0: T 1=T
As suggested in the definition, like other signals the impulse can be delayed so that
we can specify that a pulse should occur at a delayed time a use the notation:
.t
a/
a/ C .t
b/ C : : :
.t
kT /
kD0
where T is the period of the pulse train and n the number of pulses.
154
2
1:5
1
0:5
0
t
Figure 6.12 Impulse Signal, .t/
2
1:5
1
0:5
0
t
Figure 6.13 Depiction of an Impulse Signal, K.t/, with strength, K
Sifting
Probably one of the most important properties of an impulse is its ability to pick out
a value from a given function, called sifting. Consider a continuous function, f .t/
that we multiply by .t/. This product will be zero except for the point marked by
the impulse. At that point, f .t/ D f .0/. The following two representations are
therefore equivalent:
f .t/.t/ $ f .0/.t/
This can be generalized to
f .t/.t
a/ $ f .a/.t
a/
On their own, these relations are not very useful but become much more interesting
when we integrate both sides. The left and right-hand sides can now be treated as
155
2
1:5
1
0:5
0
t
Figure 6.14 A series of Impulse signals, .t
equal to each other with defined values:
Z 1
Z
f .t/ .t a/dt D
1
k/
f .a/ .t
a/dt
D f .a/
.t
a/dt
.t
a/dt D 1
then
Z
f .t/ .t
a/dt D f .a/
(6.1)
This result tells us that the area under the product of a function, f .t/ with a unit
impulse is equal to the value of that function at the point where the unit impulse
is located. In other words the impulse function plucks out or sifts out the function
value at the point of the impulse. The function, f .t/ must be continuous where the
impulse is located.
156
derivative at the step is essentially infinite. Since the point at which the step occurs
is infinitesimally small, the derivative of the step is reminiscent of an impulse. We
can think of the relationship between a unit step and impulse as:
du.t/
D .t/
dt
Summary
.t
Z
a/ D 0
for
t a
.t
a/dt D 1
f .t/ .t
a/dt D f .a/
6.6 Sinusoidal
The sine function is another signal that finds common using in engineering. The
sinusoidal signal is the most fundamental periodic signal. Any other periodic signal
can be constructed from a summation of sinusoidal signals. We can describe a
sinusoidal signal by an equation of the general form:
y.t/ D A sin.!t C /
The equation has three terms, the amplitude, A, the angular frequency, !, and the
phase, . We will describe the phase in more detail below but essentially it indicates how delayed or advanced and periodic signal may be. The angular frequency
is the rate of the periodic signal and is expressed in radians per second. A sine
wave traverses a full cycle (peak to peak) in 2 radians (circumference of a circle)
so that the number of complete cycles traversed in one second is then !=2. This
is termed the frequency, f , (cycles sec 1 ) of a sine wave and has units of Hertz.
The angular frequency is then conveniently expressed as ! D 2f .
6.6. SINUSOIDAL
157
Example 6.3
A sine wave has a frequency of 10 cycles per second. What is its angular frequency and
period of the wave?
The angular frequency is given by ! D 2f . Therefore ! D 2 10 D 62:83 radians per
second.
The period of the wave is 1=f , therefore the period will be 1=10 D 0:1 seconds.
The amplitude is the extent to which the periodic function changes in the y direction, that is the maximum height of the curve from the origin (Figure 6.17). The
horizontal distance peak to peak is referred to as the period, T and is usually expressed in seconds. The inverse of the period, 1=T sec 1 is equal to the frequency,
f.
2
Period (T )
y.t/
1
Amplitude (A)
=2
3=2
2
90
180
270
360
5=2
1
450
2
Figure 6.15 Basic Sine Wave: y.t/ D A sin.t/
Figure 6.19 shows a typical plot of the sinusoidal function, y.t/ D A sin.!t C /,
where the amplitude is set to one, the frequency to 2 cycles per second and the
phase to zero.
158
y.t/
0:5
0
0:5
1
y.t/
The left panel in Figure 6.17 shows the affect of varying the amplitude and right
panel two signals of different frequency.
2
0:5
Time (t)
1:5
0:5
1:5
Time (t)
6.6. SINUSOIDAL
159
y.t/
1
0
1
2
0:5
1
Time (t)
1:5
y.t /
0:5
0
0:5
1
0:5
1:5
Time (t)
Figure 6.19 A masked Sine Function, u.t 0:5/ sin.4 t/. Note that a frequency
of 4 ensures that a full cycle starts every 0.5.
Shifting a sine function can be accomplished using the expression:
sin.t
a/u.t
a/
160
2
1
0
1
2
0:5
1:5
2
t
2:5
3:5
2/2 u.t
2/
other sinusoidal with a different phase and amplitude but identical frequency. That
is:
A1 sin.!t C 1 / C A2 sin.!t C 2 / D A3 sin.!t C 3 /
where
A3 D
and
A1 sin 1 C A2 sin 2
3 D arctan
A1 cos 1 C A2 cos 2
(6.2)
2 /
6.7 Exponential
Although we may not think of common situations in biology where exponentials
are used as inputs to a system, their theoretical importance is crucial to the study of
systems. The common or real exponential is given by:
y D Ae ax
where A and a are constants. Typical plots of the common exponential are given
in Figure 6.21. The constant A determines the value of y at t D 0. The constant a
determines whether the exponential grows (a > 0) or decays (a < 0) and the degree
6.7. EXPONENTIAL
161
to which it does. In the science and engineering fields the exponential function is
also expressed as:
y D Ae x=
where 1= is called the time constant.
20
10
5
t
(6.3)
x2
D e x1 =e x2
What about the exponential of a complex number such as i!, that is e i ! ? How can
we interpret such an expression? If we substitute x in the exponential Taylor series
for i ! and assume for now that the series converges, then we obtain:
e i ! D 1 C i! C
i! 3
!4
i! 5 ! 6 i! 7
C
C
C
3
4
5
6
7
!2
!4 !6
!8
!3
!5
C
C
C i !
C
2
4
6
8
3
5
D 1 C i!
D 1
.i!/2
.i!/3
.i!/4
.i!/5
.i!/6
.i!/7
C
C
C
C
C
C
2
3
4
5
6
7
!2
2
D cos ! C i sin ! :
!7
C
7
162
Since both trigonometric series converge, this is therefore a legitimate proof. The
final result is called Eulers theorem and shows us that the complex exponential is
related to the trigonometric functions and therefore has periodic properties.
Eulers Theorem:
e i ! D cos ! C i sin !
We can generalize and consider the complex exponential: e . Cj!/t . This is called
the complex exponential function
Complex Exponential Function:
y D Ae .Ci !/t D Ae st
where s D C i!
The symbol s is commonly used to indicate the complex number in an exponential.
If the complex term i! is zero then the complex exponential reduces to the real
exponential. Although we know what a plot of a real exponential looks like, how
does the complex exponential behave?
We can write the complex exponential as follows (assuming A D 1):
y D e st D e .Ci !/t D e t e i !t
Using Eulers theorem e i ! D cos ! C i sin !:
y D e t .cos !t C i sin !t/
If for the moment we set D 0 then we can look at the complex number:
y D cos !t C i sin !t
This function has two sinusoidal components, one in the real domain and the other
in the complex domain. At t D 0, only the real part as a value at 1.0, the imaginary
part is zero. As we increase t, both real and imaginary part oscillate, the real part
as a sine wave and the imaginary part as a cosine wave. We can plot a 3D graph
that shows how the real and imaginary parts behave as a function of t. This plot is
shown in Figure 6.22.
If we now add back the real part of the complex exponential, that is:
y D e t .cos !t C i sin !t/
6.7. EXPONENTIAL
163
1
1
1
1
0:05Ci !/t
De
0:05t iw t
when we have is a spiral behavior that can either expand or contract (Figure 6.23).
The exponential function encompasses a wide range of of behaviors. Given the
expression y D Ae .Ci !/t , these include:
Constant: D 0I ! D 0, y = A
Monotonic exponential: 0I ! D 0
Sinusoidal: D 0I ! 0
Exponentially varying sinusoidal: 0I ! 0
A purely real sinusoid can be expressed as the sum of two complex sinusoids. Using
Eulers formula:
e i ! D cos ! C i sin !
e
i!
D cos !
i sin !
164
ei ! C e
2
i!
ei !
i!
sin ! D
e
2i
Exercises
1. Write out the signal equation for a ramp with a slope k, that starts rising at 5
seconds, lasts for 6, then stops increasing.
2. Write out the signal equation for a ramp with slope k that lasts for T seconds
then falls abruptly to zero.
3. Sketch the following signals:
(a) u.t /
u.t
5/
(b) e t u.t/
(c) .t / C .t
3/
(d) 8t u.t/
16.t
1/u.t
(e) 2t u.t/
2.t
12/u.t
1/ C 8.t
2/u.t
2/
12/
4. Write out a signal that will behave as a sawtooth. Assume the sawtooth starts
at t=0, with height 1.0 and lasts 2.0 t units.
5. Using the sifting property of an impulse, compute the sifted value for e
t D 0. Show your working.
at
6. Using the sifting property of an impulse, compute the sifted value for sin.t/
at t D 1:1. Show your working.
7. Show that by differentiating the exponential series 9.4 that the derivative of
e x is e x .
8. The frequency of a sine wave is 5 cycles per second. Compute the angular
frequency, ! and the period, T .
9. The peak to peak voltage of a sinusoidal signal is 5 volts, what it the amplitude of the signal?
6.8. APPENDIX
165
=2/.
6.8 Appendix
Proof for Equation 6.2. Show that:
A1 sin.!t C 1 / C A2 sin.!t C 2 / D A sin.!t C /
We being with the known identity:
sin. C / D sin cos C cos sin
(6.4)
166
Y D A sin
Y
X
2 /
2 /
A1 sin 1 C A2 sin 2
D arctan
A1 cos 1 C A2 cos 2
We conclude that:
A1 sin.!t C 1 / C A2 sin.!t C 2 / D A sin.!t C /
(6.5)
Linear Systems
Linear systems are important because we can solve them.
Richard Feynman, The Feynman Lectures on Physics
Most of the phenomena in nature are the result of process governed by nonlinear
laws. So why the emphasis on linear systems? The answer is simple, there is no
general theory for dealing with nonlinear systems. Whether there ever will be is
a matter of speculation but our inability to rationalize nonlinear systems is a huge
hinderance to our understanding of truly complex systems such as biological cells
but particularly the brain. All is not lost however, the last three hundred years or
has seen the development of a complete understanding of linear systems. Given
our inability to deal effectively with nonlinear systems but our complete mastery of
linear systems many studies in science and engineering focus on systems that are
linear.
Engineers are particularly fond of linear systems. Many of the devices that engineers build are by design linear for the very reason that they can be understood.
Unfortunately we dont have that luxury when studying biological systems. One
way to study nonlinear systems is to simulate it on a digital or analog computer.
However, anyone who has tried this will realize attempting to actually understand
what is going on in the system simply from time course plots knows how difficult
this is. As an adjunct to simulation, many bioengineers will linearize a nonlinear
system and study its linear properties. Between computer simulation and lineariz-
167
168
ing a system, we can get a long way to understanding how a particular system operates. In this chapter we will focus on linear systems, what they are, what properties
they have and when confronted with a nonlinear system how to use linearization to
make the system amenable to analytical study.
169
Additivity
If:
x1
System
y1
x2
System
y2
x1 + x 2
System
y1 + y2
then:
Homogeneity
If:
x1
System
y1
x 1
System
y 1
then:
y2 D mx2
If we now apply both inputs simultaneously we see that additivity is also obeyed.
y3 D m.x1 C x2 / D y1 C y2
Therefore the equation is linear.
170
Example 7.2
Consider the simple equation, y D x 2 . The homogeneity test for this equation fails because
.x/2 D 2 x 2 D 2 y y
y D x 2 is therefore a non-linear equation.
Non-linear equations are fairly easy to spot, any equation where the state variables is raised
to a non-unity power or systems that involve the state variables of trigonometric functions
are generally non-linear.
xn
xy
ex
.dy=dy/n
x
sin.x/
log.x/
Vm S=.S C Km /
Class Question 1
Show that the function e x is nonlinear.
We first form sum of apply separate inputs, that is:
e x1 C e x2
We now apply the sum simultaneously, that is:
e x1 Cx2
To obey additivity, these two expressions much be equal. However, e x1 Cx2 D
e x1 e x2 which is not the same as e x1 C e x2 . Therefore e x is a nonlinear function.
Similarly we can also easily show that homogeneity is not true because it should be
evident that:
ae x e ax
7.2. TIME-INVARIANCE
171
7.2 Time-Invariance
Another property of a system that is often included as part of linearity is timeinvariance. Systems that are linear and time-invariant are called Linear TimeInvariant Systems or LTIs. Such systems are important to engineers because the
response of a LTI system can be completely characterized by the response of the
system to an impulse or step response. We will came back to this important topic
in another chapter.
A time-invariant system is one where the system isnt an explicit function of time.
This means that if a system is subjected to a perturbation at time, t0 and experiences
a response at time t1 , then applying the same perturbation t0 at a time t0 C t will
elicit the same response but now at time t1 C t . That is the system response is
simply shifted in time with no other qualitative difference in the actual response.
In a computer simulation it would mean that we could start the simulation clock
at any time, 0; t1 ; t2 ; : : : etc and we would get the same time course solutions. For
these reasons, time-invariance is also sometimes called shift-invariance.
172
x.t /
H1
y.t/
x.t
/
H1
y.t
/
Example 7.4
Is the following system time-invariant?
y.t / D tx.t /
We first input a delayed x.t /, that is x.t a/, to generate y1 .t /. Replace every occurrence
of x.t/ with x.t a/.
y1 .t / D t .x.t a//
Next we generate a time shifted output to generate y2 .t /. Replace every occurrence of t
with .t a/:
y2 .t / D .t a/.x.t a//
We see that y1 .t/ y2 .t /, therefore the system is time varying.
Example 7.5
Is the following system time-invariant?
y.t / D e t .x.t //
7.3. LINEARIZATION
173
We first input a delayed x.t /, that is x.t a/, to generate y1 .t /. Replace every occurrence
of x.t/ with x.t a/.
y1 .t / D e t .x.t a//
Next we generate a time shifted output to generate y2 .t /. Replace every occurrence of t
with .t a/:
y2 .t / D e .t a/ .x.t a//
We see that y1 .t / y2 .t /, therefore the system is time varying.
Spotting a time variant system is often a case of observing if the system equations
are an explicit function of time. Linearity and time-invariance are important because the dynamics of systems which agree with these assumptions can be fully
described.
7.3 Linearization
To linearize means replacing the nonlinear version of the equation with a linear approximation. Such approximations are only valid when subjected to small changes
around the point of linearization, also called the operating point. It should be emphasized that in the process we loose valuable information but enough information
is retained to make linearization an extremely useful tool.
To appreciate the difference between linear and nonlinear functions, consider the
system y D x 2 . Let us apply two separate inputs, x1 and x2 to give outputs x12
and x22 . If we now apply the inputs simultaneously, that is y D .x1 C x2 /2 , we
obtain x12 C x22 C 2x1 x2 . We see that the output is not simply x12 C x22 but includes
an additional term, 2x1 x2 . This term is the nonlinear contribution. Imagine that
this difference now enters further nonlinear processes, leading to further changes.
Eventually the output looks nothing like the input. This small change makes most
nonlinear systems difficult to understand and analyze.
Unless the system is degenerate or has the trivial solution (that is the solution is
zero), linear systems generally admit only one solution. In contrast it is possible
for nonlinear systems to admit multiple solutions, that is given a single input, a nonlinear system can regurgitate one of a number of possible distinct outputs. To makes
matters worse, in the majority of mathematical models we find in biochemical networks, it is not even impossible to find the solutions mathematically. That is we
cannot actually describe mathematically how an output depends on an input other
than by doing brute-force computer simulation. Understanding nonlinear systems
whether we find them in biology or elsewhere is a huge unresolved problem.
174
Whereas there is a complete theory of linear systems, no such thing exists for nonlinear systems. When dealing with nonlinear systems we are often forced to use
computer simulation.
There is one approach we can take that can help use deal with nonlinear models. If
we were to draw a nonlinear curve on a graph and zoom into the graph the curve
would eventually look like a straight line; that is we can turn a nonlinear system
into a linear one but only in small regions of the systems behavior where linearity
dominates. This process, is called linearization and is a powerful technique for
studying nonlinear systems.
Tangent line
yo
f(x)
x
xo
xo /
(7.1)
7.3. LINEARIZATION
175
xo /
2xo2 D 2xo x
xo2
16
(7.2)
x5
x3
C
3
5
Without going into the details (See Appendix B) the ci terms in the polynomial
series (7.2) can be determined such that the series will approximate any given continuous function. The Taylor series is always defined around some operating point,
xo and is expressed in terms of a distance from that operating point, x xo . The
Taylor series is given by:
f .x/ D f .xo / C
@f
.x
@x
xo / C
1 @2 f
.x
2 @x 2
xo /2
C ::: C
1 @n f
.x
n @x n
xo /n C : : : (7.3)
176
2
Figure 7.4 Linearized sin.t/ function represented by the straight line through zero.
The various derivatives in the Taylor series must be evaluated at xo . The the function f .x/ must be continuous in order to be able to compute the derivative, that
is there should be no holes or sudden breaks (discontinuity) in the curve that the
function describes. The number of terms in the Taylor series determines the degree
of approximation, the fewer terms the more approximate the series is. For example,
the most approximate expression is given by using only the first term, f .xo /. However, f .xo / is a constant so this represents a very poor approximation. To make the
approximation more useful we can include the first two terms of the Taylor series:
f .x/ f .xo / C
@f
.x
@xo
(7.4)
xo /
This should be recognized from the tangent line equation shown previously, equation 7.1. Note that the derivative must be computed at the operating point, xo .
Provided x is close to xo , the approximation is good. For example, let us form the
Taylor series for the function, y D sin.x/ around xo D 0. We should recall that
sin.0/ D 0 and cos.0/ D 1, then:
y D sin.0/ C
@ sin.xo /
.x
@x
y D 0 C 1x C 0
0/ C
1 @2 sin.x/
.x
2 @x 2
0/ C : : :
1 3
1
x C 0 C x5 C : : :
3
5
The linear approximation is simply, y D x (Figure 7.4). We have effectively linearized the sin function.
To illustrate linearization with another example consider the simple nonlinear function, y D x 2 . To linearize we must first choose an operating point around which
7.3. LINEARIZATION
177
we will linearize, for example, xo D 2. According to the second term in the Taylor
series we need to find the derivative, df =dx so that the first two terms of the Taylor
series (Equation 7.4) become:
f .x/ D f .2/ C 2xo .x
2/
To obtain the linear approximation we evaluate the derivative at the operating point
(xo D 2), that is df =dxo D 2xo D 4 so that the final linear approximation is then
given by:
f .x/ D 4 4.x 2/ D 4x 4
Figure 7.5 shows the original nonlinear function together with the linear approximation.
40
y D x2
30
20
10
y D 4x
0
0
3
x
df
x
dxo
(7.5)
@f
@f
x C
y
@x
@y
178
xss / C
xo and u D u
@f .xo ; uo /
.u
@u
uo /
dx
d x
D
dt
dt
Note that xo is independent of time. We can therefore write the Taylor series approximation as:
@f .xo ; uo /
@f .xo ; uo /
d x
x C
u
dt
@x
@u
The important point to note about this result is that the Taylor approximation gives
us an equation that describes the rate of change of a perturbation in x, that is x.
Let us replace the partial derivative by the symbols A and B so that:
d x
Ax C Bu
dt
Around the steady state equilibrium point, the linearization no longer has an offset
as we saw in equation 7.5, f .xo /. This is an important point to remember in later
chapters when we linearize systems around the steady state.
7.3. LINEARIZATION
179
xo / C
@F .xo ; yo /
.y
@yo
yo /
dy
@G.xo ; yo /
G.xo ; yo / C
.x
dt
@xo
xo / C
@G.xo ; yo /
.y
@yo
yo /
xo / C
@F .xo ; yo /
.y
@yo
yo /
dy
@G.xo ; yo /
.x
dt
@xo
xo / C
@G.xo ; yo /
.y
@yo
yo /
3
@F 2
x
@yo 7
74
@G 5 x
@yo
xo
yo
3
5
180
If the differential equations are also a function of one or more inputs (u1 : : : um )
then the derivation can be extended such that:
3
32
2
o
2 3 2 @F @F 3
u
u
1
@F
@F
1
dx
2
3
:::
6
7
x
x
o
6
7
7
6
o
o
6 dt 7 6 @xo @yo 7
7
@um 7 6
6 @u1
6 76
6
4
5
C6
76 ::: 7
7
4 dy 5 4 @G @G 7
5 x y
4 @G
@G 5 4
5
o
:
:
:
o
o
o
@u1
@um
@xo @yo
dt
um um
where uoi is the nominal value for the ith input. If we represent the Jacobian by
the matrix symbol A, the matrix related to the input as B (often called the control
matrix), and the change in state variable and inputs by x and u, then:
d
@F i
@F i
x
x C
u
dt
@xj
@uj
(7.6)
d
x Ax C Bu
dt
(7.7)
This equation tells us how a disturbance x changes in time given some initial
disturbance to the state x and a disturbance to the system inputs u.
Example 7.6
Linearize the following system of differential equations around the steate state point (xo D
0; yo D 0), with given state x; y and input u:
dx
Du
dt
x2
dy
D x2
dt
Compute the Jacobian, A with respect to the state and the control matrix, B with respect
to the input, u.
AD
2x
2x
0
1
1
BD
0
2x
2x
0
1
x
1
C
u
y
0
7.3. LINEARIZATION
181
Exercises
1. Show whether the following relationships are linear or nonlinear with respect
to x:
(a) y D x 2
(b) y D
(c)
dy
dt
dx
dt
D ay C x
(d) y D mx C b
2. Show algebraically that the following equation is time varying:
y.t / D tx.t/
Show graphically that the equation is time varying.
3. Show whether the following systems are time-invariant or varying:
(a) y.t/ D Ax.t/ C B
(b) y.t/ D cos.3t/x.t/
(c) y.t/ D x.t
1/ C 2tx.t
2/
x at x D 4
p
(b) y D 3 x at x D 8
(a) y D
(c) y D 2 ln.x/ at x D 1
(d) f .x; y/ D y 2 e xCy at x D 2I y D 1
5. Show that the following differential equation is linear:
dx
D
dt
4.x
2/
dx
D x 2 C 10
dt
182
(b)
dx
D x 3 C 2x
dt
y2
dy
D sin.y/
dt
4x
x1, x2 , .... xn
u2(t)
State Variables
p 1, p 2 ,.... p q
ua(t)
Parameters
System Inputs
y 1(t)
y 2(t)
ym(t)
Observed Variables
183
184
times one can only indirectly access the state variable via other kinds of measured
quantities (yi ).
To given a concrete example of a system, consider a simple genetic circuit that
expresses green fluorescence protein in response to changes in an inducer molecule.
Let the state variable of the system be the concentration of expressed protein, this
would be a xi type quantity in Figure 8.1. However we cant actually observe the
protein directly, only as a fluorescent measurement in an instrument. The fluoresce
we actually observe in the experiment is represented by a yi variable in Figure 8.1.
Finally, the inducer that we can apply to the system to change the level of gene
expression is an external parameter, ui .
In the following notation, lower case bold will be used to represent vectors and
upper case bold will be used to represent matrices.
Let us represent the vector of state variables, x, by:
x.t / D x1 .t /; x2 .t/; : : : ; xn .t/T
the vector of inputs, u, by
u.t / D u1 .t /; u2 .t/; : : : ; up .t/T
and the vector of output variables by:
y.t / D y1 .t /; y2 .t/; : : : ; ym .t/T .
In general the relationship between the state variables, xi and the input parameters,
ui is nonlinear and can be written down as:
dx
D f .x.t/; u.t//
dt
where x and u are vectors. Likewise the vector of observables, y is related to the
state variables and parameters by some nonlinear function g:
y.t/ D g.x.t/; u.t//
x, u and y are in general functions of time. For brevity, we will omit the .t/ in the
following discussion.
It is common to linearize these nonlinear system equations around the equilibrium
or steady state point. If we let xo and uo be the values at steady state then by
definition it must be the case that:
0 D f .xo ; uo /
Let us now make a small disturbance around uo such that:
u D uo C u
185
and
x D xo C x
We now expand f .x.t/; u.t// around xo and uo using a Taylor series to obtain:
@f
dx
D f .xo ; uo / C
.x
dt
@xo
We now define:
xo / C
@f
.u
@uo
uo / C : : :
@f
@xo
@f
BD
@uo
AD
Since x D x
xo and xo is a constant:
dx
d x
D
dt
dt
xo and u D u
uo we obtain:
d
x D Ax C Bu
dt
This is a linear time independent equation that describes the rate of change of the
perturbation in x around the steady state as a result of perturbations in x and/or u.
The function g can also be linearized to obtain:
y D C x C Du
The matrices A; B; C and D and constant and do not depend on time or the state
variables. In the control literature, these matrices are given the following labels:
A
B
C
D
nn
np
mn
mp
If the C matrix is the identity matrix and the D matrix is zero, then the state variable
vector is equal to the output vector, y. The linearized state equations are often
represented in the literature by equation (8.1). For many systems the D matrix
is empty. A block diagram representing of equations (8.1) is given in Figure 8.2
which shows why the D matrix is called the feed-forward matrix.
186
D
.
u(t)
-1
y(t)
dx
D Ax.t/ C Bu.t/
dt
(8.1)
As we will discover in the next section, these equations are both linear and timeinvariant and are called the state space equations.
Written out, the state space equation for x would look like:
2 3 2
32 3 2
32 3
x1
a11 a12 : : : a1n
x1
b11 b12 : : : b1p
u1
6x2 7 6a21 a22 : : : a2n 7 6x2 7 6b21 b12 : : : b2p 7 6u2 7
6 7 6
76 7 6
76 7
6 :: 7 D 6 ::
:: 7 6 :: 7 C 6 ::
:: 7 6 :: 7
4:5 4 :
: 54 : 5 4 :
: 54 : 5
xn
an1 an2 : : : ann
xn
bn1 bn2 : : : bnp
up
Note that the A matrix is always square. The A matrix is also called the Jacobian
matrix. Note that the number of inputs, u is entirely at the discretion of the model
builder. It is possible for example to have no inputs, one input or many inputs. The
easiest way to construct the A and B matrices is to differentiate the differential
equations with respect to the state variables for the A matrix and the inputs for the
B matrix.
Each entry in the A is given by:
A ij D
dxi =dt
dxj
(8.2)
187
The A matrix in the state space representation is called the Jacobian Matrix,
where:
dxi =dt
A ij D
dxj
if fi D dxi =dt then:
3
@f1
@f1
:
:
:
6 dx1
dxn 7
6 :
:: 7
:
6
::
A D 6 ::
: 7
7
4 @fm
@fm 5
:::
dx1
dxn
2
dxi =dt
duj
(8.3)
188
d x1 C x2
D A.x1 C x2 / C B.u1 C u2 /
dt
To show additivity we now apply the sum of the inputs, u1 C u2 . Let us propose
that the response to this input is given by x1 C x2 . The derivative of the response
with respect to time is:
d x1 C x2
dt
However this equals A.x1 C x2 / C B.u1 C u2 / for when the inputs are applied
separately, therefore x1 C x2 is a solution when u1 C u2 is applied. This tells us
that the system shows additivity. We can generalize this by considering u1 and
u2 inputs. This shows that the system satisfies both additivity and homogeneity,
therefore the system is linear.
K
M
x1
Distance
Equilibrium
Point
Friction, f
Figure 8.3 A mass, M, is attached to spring (K) which in turn is attached to a fixed
point. The mass rests on a surface with friction, f .
Let us designate the distance moved relative to the equilibrium point as x1 . If the
system moves to the left of the equilibrium point we assume x1 < 0 otherwise we
assume x1 > 0. At equilibrium however, x1 D 0. If we apply an impulse, .t/, to
the mass, it will begin to oscillate as it attempts to return to the equilibrium point. If
189
we include a frictional force on the table then the movement of the mass will slow
down until it resets again at the equilibrium point. We will assume that the velocity
of the mass is given by x2 . This means that the acceleration will be given by the
rate of change of the velocity:
dx2
aD
dt
Our aim is to write down two differential equations, one for the velocity and another
for the acceleration. To do this we must consider all the forces that contribute to the
dynamics in this system. There are three forces involved, the force we apply when
we displace the mass, the force that is exerted by the spring and the force exerted
by the friction on the table. The total forces, F t , on the mass is therefore:
F t D Fimpulse C Fspring C Ffriction
First consider the spring. Hookes law states that the displacement experienced by a
spring from its equilibrium point is propositional to the applied force, F . Moreover
because the force exerted by the spring opposes the direction of the displacement
we also include a negative sign in the relationship. That is:
x/
where k is called the spring constant. That is, Fs D kx. The next force to
consider is the frictional force. Friction always opposes motion and studies indicate
that the frictional force is proportional to the rate at which the mass moves, in other
words:
dx1
dv
D f
D f x2
Ff D f
dt
dt
The force relationship is negative because friction opposes movement. The final
force to consider is the displacement we apply to the mass at t D 0. This will be
an impulse such that at t < 0 the system is at equilibrium but at t D 0 we apply
an impulse, .t/ that causes the system to respond. We can therefore write that the
total force on the mass is:
Ftotal D .t/
kx1
f x2
From Newtons second law we know that the force on a mass is equal to the acceleration times the mass of the object. That is:
dx2
M D .t/
dt
or
.t/
dx2
D
dt
M
kx1
k
x1
M
f x2
f
x2
M
190
This equation combined with the velocity equation gives us the complete system of
equations for this system:
dx1
D x2
dt
dx2
.t/
D
dt
M
k
x1
M
f
x2
M
In this case it should be evident how to construct the A and B matrices by inspection. However they can also be construct by differentiation (See equations 8.2
and 8.3). This would be a necessity for nonlinear models. We can rewrite these
equations in the state space representation as :
3
2
2
32 3 2 3
dx1
0
1
0 h
x
6 dt 7
i
7 6
6
7 4 15 6 7
C 4 1 5 .t/
7D4 k
6
f 5
4 dx2 5
x2
M
M
M
dt
Note that the impulse function is separated out from the rest of the model because
it represents the input to the system.
Electrical Model
eL
L
iL
u(t)
ec
eo
ic
Figure 8.4 Electrical circuit comprising a fixed current course, u.t/ together with
a capacitor (C ), inductor (L) and resistor (R). i is the current and e the voltage.
Consider the electrical circuit shown in Figure 8.4. Let us note the following basic
191
relationships:
eo D RiL
dec
dt
diL
eL D L
dt
ic D C
ec D eL C eo
u.t/ D ic C iL
The circuit can be described by two differential equations, one describing the rate
of change of the voltage across the capacitor and another that describes the rate of
change of current through the inductor. From the capacitance law we can write:
dec
ic
D
dt
C
but since ic D u.t/
iL :
dec
u.t/ iL
D
(8.4)
dt
C
C
This provides one of the equations we need. Given that ec D eL C eo the inductor
law can be rewritten as:
L
or
d iL
D ec
dt
eo D ec
d iL
ec
D
dt
L
which yields the second equation.
Ri
iL
L
RiL
(8.5)
We will designate the capacitor voltage (ec , equation 8.4) by x1 and the inductor
current (iL , equation 8.5) by x2 . We also note that the input, u.t/ acts on the
capacitor, C , therefore we can rewrite state space equations as:
2
3 2
3
dx1
1 2 3 213
6 dt 7 6 0
h i
7 x1
6
7 6
6 7
C
7
4
5
C 4C 5 u
6
7D
R5
4 dx2 5 4 1
x2
0
L
L
dt
192
Dosage
Absorption
x1
u(t)
k1
Blood
k4
x2
k2
Urine
x4
k3
x3
Tissues
Pharmokinetic Model
Consider the pharmokinetic system shown in Figure 8.5.
We will assume first-order kinetics for each of the processes. The dosage will be
assumed to be a step response so that:
dx1
D u.t/
dt
k1 x1
k2 x2
dx3
D k2 x2
dt
k3 x3
dx4
D k4 x2
dt
k4 x2 C k3 x3
193
2 3 2 3
3 x1
0 6 7 617
76 7 6 7
76 7
07 h i
07 6x2 7 6
7
76 7 C 6
7 u
76 7 6
0
7
07 6x3 7 6
6
56 7 4 7
5
4 5
0
0
x4
dt
Biochemical Model
Consider the simple biochemical pathway shown below:
v1
v2
v3
X0 ! S1 ! S2 ! X1
where we assume that Xo and X1 are fixed species. Assume that the rate laws are
given by:
v1 D E1 .Xo k1
S1 k2 /
v2 D E2 .S1 k3
S2 k4 /
v3 D E3 k5 S2
where Ei is the concentration of enzyme i in reaction i . We will build the state
equations around the steady state operating point. We could have used nonlinear
Michaelis-Menten rate laws but we can make the algebra simpler by assuming simple linear rate laws. The differential equations for S1 , and S2 are:
dS1
D v1
dt
v2
dS2
D v2
dt
v3
194
differential equations with respect to the state variables for the A matrix and the
inputs for the B matrix. For these equations, A can then be derived as:
"
#
.E1 k2 C E2 k3 /
E2 k4
AD
E2 k3
.E2 k4 C E3 k5 /
B is derived by assuming E1 ; E2 and E3 are the inputs which leads to a 2 by 3
matrix:
#
"
k1 Xo k2 S1
.k3 S1 C k3 S2 /
0
BD
0
k2 S1 k4 S2
k5 S2
The uo vector is E1 E2 E3 T . Fully written out the state space equation looks like:
3
2
dS1
"
#2 3
S1
6 dt 7
.E1 k2 C E2 k3 /
E2 k4
6
7
4
5
D
6
7
4 dS2 5
E2 k 3
.E2 k4 C E3 k5 /
S2
dt
"
C
k1 Xo
.k3 S1 C k3 S2 /
k2 S1
0
k2 S1
k4 S2
2 3
E1
#6 7
0
6 7
6E2 7
7
k5 S2 6
4 5
E3
One important point to bear in mind about the above equation. Even though matrix
B matrix appears to be a function of S1 and S2 , this is not strictly true. The values
for S1 and S2 must be taken from their values at the operating point (steady state).
As a result, the B matrix contains only constant values.
Exercises
1. Consider the following two species biochemical pathway:
vo
! x1
k1 y1
! x2
k2 x 2
k1 x1
k 2 x2
195
Write out the state space representation of this model. You can assume that
the term, vo , is the input to the system and that the operating point is the
steady state. Assume the C matrix is the identity matrix and the D matrix is
zero.
2. Write out the state space model for the gene cascade show in Figure 8.6.
Assume the input to the model is Xo and that the state variables are S1 and
S2 . Assume the C matrix is the identity matrix and the D matrix is zero.
The operating point for the system will be the steady state. The individual
rate laws are given by:
v1 D Vm1
Xo
Km1 C Xo
v2 D k1 S1
v3 D Vm2
S1
Km2 C S1
v4 D k2 S2
Xo
S1
v1
v2
S2
v3
v4
196
(9.1)
So long as the terms A and B are constant and not time dependent, this is a linear
time invariant system. In text books on ordinary differential equations, a linear
differential equation is usually written in the form:
dx
C P .t/x D Q.t/
dt
If we remove the time dependence from P and replace it with A and replace Q.t/
with Bu.t / we obtain:
dx
Ax D Bu.t/
dt
where u.t / is some time dependent input. This equation can of course be rearranged
to equal 9.1. That is, in its time invariant form (where P .t/ is replaced by P ),
197
198
k1 X o
!x
k2 x
k2 x
is given by:
Xo k1
(9.2)
1 e k2 t
k2
Solution 9.2 is called the total or complete solution because we can see that it is
make up of two terms that describe the dynamics of the systems. One term, on the
left, is concerned with the non-zero initial conditions and a second term on the right
is related to the input to the system (in this case a step function). We can summarize
some generalities from equation 9.2 that apply to all LTI systems. The first thing to
note is that the equation is in two parts. The first part is given by:
x.t/ D x.0/e
k2 t
x.0/e
k2 t
This is called the zero-input response (also called the free system or response)
because it doesnt include the input step function and assumes that the initial conditions are not necessarily zero. It is the response of the system in the absence of
any input. Figure 9.1 shows the zero-input response for a positive initial condition.
The natural response is to decay to zero.
199
x.t /
4
x.t/ D 5 e
0:5t
2
0
10
t
Figure 9.1 Zero-Input Response for a First-Order System
The second part of the total response is given by:
Xo k1
1 e k2 t
k2
This part of the equation is called the zero-state response. In this case the equation
assumes that the initial condition, x.0/ is zero but with an input step function with
respect to Xo . Figure 9.2 shows an example of a zero-state response on the firstorder system.
10
x.t/
8
6
x.t/ D
Xo k1
1
k2
k2 t
2
0
10
t
Figure 9.2 Zero-State Response for a First-Order System, vo D 5; k1 D 0:5
The zero-state response equation can also be divided into two parts:
Xo k1
k2
Xo k1
e
k2
k2 t
These two parts refer to the steady state and transient responses. Thus the first term
Xo k1 =k2 corresponds to the steady state solution that is obtained at infinite time.
200
We can see this if we take t ! 1 where the second part, .Xo k1 =k2 /e
zero.
k2 t
tends to
When we combine both the zero-input and zero-state parts of the equation we obtain
the total response shown in Figure 9.3.
15
x.t/ D vo e
k2 t
k2 t
x.t/
10
Xo k1
1
k1
10
t
Figure 9.3 Total Response for a First-Order System, Xo D 5; k1 D 1; k2 D 0:5
x.t / D
x.0/e k2 t
Zero-Input Response
Xo k1
k2
Steady State
Xo k1 k2 t
e
k
2
(9.3)
Forced Transient
Zero-State Response
Equation 9.3 summaries the various parts in the full response equation. That is:
Total Response D Zero-Input Response C Zero-State Response
Finally, the reason why we can split the total response into the various behavioral
modes is because of the principle of superposition.
201
Zero-input response:
1. Input is zero
2. Initial conditions are non-zero
Zero-state response:
1. Input is non-zero
2. Initial conditions are zero.
t 2 a2
t 3 a3
C
C :::
2
3
202
(9.5)
!
A2t 2
A3t 3
x.t/ D I C At C
C
C : : : x.0/
2
3
(9.6)
We can check if this is indeed the solution by differentiating the proposed solution
with respect to t , and see if we recover the differential equation 9.5. Differentiating
equation 9.6 yields:
!
3 2
dx
A
t
D 0 C A C A2t C
C : : : x.0/
dt
2
!
A2t 2
D A I C At C
C : : : x.0/
2
D Ax.t/
That is, equation 9.6 is the solution to d x=dt D Ax. By analogy, the matrix series
in equation 9.6 is referred to as the matrix exponential and is represented by the
curious expression:
e A t D I C At C
1 2 2
1
A t C A3t 3 C : : :
2
3
(9.7)
Although it may seem odd to raise e to the power of a matrix, the term, e A t has
properties very similar to a usual power. Given the definition of the matrix exponential, we can write the solution to d x=dt D Ax, (equation 9.6) in a very compact
way, namely:
x.t/ D e A t xo
(9.8)
203
D e At
Non-Homogeneous Solutions
Let us now consider the solution to the full state space system:
dx
D Ax.t/ C Bu.t/
dt
First we rewrite the above equation as:
dx
dt
Ax.t/ D Bu.t/
e A t Ax.t/ D e A t Bu.t/
dx
e A t Ax C e A t
dt
(9.9)
204
t0
that is:
e A t x.t/
e A t0 x.t0 / D
e A Bu./d
t0
Premultiply by e A t yields:
x.t/
e A .t
t0 /
x.t0 / D e A t
e A Bu./d
t0
or
x.t/
Total Response
D e A .t t0 / x.t0 / C
Zero-Input Response
e A .t
t0
/
Bu./d
(9.10)
Zero-State Response
We can split the solution into the zero-input and zero-state responses. The zeroinput response depends only on the initial conditions with the inputs set to zero.
The zero-state response depends on the inputs but where the initial conditions are
zero. It is worth noting that the times on the right-hand side are all relative, that is
the equation involves .t t0 / and (t ) terms, this is a result of time invariance.
In other words, the actual start time where the solution is evaluated does not affect
the qualitative form of the solution curves.
Many textbooks will set t0 D 0 so that one also finds the solution written as:
x.t/ D e A .t / x.0/ C
Z
0
e A .t
/
Bu./d
205
Example 1
Find the solution to the following system of differential equations assuming initial conditions, x1 .0/ D 2; x2 .0/ D 3:
dx1
D
dt
2x1
dx2
D x1
dt
x2
From the equations we can determine that the A matrix is given by:
2 0
1
1
Using equation 9.6 we see can write out the matrix exponential:
!
A2t 2
A3t 3
A
t
e
D I C At C
C
C :::
2
3
1 0
2
D
C
0 1
1
2
4t 2
2t C
2
61
6
6
D6
6
40Ct
8
0
4 0 t2
C
tC
7
1
3 1 2
8t 3
C :::
3
0 t3
C :::
1 3
3
3t 2
7t 3
C
C :::
2
3
tC
7
7
7
7
3
t
7
C : : :5
3
t2
2
(9.11)
The elements in the final matrix can be seen as series representations for exponentials. In
particular, the first series in the matrix represents 2e 2t and the forth element e t . The third
element is a bit more difficult to spot but can easily be shown to be the sum e t e 2t .
Therefore the solution to the original set of differential equations that incorporates the initial
conditions is:
x1 .t / D 2e
2t
t
x2 .t / D 2 e
D 5e
2t
2e
2t
C 3e
(9.12)
The approach used in this example is not simple because it involves spotting the appropriate
patterns in the series solution. There are other ways to solve for the solutions, most notably
the use of the Laplace transform which we will visit in Chapter 10.
206
Example 1 shows the solution to a simple set of linear differential equations without
any forcing terms. In state space these solutions represent the zero-input response.
The solution to the equations was a weighted sum of exponentials which can be
generalized to:
n
X
xj .t/ D
j k e k t
kD1
Exercises
1. Define the terms zero-input response, zero-state response and total response.
2. Why is it possible to separate out the response of a linear system into the
zero-input and zero-state responses?
3. Prove equation 9.7?
4. Given the following system:
dx1
D x1
dt
dx2
D x1 C x2
dt
207
208
10
Laplace Transforms
10.1 Introduction
As we saw from the last chapter, solving linear differential equations can be difficult, although with modern software such as Mathematica it can now be done relatively easily. However a classical approach to solving linear differential equations
and which has considerable other advantages, is the Laplace transform, developed by the 19th century French mathematician and astronomer, Pierre Simon de
Laplace. In this chapter we will describe the use of the Laplace transform in solving linear differential equations. The method is analogous to using logarithms to
make multiplication easier. When using logarithms we take log of the two numbers
we wish to multiply, add the resulting values, then take the antilog to obtain the
product. The argument is that simple addition is much easier compared to multiplication. The same strategy can used in solving linear differential equations. In
this case we take the Laplace transform of the differential equations which turns
them into algebraic equations, we solve the algebraic equations for the dependent
variables and then take the inverse Laplace transform to recover the solution.
The Laplace transform converts integral and differential equations into algebraic equations.
209
210
In this form the Laplace transform is also refereed to as the unilateral Laplace
transform to distinguish it from the bilateral form where the integration goes from
minus infinity to plus infinity. The variable s is a complex variable of the form
s D C j!.
Notation: By convention the Laplace transform of a function is often indicated with a capital letter, for example, F .s/; H.s/ or X.s/ while the original
function by lower case letters, for example, f .t/; h.t/ or x.t/.
The Laplace transform of a variable f .t/ is also denoted by the symbol
Lf .t /. The use of the two notations is generally clear from the context.
Unit Step Function Lets look at a very simple function and compute its Laplace
transform. The simplest function is f .t/ D 1 for t > 0. That is at t 0 f .t/ D 0.
This represents the step function, u.t/ (Figure 10.1) of size one at t 0. Since
f .t / D 1, the Laplace transform will look like (See Table 10.1):
Z 1
1
1 st t D1
st
D
e
F .s/ D
e
1 dt D
s
s
0
t D0
Lu.t/ D
1
s
u(t)
1
211
Derivative To illustrate deriving the Laplace transform for another function, let us
consider the Laplace transform for the derivative. In the context of solving differential equation this is probably the most important transform to consider. Consider
the first order derivative:
df
f .t/ D
dt
The Laplace transform for the derivative is given by:
Z 1
df
Lf .t/ D
dt
e st
dt
0
R
R
Using integration by parts: udv D uv
vdu where:
uDe
gives:
st
st
v D f .t/
df
dt D e
dt
D e
D e
st
st
e
0
st
f .t/
st
se
Z
1
f .t/. se
df
dt
dt
dv D
dt
st
dt/
0
1
f .t/.e
st
dt/
1
f .t/ 0 C sF .s/
f .1/
D
e1
Z
st
1
f .t/ 0 C s
That is:
du D
df
dt D
dt
f .0/
C sF .s/
e0
f .0/ C sF .s/ D sF .s/
df .t/
L
D sF .s/
dt
f .0/
Example 10.1
Find the Laplace transform for the following equation:
dx
D
dt
sX.s/
kx
x.0/ D
kX.s/
f .0/
212
Inverse Transform
If we determine the Laplace transform for a function, f .t/ to be F .s/ then we can
also define the inverse transform of F .s/ to be f .t/. The inverse transform is often
denoted by:
f .t/ D L 1 F .s/
Convergence
Throughout this book we will assume that all the Laplace transforms will converge.
This is possible if either f .t/ decays to zero at infinite t, or increases at a slower
rate than the rate of decay of e st . If this is the case then F .s/ can be defined and is
called the transform of f .t/, often symbolized using L. /. Note that integral starts
at t D 0 and is therefore only defined for positive t.
Tables of Transforms
Most of the time there is no need to derive the transforms directly because there are
published tables that list common functions and the corresponding Laplace transform (See Table 10.1 and Appendix H). In addition the Laplace transform is a linear
operation which means there are many useful properties of the transform which can
also greatly help when deriving the transform for a particular function.
1
L .1
b
then
L
ha
b
.1
bt
bt
/ D
i
/ D
1
s.s
b/
a
s.s
b/
213
f .t/
Lf .t/
.t/
au.t/
a
s
a
sCk
kt
ae
1
s2
t
a
.1
b
bt
a
s.s C b/
df .t/
dt
sF .s/
f .0/
Summation Rule If f .t/ and g.t/ are two functions then the Laplace transform of
their sum f .t / C g.t/ is;
Lf .t/ C g.t/ D Lf .t/ C Lg.t/
More generally
Laf .t/ C bg.t/ D aLf .t / C bLg.t /
Likewise we can also state:
L
For example:
L 3t C 8t 2 D 3L t C 8L t 2
If
F .s/ D
3
s
1
s
4
s
214
L F .s/ D 3L
D3
1
s
e 2t
1
s
C 4L
1
s
4e 3t
Change of Scale
If the time scale of time domain, t is changed, that is t is scaled by a, then the
Laplace transform is modified as follows:
L f .at/ D
1 s
F
a
a
For example, find the Laplace transform of sin.3t/. The transform of sin.t/ is:
F .s/ D
1
1 C s2
1
3
1C
1
s 2 D
3
9 C s2
at
f .t/ D F .s C a/
For example, find the Laplace transform of e at t 3 . First we find the transform of t 3
which is 6=s 4 . We then form F .s a/ so that:
L.e at t 3 / D
6
.s
a/4
215
a/ D F .s/e
a/u.t
as
This theorem can be modified as follows. Let g.t/ D f .t C a/, then g.t
f .t /. Using the second time shifting theorem we can write:
L u.t
a/ D
This result can be useful for finding the Laplace transforms of shifted functions. For
example, consider tu.t 120/. We can rewrite this as (Noting that L t D 1=s s ):
L tu.t
120/ D e
120s
De
120s
L t C 120
120
1
120s
De
C
s2
s
1
.1 C s 120/
s2
f .t/ D e
.kCs/t
1
Lf .t/ D
e
sCk
.sCk/t
kt
1
sCk
1
dt
0
dt
216
.t/ e
st
dt
Because the delta function only has a value at t D 0, the above equation becomes:
Z 1
Z 1
Lf .t/ D
.t/ 1 dt D
.t/ dt D 1
0
That is:
L.t/ D 1
For the shifted impulse, .t
a/ D e
as
5
s
Using the third row in Table 10.1 we can see that the inverse transform is given by:
y D 5e 2t
More often however finding the inverse transform directly for a table is not always
possible. Transforms obtained from differential equations will often be a quotient
of two polynomials:
F .s/
Y.s/
G.s/
We can therefore write Y.s/ as a sum of partial fractions.
217
Partial Fractions
Computing partial fractions is a way of representing a complicated fraction as a
sum of simpler terms. It is relatively easy sum a series of fractions into a complex
fraction, for example:
4
12x 32
8
C
D
xC2 x 5
.x C 2/.x 5/
The question to address here is how we can go the other way? There are a number of different ways to determine partial fractions but only one will be described
here. The first important rule is that the degree of the numerator must be less than
the degree of the denominator. If not, then the numerator must be divided by the
denominator using polynomial long division. For example in the following, the
numerator has a higher degree (3) than the numerator (2):
yD
x3
x2
x
x2
Cx
We would now determine the partial fraction for the first term x=.x 2 1/. Once
the expression has the right degree structure we then examine each factor in the
denominator. Each factor will result in a particular term in the partial fraction.
Factor in Denominator
ax C b
.ax C b/n
.ax 2 C bx C c/n
218
8x 42
C 3x 18
8x 42
8x 42
D
C 3x 18
.x C 6/.x 3/
12 D A.x
3/ C B.x C 6/
The key here is to realize that this expression is true what ever the value of x.
Therefore we can set values to x to selectively eliminate terms. For example, let
use set x D 3. This will eliminate the A factor:
12 D B.3 C 6/
8.3/
Solving for B gives:
BD
12 D A. 6
6:
3/
This yields:
A D 10
The final partial fraction is therefore:
10
8x 42
D
x 2 C 3x 18
xC6
2
x
2x 2
219
9x D A.x C 4/ C B.2x
1/
2/2 C B.x
2/ C C.x
1/.x
1/
As before we can eliminate the B and C terms by setting x D 1 and the A and
B terms by setting x D 2. This allows us to determine the values for A and C
respectively:
AD4
C D 16
The problem now is how to determine B since there is no value for x that will
simultaneously eliminate the A and C terms. In this case a different strategy is
used. Since we know the values of A and C , all we need to do is pick a value for
x. This means that all terms are set except for B. An easy value to assign to x is
zero. This allows us to write:
4.0/2 D .4/. 2/2 C B. 1/. 2/ C 16. 1/
D 162B 16
0DB
That is there is no B term. We can now write the final partial fraction as:
.x
4x 2
1/.x
2/2
4
x
16
.x 2/2
220
Linear Differential
Equation, y(t)
Apply
Laplace Transform
Time Domain
Solution, y(t)
kx
Taking the Laplace transform on both sides (using the Laplace transform table)
yields:
sX.s/ x.0/ D kX.s/
Solving for X.s/:
X.s/ D
x.0/
sCk
221
Taking the inverse Laplace transform (again using the Laplace transform table)
yields the solution:
x.t/ D x.0/e k t
Let us now apply the Laplace transform method to a more complex problem. Consider the three step pathway:
v3
v2
x1 ! x2 !
x2 drains into the environment. The system therefore has two state variables, x1
and x2 . The system of differential equations for this model is:
dx1
D
dt
k1 x1
dx2
D k 1 x1
dt
k2 x2
x1 .0/ D
sX2 .s/
x2 .0/ D k1 X1 .s/
k1 X1 .s/
k2 X2 .s/
The above set of equations represent a set of linear algebraic equations in X1 .s/
and X2 .s/. x1 .0/ and x2 .0/ represent the initial conditions for x1 and x2 at time
zero. If we assume that x1 D a and x2 D 0 at time zero, then:
sX1 .s/
aD
k1 X1 .s/
k2 X2 .s/
a
s C k1
X2 .s/ D
k1 a
.s C k1 /.s C k2 /
Now that we have solved for X1 .s/ and X2 .s/ we must apply the inverse Laplace
transform to recover the time domain solution. The inverse transform can be obtained by using the Laplace transform table in reverse. However we do not have
entries in the table that resemble the right-hand side of X2 .s/ (A more complete
table is given in Appendix H). In order to proceed we must split the equation into
simpler terms. The hope is that the simpler terms will be found in a table of Laplace
222
transforms. To do this we can use partial fractions. Let us first however find the
inverse transform of X1 .s/, this is straight forward because the term a=.s C k1 / is
the 3rd entry in the table, that is:
x1 D a e
k1 r
k1 a
.s C k1 /.s C k2 /
(10.1)
k2 . This allows us to
k1 a D B. k2 C k1 /
So that:
k1 a
k1 k2
A can be determining in a similar way by setting s D
BD
AD
k1 which yields:
k1 a
k1 k2
k1 a
1
.k1 k2 / s C k2
k1 a
1
.k1 k2 / s C k1
We can now use the reverse transform of a=.s C k/ to yield an equation for x2 .t/:
i
k1 a h k2 t
k1 t
x2 .t/ D
e
e
.k1 k2 /
Recall that a D x1 .0/. The use of the inverse transform to derive differential equations is perhaps not so important today given the availability of symbolic applications such as Mathematica or the open source tools such as Sage (http://www.
sagemath.org/) or wxMaxima (http://andrejv.github.com/wxmaxima/ and
Maxima http://maxima.sourceforge.net/). However, the ability to form the
transforms is still very important because the elimination of time and its replacement with s leads to a host of new analysis methods.
223
a// D e
sa
X.s/
For example the Laplace transform of a delayed unit step response, u.t a/ is given
by:
e at
L u.t a/ D
s
The Laplace transform for a pulse, u.t/ u.t 2/ can be evaluated using the
linearity property:
F .s/ D L u.t/
D
1
s
u.t
e 2s
1
D
s
2/ D L u.t/
e
s
L u.t
2/
2s
Sine Wave The Laplace transform for a sinusoidal wave, A sin.!t/ is given by
F .s/ D
A!
.s 2 C ! 2 /
224
kx
with initial condition x.0/ D 0. Apply the Laplace transform to both sides of the equation:
sX.s/ D
v
s
kX.s/
v
s.s C k/
X.s/ D
3x
with initial condition x.0/ D 6. Apply the Laplace transform to both sides of the equation:
sX.s/
6D
1
s
.s
3X.s/
s 6 11
2/.s C 3/
1 1
1 29
C
5 .s 2/ 5 .s C 3/
1
e
5
3t
e 5t C 29
225
Example 10.4
Solve the following system of differential equations:
dx
D x C 5y
dt
dy
D x y
dy
x.0/ D 0
y.0/ D 1
sY .s/
y.0/ D
X.s/
Y .s/
X.s/ D
s2
sin.2t //
Example 10.5
Solve the following second -order differential equation:
d 2x
dx
C
2
dt
dt
2x D 4
x.0/ D 2
dx
.0/ D 1
dt
2s
1/ C .sX.s/
2/
2X.s/ D
Combine terms:
.s 2 C s
2/X.s/ D
X.s/ D
4
C 2s C 3
s
2s 2 C 3s C 4
s.s 1/.s C 2/
4
s
226
Type
Multiplier x D Kv
X.s/ D K V .s/
Summer
x D v1 C v2
Cascade
1/
From the last equation we can determine the unknown coefficient, A, B and C .
Setting s D 0 eliminates B and C terms so that A D
2.
3
1
2
C
C
s
s 1 sC2
Taking the inverse Laplace transform for each term yields the final solution:
X.s/ D
y.t / D 3e t C e
2t
227
vo
! x1
k1 y1
! x2
k2 x 2
with ODEs:
dx1
D vo
dt
k 1 x1
dx2
D k 1 x1
dt
k2 x2
Derive the free, forced and full time-domain solutions for this system for both variables, x1 and x2 in response to a step function in vo .
Cascade
U(s)
X(s)
G1(s)
G2(s)
U(s)
X(s)
G1(s) G2(s)
Parallel
G1(s)
U(s)
X(s)
G2(s)
U(s)
G1(s) G2(s)
X(s)
Feedback
U(s)
E(s)
X(s)
G(s)
U(s)
G(s)
1 + G(s)H(s)
H(s)
X(s)
228
U(s)
G2(s)
G1(s)
X(s)
H1(s)
H2(s)
We can first simplify the inner loop using the closed loop feedback function (Figure 10.3
Feedback).
U(s)
G1(s)
G2(s)
1+H1(s) G1(s)
X(s)
H2(s)
We combine the blocks G1 .s/ and G2 .s/ using the cascade rule:
U(s)
G1(s) G2(s)
X(s)
1+H1(s) G1(s)
H2(s)
U(s)
X(s)
Exercises
1. Using tables, find the Laplace transform for the following functions:
(a) pulse, u.t
(b) t 2
2/
u.t
4/
3t C 4
(c) e u.t
t
/
a)
b)
229
xC4
.s C 1/.s C 2/.s C 3/
3s C 1
s.s C 1/.s C 2/
c)
s.s 2
d)
3s C 1
C 4s C 4/
3s 2 C 10s C 10
s 3 C 4s 2 C 6s C 4
Hint: s 2 C 2s C 2 D .s C 1/2 C 1
2s C 4 D 2.s C 1/ C 2
2t
x.0/ D 0
b)
dx
D e 5t C 5x
dt
x.0/ D 0
230
and x2 D e t .t C 1/
11
Zero-State Response
11.1 Introduction
In an earlier chapter (Chapter 9) we introduced the notion of the zero-input and
zero-state response. In this chapter we will look in more detail at the zero-state
response. To remind ourselves, the zero-state response is the response of a system
to an input when the initial conditions are zero.
x
v2
v1
where I is the input and x the single state variable. We will assume that v2 D k2 x.
The differential equation for this system is given by:
dx
D u.t/
dt
k2 x
231
232
where u.t / is an input signal that determines v1 . Let us compute the zero-state
response, that is we set x.0/ D 0. Taking the Laplace transform on both sides
yields:
sX.s/ D U.s/ k2 X.s/
In the equation we represent the input signal by the transform, U.s/. Solving for
X.s/ yields:
X.s/ D U.s/
1
s C k2
We can see that the transform for the state variable x is a product of the transform
of the input signal, U.s/, and a second term, 1=.s C k2 /. To obtain the response
of the system to an arbitrary input we need only replace U.s/ with the appropriate
transform. Of particular importance is the impulse response because its transform
is L..t // D 1. This means that the response of the system to an impulse is equal
to the second term in the expression, that is:
1
s C k2
If we applied a different kind of input, for example a unit step response, u.t/, then
U.s/ D 1=s, so that the response of the system is now:
1
1
X.s/ D
s s C k2
X.s/ D
Given that the application of an impulse means that U.s/ D 1, the response of a
system to an impulse is of particular importance. If we know the response of a
system to an impulse we can use this information to discover the response to any
input simply by multiplying the impulse response by the new input. The impulse
response of a system therefore fully characterizes a system. As will be repeated
elsewhere, the impulse response uncovers the transfer function of the system, in
this case 1=.s C k2 / and is of central importance to control theory. In the control
literature the transfer function is often represented by the symbol H.s/.
The impulse response of a system is the output of the system when we apply
an impulse, .t /, to the system with all initial conditions set to zero.
The response of a system to an impulse in zero state is called the impulse
response and is described by the transfer function, H.s/.
11.3. CONVOLUTION
233
11.3 Convolution
In the last section we saw that if an impulse, .t/ is applied to a system with all
initial conditions set to zero the output is called the impulse response. In the
Laplace domain this response is often represented by the symbol H.s/. Here we
wish to consider the same concept but now in the time domain where the impulse
response is represented by the symbol h.t/.
We begin by approximating a function q.t/ by a series of pulses shown in Figure 11.2. To do this we first define a pulse of unit height and width t by the
symbol (See Figure 11.1):
p.t n/
Note we have shortened t to .
p(t - nt)
nt
1
X
q.n/p.t
n/
where we sum over all pulses. The thinner the pulses the better the approximation.
In the limit as goes to zero, we can write the sum as a limit, that is:
q.t/ D lim
!0
1
X
1
q.n/p.t
n/
234
q(nt)
q(t)
nt
Figure 11.2 Approximation of a function q.t/ using a series of pulses of width t.
The height of the nth pulse is q.nt/.
We now make one important change, we multiply top and bottom by to obtain:
1
X
1
q.t/ D lim
q.n/
p.t
!0
1
n/
(11.1)
Recall that p.t n/ has unit height. If we divide this by , the height becomes
1=. Since the width of the pulse is , the area of p.t n/= must be 1= ,
which is one and unitless. This is an important change because in the limit as
! 0, p.t n/= will converge to an impulse .t / where we have replaced
n with :
1
lim p.t n/
!0
becomes an impulse:
1
.t / D lim p.t n/
!0
Taking the limit of equation 11.1 we note that the sum becomes an integral, such
that we the input function q.t/ can be written as:
Z 1
q.t/ D
q./.t /d
1
11.3. CONVOLUTION
235
We will define the impulse response, h.t/ as the output resulting from the application of an impulse, .t/, that is:
.t/ ! h.t/
If we assume the system is time invariant, then applying an impulse at a delayed
time will yield the same impulse response but delayed:
/ ! h.t
.t
/
q./.t
/
q./h.t
/d
q./.t
/d D q.t/
y.t/ D
q./h.t
/d
This integral is called the convolution integral and is often represented more simply using the notation:
y.t/ D q.t/ h.t/
Without proof we will also state the following useful properties:
(11.2)
236
237
(11.3)
F .s/ G.s/
(11.4)
1
s2
H.s/ D
1
1 C s2
s 2 .1
1
1
D
2
Cs /
s2
1
1 C s2
238
sX .s/
xo D AX .s/
where xo is the vector of initial conditions. Collecting the term X .s/ yields:
.sI
A/X .s/ D xo
X .s/ D .sI
(11.5)
A/ 1 xo
Taking the inverse transform on both sides (note that xo is a constant) yields the
solution x.t /:
x.t/ D L 1 ..sI
(11.6)
A/ 1 /xo
A/ 1 /
A/ 1 /
x.t/ D .t/xo
Note that from equation 9.8 the state-transition matrix is also the exponential matrix,
.t/ D e A t
(11.8)
Example 3
Apply 11.6 to the homogenous example given Chapter 9 in 1. In this example:
2 0
AD
1
1
Applying equation 11.6 we obtain:
x.t / D L
1 0
s
0 1
2
1
0
1
!
xo
239
1
sC1
0
x
1
.s C 2/ o
.s C 2/.s C 1/
Multiplying out each row and noting the initial conditions of x1 .0/ D 2; x2 .0/ D 3, gives:
2
1
x1 D L
sC2
x2 D L
3s C 8
.s C 1/.s C 2/
DL
5
.s C 1/
2
sC2
The inverse transforms for these expressions can be easily determined from standard Laplace
tables to yield:
x1 D 2e
x2 D 5e
C 2e
2t
These solutions correspond to those obtained using the matrix exponential approach (Equation 9.12).
Let us take the Laplace transform for the full state equation for x. That is:
dx
D L .Ax C Bu/
L
dt
sX .s/
xo D AX .s/ C BU .s/
where xo is the vector of initial conditions. Collecting the term X .s/ yields:
.sI
240
X .s/ D .sI
A/ 1 xo C .sI
A/ 1 BU .s/
(11.10)
A/ 1 /xo C L 1 ..sI
Recall that:
.t/ D L 1 ..sI
A/
A/ 1 BU .s//
D eAt
F .s/ G.s/
/Bu./d
We can also substitute the state transition matrices with the exponential matrices to
finally give:
Z t
A
t
x.t/ D e x C
e A .t / Bu./d
o
We can use the sifting property (6.1) to evaluate this integral to be:
x.t/ D e A t B
This result corresponds to equation 9.10.
241
(11.11)
The term .sI A/ 1 B is called the transfer function matrix, sometimes indicated
by H .s/:
X .s/ D H .s/U .s/
where
A/ 1 B
H .s/ D .sI
(11.12)
This expression corresponds to the forced component of the system and splits the
response into two parts, H .s/ which is an intrinsic property of the system and U .s/
which is the perturbation we apply to the system. Both together give us the effect a
perturbation has on the output of the system, X .s/.
Example 4
Derive the transfer function for the following system:
x
v2
v1
where I is the input and x the single state variable. We will assume that v2 D k2 x. The
differential equation for this system is given by:
dx
D u.t / k2 x
dt
where u.t/ is an input signal. In order to derive the transfer function we must set the initial
conditions, x.0/ to zero. Taking the Laplace transform on both sides yields:
sX.s/ D U.s/
k2 X.s/
At this stage we are not interested in any specific input signal, we therefore represent the
input signal by the symbol, U.s/. Solving for X.s/ gives:
X.s/ D U.s/
1
s C k2
1
s C k2
242
Example 5
Derive the transfer function for the following system using the state space representation:
x
v2
v1
with ODE:
dx
D k1 I k2 x
dt
We can cast the ODE in the state space representation:
dx
D k2 x C k1 I
dt
The A and B state variable matrices are given by:
A D k2
B D k1
k1 U .s/
k1
s C k2
Example 6
Find the step response to I for the system that has the transfer function (see previous
example):
k1
H.s/ D
s C k2
The Laplace transform of the step input, I is given by I =s, therefore the step response,
X.s/ is given by:
I k1
X.s/ D
s s C k2
The inverse transform is then given by:
I k1
x.t / D
1 e k2 t
k2
243
dx
L
dt
D L .Ax C Bu/
In deriving this equation we assumed that the initial conditions for the state variables were zero. We can therefore define the transfer function as:
The transfer function for a given state variable is defined as the ratio of
the output, X.s/ to the input, U.s/, of the system in the Laplace domain.
All initial conditions are assumed to be zero.
H.s/ D
X.s/
U.s/
For a given state variable, the transfer function is generally a ratio of two polynomials in s. The powers of s will all be integral powers. Another condition satisfied by
physical systems is that the order of the numerator polynomial is less than or equal
to the order of the denominator polynomial. It is this property that also allows us
to use partial fractions to separate the terms when we wish to apply the inverse
Laplace transform.
Since the Laplace transform of the impulse response .t/ is 1, we can also state
that the output function, X.s/ is equal to the transfer function when we apply an
impulse function to the system:
X.s/ D H.s/ when the input is an impulse function
In general, the transfer function will always be a ratio of polynomials:
244
H.s/ D K
.s
.s
z1 /.s
p1 /.s
z2 / : : : .s
p2 / : : : .s
zm 1 /.s
pn 1 /.s
b.s/
zm /
D
pn /
a.s/
The denominator, a.s/, of the transfer function is often called the characteristic
polynomial. If we isolate the characteristic polynomial and set it to zero:
a.s/ D 0
the equation is called the characteristic equation. In engineering, the roots of
the characteristic equation are called the poles. That is values for pi that set the
characteristic equation to zero. The roots of the equation in the numerator, b.s/, are
called the zeros. That is values of zi that set the numerator to zero. For example, if
p2 is 5, then one of the poles must be s D 5 since s 5 D 0.
The order of a transfer function is the order of the characteristic polynomial. That is the highest power of s in the denominator.
Example 7
1. What is the order of the transfer function:
a
sCb
The order of the characteristic polynomial (denominator) is one.
2. What is the order for the transfer function:
H.s/ D
s2
2s C 1
C 5s C 6
245
Example 8
1. Determine the poles and zeroes for the first order transfer function:
a
sCb
The system has no zeros but has one pole at s D
2. What are the poles and zeros for the transfer function:
H.s/ D
2s C 1
s 2 C 5s C 6
The first thing we can do is convert the transfer function to the general form:
H.s/ D
1
s C 1=2
2 .s C 3/.s C 2/
3 and
Final-Value Theorem:
The final-value theorem is very useful for computing the behavior of the model at
infinite time when the system might be at thermodynamic equilibrium or steady
state. The theorem is defined by:
lim f .t/ D lim sF .s/
t !1
s!0
v2
v3
x1 ! x2 !
where we assume that x1 is transformed into x2 and drains into the environment.
The system therefore has two state variables, x1 and x2 . The system of differential
246
k1 x1
dx2
D k1 x1
dt
k2 x2
a
s C k1
F2 .s/ D
k1 a
.s C k1 /.s C k2 /
If we apply the final-value theorem we find that both x1 and x2 tend to zero. For
example:
a
D0
lim x.t/ D lim s
t !1
s!0 s C k1
This is as expected, because all the mass will drain from the pathway leaving nothing in x1 and x2 ,
Class Question 1
Consider the following two gene cascade:
Xo
x1
v1
v2
x2
v3
v4
To make the calculations simpler we will assume simple mass-action kinetics for
all steps. That is:
v2 D k2 x1 v3 D k3 x1 v4 D k4 x2
The input signal is supplied to v1 .
a) Find the transfer function H .s/ for this system.
b) Obtain the solution to x1 .t/ and x2 .t/ assuming a simple step response in v1 ,
that is k1 Xo .
c) Use the final-value theorem to determine the steady-state level of x1 and x2 .
247
k1
s C k2
We can consider the input Xo to be a step function in the sense that at time=0, we
apply Xo . The response, X.s/ is then given by:
k1
Xo
X.s/ D U.s/H.s/ D
s s C k2
The steady state for this system can be determined from the final-value theorem:
Xo k1
sXo k1
D
yss D lim sX.s/ D lim
s!0
s!0 s.s C k2 /
k2
The inverse Laplace transform of X.s/ is given by:
x.t/ D
Xo k1
.1
k2
k2 t
(11.13)
t =
Comparing this to equation 11.13 we can see that the constant K is the steady state
solution, that is:
lim x.t/ D K
t !1
248
t =
//
K
D e
t =
D
t D0
K
That is, indicates how fast the system can relax and is often called the time
constant. Consider an electrical circuit that comprises of a resister, R, in series
with a capacitor, C , and a voltage source, V . By considering the resistance and
capacitance laws together with Kirchloffs voltage law we can derive the following
differential equation:
dvc
V D vc C RC
dt
In this equation vc the voltage across the capacitor is the state variable. This is a
first-order differential equation where the transfer function, H.s/ is given by:
H.s/ D
1=RC
1
D
s C 1=RC
RC s C 1
Exercises
1. Define the impulse response for a system.
2. Define the transfer function for a system.
3. What is the relationship between the impulse response and the transfer function?
4. Find the convolution of the functions u.t/ and u.t/.
5. What is the equivalent mathematically of convolution in the Laplace domain?
6. Write down the general form the transfer function.
7. What are the poles and zeros of the transfer function?
8. State whether the following transfer functions are first or second order:
H.s/ D
H.s/ D
s2
1
sCa
1
C sa C b
249
9. Using the final-value theorem determine the steady state values for x:
X.s/ D
X.s/ D
4
s.2s C 1/
6.s C 1/
s.9s C 0:25/
2
4s C 1
what is the gain K and the time constant for this system?
11. What does the time constant represent in a first-order system?
12. It has been determined that the transfer function that determines the concentration of acetaminophen in patients liver as a function of a constant infusion
of acetaminophen is given by:
H.s/ D
0:03
s C 0:06
250
12
Cellular Networks
12.1 Systems Equation
As described in Chapter 4, the cellular network equation (4.20) is given by:
dx
D Nv.x; p/
dt
(12.1)
251
252
@f
@f
x C A y
@x
@y
(12.2)
As with any linearization, the result is only approximate for small changes in x
and p. Note that the equation includes the stoichiometry matrix and two partial
derivatives and represents a set of linear ordinary differential equations. Such equations can be solved where the initial conditions are given by x.0/ D xo . The set
of parameters and inputs will be p D po .
Note that the equation only describes the evolution of x, not p. Whatever the
initial condition for p, it remains constant in the example considered here. In
253
xo
xo
xo
xo
0
D0
0
D0
po
po
po
po
D0
0
0
D0
Vm x
Km C x
The steady state level for x is denoted by xss . Let us assume we are going to change one
input, the Vm . To linearize we evaluate the various matrices defined in equation 12.2 around
xss and Vm :
d x
@v1 =@xss
@v1 =@p
D 1; 1
x C 1; 1
Vm
@v2 =@xss
@v2 =@p
dt
Each derivative is computed at xss and pss . We can now evaluate the individual partial
derivatives to obtain:
d x
0
D 1; 1
x
Vm Km =.Km C xss /2
dt
C 1; 1
Vm
(12.3)
which describes the rate of change of the perturbation in response to a step change in Vm
and an initial perturbation to x. Note that the terms that include xss are constant so that the
equation can be reduced to:
d x
D
dt
.C1 x C C2 Vm /
254
The solution to this equation with initial condition xo in xss , is given by:
h
x.t / D e
C1 t
iC
2
1
Vm C xo e
C1
C1 t
C2
Vm
C1
From the last example and the linearized equation 12.2, it is possible to define the
A and B matrices for a cellular model in the state space representation. Given the
linearized equation:
@v.xss ; pss /
@v.xss ; pss /
d x
DN
x C N
p
dt
@x
@p
we see that the A matrix or Jacobian, is given by:
ADN
where the term:
@v.xss ; pss /
@x
@v.xss ; pss /
@x
255
@v1
@x2
@v2
@x2
@vm
@x2
3
@v1
:::
@xn 7
7
7
@v2 7
7
:::
@xn 7
7
7
7
7
7
@vm 7
5
:::
@xn
In future examples we will also used the symbol Eji to denote the unscaled elasticity
for reaction i with respect to species j . Note that many of the terms in the elasticity
matrix are likely to be zero.
In future examples Eji will sometimes be used to denote the unscaled elasticity
for reaction i with respect to species j , that is:
Eji D
The B matrix is:
BDN
@vi
@xj
@v.xss ; pss /
@p
We will return to the C and D matrices later. We can now state that the state space
representation for a cellular network is given by the equation:
d x
@v
@v
D N x.t / C N p.t/
dt
@x
@p
Taking the Laplace transform of this equation will give us the transfer function for
this system as:
H .s/ D sI
@v
N
@x
@v
@p
(12.4)
256
or
H .s/ D .sI
A/
(12.5)
Example 12.2
Consider the simple gene regulatory network with a single transcription factor, x:
x
v2
v1
We will assume that the expression rate for x is controlled by a inducer I . We will also
assume that the first step is governed by the rate law v1 D k1 I and the degradation step by
v2 D k2 x.
We will compute the transfer function given by:
@v
N
@x
H .s/ D sI
@v
@p
First we need to collect the three matrix terms, N, @v=@x and @v=@p.
N D 1
2 @v 3
1
6 @x 7
@v
7D 0
D6
5
4
k2
@x
@v2
@x
2 @v 3
1
6
@I 7
@v
k1
6
7
D4
5D 0
@p
@v2
@I
Inserting these into the transfer function yields:
H .s/ D s
D .s C k2 /
0
1
k2
1
k1 D
1
k1
s C k2
k
1 1
0
257
Example 12.3
Consider the two gene cascade regulatory network with transcription factors, x1 and x2 :
x1
x2
v2
v1
v4
v3
v1 represent the expression rate for protein x1 . v2 represents the degradation rate of x1 . v3
is the expression rate of protein x2 . v4 is the degradation rate of x2 . I is the input inducer
to the cascade. We will assume that there are two inputs that can perturb the activity of v1
and v3 . We will compute the transfer function given by:
H .s/ D sI
@v
@x
@v
@p
First we need to collect the three matrix terms, N, @v=@x and @v=@p.
1
1
0
0
ND
0
0
1
1
@v1
6 @x1
6
6
6 @v2
6
6 @x1
6
@v
D6
6 @v3
@x
6
6 @x1
6
6
6 @v4
4
@x1
2
3
@v1
2
0
7
@x2 7 6
7 6
@v2 7 6 @v2
7 6
6
@x2 7
7 6 @x1
7D6
6
@v3 7
7 6 @v3
7
@x2 7 6
6 @x1
7 6
4
@v4 7
5
0
@x2
@v1
6 @p1
6
6
6 0
6
@v
D6
6
@p
6
6 0
6
4
2
7
7
7
0 7
7
7
7
@v3 7
7
@p2 7
5
0
7
7
7
0 7
7
7
7
7
0 7
7
7
7
@v4 5
@x2
258
@v2 @v4
@v2
@v4
C
sC
s C s2
@x1 @x2
@x1
@x1
2
@v2
6s C @x
1
6
H .s/ D 6
4 @v3
@x1
2
@v4
sC
6
@x2
1 6
D
6
D 4 @v3
@x1
12
0
7
7
7
@v4 5
sC
@x2
3
@v1
6 @p1 7
7
6
7
6
4 @v3 5
@p2
@v1
7 6 @p1
76
76
@v2 5 4
0
sC
@x1
@v4 @v1
sC
@x2 @p1
1 6
6
D
6
D4
@v3 @v1
32
3
0 7
7
7
@v3 5
@p2
2
3
0
@x1 @p1
sC
@v2
@x1
7
7
7
@v3 5
@p2
@v
@v
x C p
@x
@p
259
H J .s/ D C .sI
or (See equation 12.5):
H J .s/ D C H x .s/ C D
Substituting the various state space terms with the network terms yields:
@v
sI
H J .s/ D
@s
@v
N
@s
@v
@v
C
@p @p
A subscript, J , has been added to the transfer function to indicate that this function
refers to the flux rather than the individual state variables. In future we will also
refer to the state variable transfer function as H x .s/ to distinguish the two.
Example 12.4
The use of the flux transfer function, H J .s/ makes most sense when applied to metabolic
pathways. Consider then the following simple metabolic sequence of enzyme catalyzed
reactions:
v1
v2
v3
X0 ! x1 ! x2 ! X1
We will assume that the metabolites, Xo and X1 are fixed and that v1 to v3 are governed
by the rate laws:
v1 D E1 .k1 Xo
k2 x1 /
v2 D E2 .k3 S1
k4 x2 /
v3 D E3 k5 x2
where Ei is the concentration of the i t h enzyme and ki are the kinetic rate constants. m
the number of state variable is two and n the number of reactions is three. We will assign
260
three inputs to the model, corresponding to step functions in E1 , E2 and E3 . The number
of inputs, q, is therefore three.
First we write down the stoichiometry matrix for this system. This will be a two by three
matrix. Note that since Xo and X1 are fixed they will not appear in the stoichiometry
matrix:
1
1
0
ND
0
1
1
We must also write down the @v=@x which will be a n by m matrix:
2
3
@v1 @v1
6 @x1 @x2 7
6
7
6
7
6
@v
@v2 @v2 7
6
7
D6
7
@x
6 @x1 @x2 7
6
7
4 @v3 @v3 5
@x1
Two of the terms in this matrix are zero so that:
2
@v1
6 @x1
6
6
6 @v2
@v
D6
6 @x
@x
6 1
6
4
0
In addition @v=@p will be a n by q matrix:
2
@v1
6 @E1
6
6
6 @v2
@v
6
D 6 @E1
6
@p
6
6 @v3
4
@E1
@x2
3
0 7
7
7
@v2 7
7
@x2 7
7
7
@v3 5
@x2
@v1
@E2
@v2
@E2
@v3
@E2
3
@v1
@E3 7
7
7
@v2 7
7
@E3 7
7
7
@v3 7
5
@E3
Six of the terms in the @v=@p matrix are however zero, so that:
2
3
@v1
0
0
6 @E1
7
6
7
6
7
6
7
@v2
@v
6 0
0 7
D6
7
@E2
6
7
@p
6
7
6
7
@v
3
4 0
5
0
@E3
261
Next we construct the H x .s/ transfer function (12.5), given by H x .s/ D .sI
using the relations:
@v
ADN
@s
@v
BDN
@p
dv2 dv2
dv2 dv1
dv3 dv2
DD
C
Cs
Cs
dx1 dx2
dx1 dx1
dx2 dx2
H x .s/ D
2 dv dv
dv2
3
1
Cs ;
6 dE1 dx2 dx2
6
6
4
dv2 dv1
;
dx1 dE1
A/
1 B,
1
D
dv3
dv2
Cs
;
dx2
dE2
dv2
dv1
s
;
dE2
dx1
dv2 dv3
dx2 dE3
dv3
dE3
dv2
dx1
dv1
Cs
dx1
3
7
7
7
5
The flux transfer matrix, H J .s/ can be obtained by inserting H x .s/ into C H x .s/ C D
where
@v
C D
@x
@v
DD
@p
These equations are unwieldy and we will show in the next couple of chapters how their
structure can be greatly simplified to reveal useful information.
@v
N
@x
@v
Cx C I
@x
262
In the biology community, these canonical transfer functions are also refereed to as
the unscaled concentration and flux control coefficients when s is set to zero. We
will return to this point in a later chapter.
Canonical transfer functions are parameter independent in that the particular parameter used to perturb the system is not explicitly specified. All that is assumed is
that a reaction is perturbed, but by what is not indicated.
Example 12.5
13
Frequency Response
13.1 Introduction
In a previous chapter we saw that a general sinusoidal signal is given by the relation:
y.t/ D A sin.!t C /
where A is the amplitude, ! the frequency and the phase shift of the sinusoidal
signal. Of particular interest is that multiplication of a sinusoidal by a constant will
only change its amplitude, addition of two sinusoids, assuming they have the same
frequency will only change the amplitude and phase, while integration will only
change the phase and amplitude. None of these operations change the frequency
component. These are also linear operations. This means that if we input a sinusoidal signal into a linear system, the sinusoidal signal that is output will only see
changes to its amplitude and phase. This is a key point in the analysis of linear systems. Of particular interest is how the steady state responds to a sinusoidal input,
termed the sinusoidal steady state response.
263
264
P .t /dt
De
adt
D e at
d
ye at D e at b sin.!t/
dt
d
.ye at / D dy
(Note that dt
e at C yae at ). Assuming an initial condition of y.0/ D 0
dt
and integrating both sides gives:
yDb
!e
at
C a sin.!t/ ! cos.!t/
a2 C ! 2
At this point we only want to consider the steady state sinusoidal response, hence
as t ! 1, then
b
a sin.!t/ ! cos.!t/
yDp
p
2
2
a C!
a2 C ! 2
To show that the frequency of the input signal is unaffected, we proceed as follows.
We start with the well known trigonometric identity:
A sin.
265
! cos./ D A sin.
yDp
/ where D !t then
b
a2 C ! 2
sin.!t
(13.1)
!
(13.2)
tan
266
A sin (wt)
Linear System
A sin (wt+)
267
k1
s C k2
Let us apply a sinusoidal signal, cos.!t/ u.t/ to the system. To obtain the output
response we multiply the Laplace transform of the input by the transfer function,
H.s/:
k1
s
X.s/ D
s C k2 s 2 C ! 2
Multiplying this out yields:
X.s/ D
k1 s
.k2 C s/.s 2 C ! 2 /
.k12
k1
.k2 cos.t!/ C ! sin.t!/
C ! 2/
k2 t
k2 /
.k12
k1
.! sin.t!/ C k2 cos.t!//
C ! 2/
The sum of cosine and sine terms can be reexpressed as a sine and a phase change
using the trigonometry identity:
p
a cos.x/ C b sin.x/ D a2 C b 2 sin.x tan 1 .b=a//
Using this identity we can reexpress the response, x.t/ as:
q
k1
!
1
2
2
x.t/ D 2
k1 C ! sin !t tan
2
k2
.k1 C ! /
p
p
Note that x=x D 1= x, therefore:
k1
!
1
x.t/ D p
sin !t tan
k2
k12 C ! 2
268
!
k2
These results correspond exactly to the results (13.2 and 13.1) we obtained from
doing the same analysis on the first-order differential equation. However both approaches are tedious to use. Fortunately there is a much simpler way to obtain the
amplitude and phase change for a given transfer function which we will consider in
the next section.
' D tan
st
dt
we see that they are very similar. s in the Laplace transform is usually a complex
number C j!. In the case of the unilateral Fourier transform, s D j!. To
compute the Fourier transform we can therefore take the Laplace transform and
substitute s with j!. The reason this works is because the real part represents the
transient or exponential decay of the system to steady state, whereas the imaginary
part represents the steady state itself. When injecting a sinusoidal signal into a
269
system we are primarily interested in the sinusoidal steady state, hence we can
set D 0. In Fourier analysis, harmonic sine and cosines are multiplied into the
system function, f .t/ and then integrated. The act of integration picks out the
strength of the response to a given frequency.
Standard form for a complex number:
a C jb
Let us use the Fourier transform to obtain the amplitude and phase change for a
general linear first-order differential equation:
dx
C ax D f .t/
dt
Taking Laplace transforms on both sides yields:
sX.s/ C aX.s/ D L.f .t//
so that
L.f .t//
sCa
The transfer function of the system, H.s/, is however the ratio of L.output=.input/
so that
X.s/
1
D
H.s/ D
L.f .t//
sCa
To obtain the frequency response we set s D j!:
1
H.j!/ D
j! C a
X.s/ D
tan
b
a
270
From this complex number we can compute both the amplitude and phase shift.
First we must get the equation into a standard form:
H.j!/ D
1 .a
a C j! .a
j!/
a
D 2
j!/
a C !2
a2
j!
C !2
271
ImH.j!/
ReH.j!/
Decibels
The decibel (dB) is the magnitude in log space of the ratio of two measurements. In practice the measurements are often electrical power but can be
others such as sound. In terms of power the decibel is defined by:
P1
dB D 10 log10
Po
when referring to other quantities such as voltage it is common to use the fact
that power is proportional to the square of the voltage such that:
2
V1
V1
dB D 10 log10
D 20 log10
2
Vo
Vo
Traditionally Bode plots measure amplitude changes in voltage so that the
decibel used in plotting Bode plots is defined using the later expression. It is
important to note that because a decibel is related to a ratio, if a given decibel
is negative then the signal is attenuated; if the ratio is positive then the signal
is amplified. Common Values:
20 log10
p
2 D 3 dB
1
20 log10 p D 3 dB
2
p
20 log10 10 D 10 dB
20 log10 10 D 20 dB
20 log10 100 D 40 dB
Rules of thumb:
1. As a number doubles, the decibel values increases by 6 dB
2. As a number increases by a factor of 10, the decibel values increases by 20
dB.
272
First-Order System
To illustrate a Bode plot let us consider the first-order system we have looked at
before in Figure 13.2.
Xo
V2
V1
k1
s C k2
(13.3)
k1
j! C k2
To obtain the amplitude and phase shift we must convert the above transfer function
into the standard form by multiplying top and bottom by the conjugate complex
number:
k1
k2
j! C k2 k2
j!
k1 k2 k1 j!
k1 k2
k1 j!
D
D 2
2
2
2
2
j!
k2 C !
k2 C !
k2 C ! 2
p
From this we can determine the amplitude ( a2 C b 2 ) given by:
s
k12
k1
Amplitude D jH .j!/j D
D
p
2
k2 C ! 2
k22 C ! 2
Likewise the phase change can be computed from tan 1 .b=a/:
!
1
Phase D tan
k2
(13.4)
273
Amplitude
Figure 13.3 shows a plot of the amplitude as a function of the frequency. Although
it looks as if the plot starts at zero amplitude, it doesnt. At zero frequency (! D 0)
the y-axis intercept is k1 =k2 (13.4). The other thing to note is that the units of
frequency are in rad/sec. If for some reason the graph needs to be redone in Hertz,
one only needs to express the frequency in terms of Hertz. This can be done by
substituting 2f , where f is the frequency in units of Hertz where ever ! appears.
s
k12
k1
D
Amplitude D jH .j!/j D
p
2
k2 C ! 2
k22 C .2f /2
Magnitude (dB)
10
0
10
20
30
10
10
10
100
101
102
103
104
Phase Shift
Note that the phase shift in Figure 14.2 starts at zero and the goes negative, meaning
the phase lags. At high frequencies the phase shift tends to 90o . The way to
understand this is to look at the phase expression and realize that the smaller k2 the
more likely the phase shift will be 90o . If k2 is very small then we can assume
there is very little degradation flux, this means that the change in x is dominated by
the input sine wave. The maximum rate of increase in x is when the sine wave is at
its maximum peak. As the input sine wave decreases the rate of increase in x slows
until the input sine wave crosses the steady state level of x. Once the input sine
Phase (Degrees)
274
50
90o line
100
10
10
10
10
Xo
Xo and x
Zero Rate
0
0:2
0:4
0:6
Time (t)
0:8
Figure 13.5 90o phase shift at high frequencies or low degradation rates for the
model shown in Figure 13.2. Note that the concentration of x peaks when the input
rate is at zero.
275
Bandwidth
In electrical engineering, a common measure of an amplifiers performance is its
bandwidth. This is the point in the frequency response where the power output from
the amplifier has halved. In terms of voltage this represents a drop to 70.7% of the
peak value. For example if the peak voltage is 5 volts and it drops to 70.7% of its
value we get a voltage of 3.535 volts. Recall that power is given by P D V 2 =R.
Assuming a constant resistance of 1 ohm, the power output at 5 volts will be 25
watts, whereas at 3.535 volts, the power output is 12.5 watts, that is the power has
exactly halved. In terms of decibels, a reduction of an output to 70.7% of its peak
value is given by:
0:707
dB
change D 20 log10
1
D
20 log10 .1:414/
3 dB
The last calculation allows us to express the bandwidth in terms of the drop in
decibels, namely 3 dB. That is the frequency at which the the amplitude has dropped
by 3 dB is the bandwidth.
The bandwidth of a system is the frequency
p at which the gain drops below the
3 dB peak. 3 dB is when the signal is 1= 2 of the maximum signal amplitude
(about 70% of the signal strength).
log.B/
276
One of the building blocks of a typical transfer function is the expression: 1 i!=k.
Let us begin by considering the Bode plot for this expression. The approach to take
is:
Consider the expression at low frequency (! ! 0)
Consider the expression at high frequencies (! ! 1/)
We compute the absolute values for the term, 1 i!=k:
s!0
Xo
D Xo
s
A Xo input will result in a xss change. This change can be calculated using the
finial-limit theorem:
xss D lim sX.s/ D lim sH.s/U.s/
s!0
D lim sH.s/
s!0
s!0
Xo
D H.0/ Xo
s
277
That is the steady state level of the output, xss is given by that ratio:
xss
D H.0/
Xo
That is, the gain at zero frequency is the ratio of the change in the steady state output
to the change in the input signal. As the change in the input signal gets smaller, the
ratio of the input to output converges to the derivative of the state variable with
respect to the parameter undergoing a step response:
H.0/ D
dx
dp
(13.5)
x
v2
v1
k1
s C k2
The zero frequency response is determined by setting s D 0 so that the amplitude component is given by:
k1
jH x .0/j D
k2
Clearly the amplitude change is not zero, so what is it? If we look at the steady state
solution, the answer will become clear. At steady state, v1 v2 D 0, that is:
k1 Xo D k2 x
278
k1 X o
k2
(13.6)
which is the amplitude change we observed at zero frequency. Therefore the amplitude
change at zero frequency is the sensitivity of the particular state variable (species concentration) with respect to the input signal.
We can use the final limit theorem to show that we can get the same result if we compute
Hx .s/. Consider the same transfer function again:
Hx .s/ D
k1
s C k2
Xo
s
We now use the final limit theorem to find the value, ss , at t ! 1, that is to find the limit
to:
Xo
k1 Xo
k1
xss D lim sHx .s/
D lim s
D
Xo
s!0
s!0 s C k2
s
s
k2
That is xss , the steady state change in the output is Xo k1 =k2 . Dividing both sides by
Xo we obtain the gain which is that same as the gain we observed zero frequency (13.6).
gain D
xss
Xo
Xo
279
x1
x2
v2
v1
v4
v3
@v
N
@x
H .j!/ D j!I
@v
@p
First we will collect the three matrix terms, N, @v=@x and @v=@p.
ND
@v1
6 @x1
6
6 @v2
6
6
@x1
@v 6
D6
6
@x 6 @v3
6 @x1
6
6 @v
4 4
@x1
2
1
0
1
0
0
1
0
1
3
@v1
@x2 7
7
2
3
@v2 7
7
0 0
7
@x2 7 6k2 0 7
7D6
7
4k3 0 5
@v3 7
7
0 k4
@x2 7
7
7
@v4 5
@x2
280
3
@v1
6 @Xo 7
6
7
6
7 2 3
6 @v2 7
k1
6
7
6 @Xo 7 6 0 7
@v
7D6 7
D6
7 405
@p 6
6 @v3 7
6 @X 7
0
6 o7
6
7
4 @v4 5
@Xo
and insert the terms into the transfer function to yield:
0
B
B
H .j!/ D B
@
k1
k2 C jw
k1 k3
.jw C k2 /.jw C k4 /
1
C
C
C
A
It is far easier however to compute the amplitude and phase numerically. Many
computer languages provide a function called ArcTan2, or more commonly, atan2
that will compute the arcTangent of a value while taking into account the four quadrants. These functions take two arguments, the first is the imaginary part and the
second argument the real part. For example, using the Matlab atan2 and converting
radians to degrees:
z = 4 + 3 i
theta = atan2 (imag 9z), real (z))*180/3.1415
36.87
@v
@v
x C p
@x
@p
281
Magnitude (dB)
50
50
10
10
10
10 2 10 1 100
Frequency (rad/sec)
101
Figure 13.8 Bode Plot for a two gene cascade: Magnitude Plot. All rate laws are
simple irreversible first-order with values: v1 W k1 D 0:1; v2 W k2 D 0:2; v3 W k3 D
0:25; v4 W k4 D 0:06; Xo D 1, see Figure 13.7
Phase (degrees)
0
50
100
150
200
10
180o line
5
10
10
10
10
100
101
Frequency (rad/sec)
282
A/ 1 BP.s/ so that:
V .s/ D C .sI
A/ 1 B C D P.s/
A/ 1 B C D
Substituting the various state space terms with the network terms and setting s D
j!, finally yields:
@v
H J .j!/ D
j!I
@x
@v
N
@x
@v
@v
C
@p @p
Exercises
1. An amplifier has a peak power output of 50 mW but drops to 25 mW when
the input signal frequency reaches 20 kHz. What is the change in dB between
the two power levels?
14
283
284
1 @v
@x
1 @v
@p
diag.x/
diag.p/
Let us now consider the scaled transfer function at zero frequency. At zero frequency, we know that H.0/ D d x=d p, so that.
dx
@v 1 @v
D
N
N
dp
@x
@p
We can pre-multiply by diag.x/
diag.x/
285
@v
N
@x
@v
diag.p/
@p
N diag.v/ diag.v/
@v
@x
to obtain:
N diag.v/ diag.v/
@v
Q x .0/ D
diag.x/ H
@x
Q x .0/ D
.N diag.v/ "x / H
N diag.v/ diag.v/
@v
diag.p/
@p
1 @v
@p
diag.p/
where "x is the scaled species elasticity and "p is the scaled parameter elasticity.
Finally multiplying both sides by .N diag.v/ "x / 1 :
Q x .0/ D
H
.N diag.v/ "x /
N diag.v/ "p
From this we will define the scaled canonical transfer function as:
C x .0/ D
.N diag.v/ "x /
N diag.v/
Example 14.1
Compute the frequency response for the single gene circuit:
v1
x
v2
We will assume that the expression rate for x is controlled by a inducer I . We will also
assume that the first step is governed by the rate law v1 D k1 I and the degradation step by
v2 D k2 x.
286
In a previous example, the transfer function for this system with respect to the input I was
shown to be:
@v1
@I
H .s/ D
@v2
sC
@x
Multiplying both sides by I and dividing both sides by x yields:
I
H .s/ D
x
@v1 I
v1 "Iv1
@I v1
D
@v2 x
s C v2 "vx2
s C v2
@x v2
v1
Q x .s/ D
H
.v1 =x/"Iv1
s C .v2 =x/"vx2
Magnitude (dB)
10
0
10
20
30
10
10
103
104
Figure 14.1 Bode Plot: Magnitude (dB). The left most plot has k1 D 0:4 and
k2 D 0:2. The right most plot, the rate constants are 100 times bigger. This shows
us that at higher rates, the system is able to faithfully reproduce the amplitude of
the input at much higher higher frequencies. That is the system can be much more
responsible to external changes.
Figure 14.2 shows the amplitude Bode plot for this system and highlights the effect of
higher rates (perhaps brought about by larger rate constants). In this case higher rates mean
that the network is able to respond better at faster changing inputs since the amplitude starts
to drop at much higher frequencies.
287
Magnitude (dB)
10
0
10
20
30
10
10
103
104
Figure 14.2 Bode Plot: Magnitude (dB). Effect of a more responsive degradation
step, "2x . Reduces the DC gain and extends out the frequency response.
The previous example showed that the scaled transfer function for the transcription
factor, x with respect to the inducer, I , was given by:
Q x .s/ D
H
.v1 =x/"Iv1
s C .v2 =x/"vx2
The zero frequency response can be obtained by setting s D 0 and known that
v1 D v2 at steady state, we obtain:
"Iv1
Q
H x .0/ D v2
"x
This tells us the relative sensitivity of the steady state to changes in inducer, that is:
dx I
dI x
If the degradation reaction, v2 is first-order, the elasticity, "vx2 D 1, therefore the
sensitivity is equal to:
Q x .0/ D "v1
H
I
That is the sensitivity of x is equal to the response of gene expression to inducer.
By way of example, if the effect of inducer on v1 is governed by a Hill equation
288
Y/
15
Structural Constraints
15.1 Introduction
In Chapter 12 a brief mention was made of the canonical transfer functions. These
were defined as:
@v 1
N
C x D sI N
@x
CJ D
@v
Cx C I
@x
In the biology community, these canonical transfer functions are also refereed to
as the unscaled concentration and flux control coefficients when s is set to zero.
Canonical transfer functions are parameter independent in that the particular parameter used to perturb the system is not explicitly specified. All that is assumed
is that a reaction is perturbed, but by what is not indicated. Because we can define
transfer functions in terms of the topological structure of biochemical networks,
there are a range of interesting constraints, called structural constraints that emerge
from the canonical transfer functions.
289
290
@v
N
@x
NK
yielding:
C x .j!/K D 0
Post multiplying the canonical equation, C J by K gives:
C J .j!/K D
@v
C xK C K
@x
so that:
C J .j!/K D K
Connectivity Constraints
Given the canonical concentration transfer function:
@v
C x .j!/ D j!I N
@x
@v
N
@x
j!I
@v
N
@x
DI
or
j!I
@v
N
@x
j!I
@v
N
@x
291
j!I
@v
N
@x
j!I
@v
DI
@x
Rearranging:
@v
D j!I
N
@x
@v
N
@x
j!I
@v
N
@x
j!I
so that at ! D 0 we obtain:
C x .0/
@v
D
@x
@v
Cx C I
@x
@v
:
@x
C J .j!/
But at ! D 0, C x .0/@v=@x D
@v
@v
@v @v
D Cx C
@x
@x
@x @x
Therefore:
C J .0/
@v
D0
@x
Scaled Constraints
In Chapter 14 the scaled canonical transfer function with respect to the species was
defined by:
C x .0/ D .N diag.v/ "x / 1 N diag.v/
292
For clarity we will drop the .0/ notation. Let us multiply both sides of above
equation by the one vector, 1; 1; T , so that:
@v 1
N diag.v/ 1
C x1 D
N diag.v/
@x
@v
N diag.v/
@x
C x1 D
Nv
@v
Cx C I
@x
N diag.v/ C I
Cix D 0
CiJ D 1
293
where the summation is over all reaction steps in the pathway. These are called the
flux and concentration summation theorems. Operationally they have an interesting
interpretation. The way to understand the canonical transfer functions is to think of
Cix as:
Cix D
x vi
dx vi
'
dvi x
vi x
where vi is called the local reaction rate of the the ith reaction. A local reaction rate
has the following operational interpretation. For the ith reaction, let us clamp (fix)
all state variable molecular species that might affect the reaction rate. We now find
a suitable parameter, for example the enzyme concentration, and increase this by
some small factor. This will cause the reaction rate to increase by some value vi .
We now release the clamps and let the system evolve. With a higher reaction rate,
substrate will be consumed at a higher rate and product produced at a higher rate.
Once the system reaches its new steady state, the reaction velocity vi has settles
to a new rate with changes in concentration x by x. To compute the canonical
transfer function, we take the ratio of the x to vi . We stress again that the vi is
the change in reaction rate we impose before the system is allowed to evolve to the
new steady state. We do not specify how the vi rate change is brought about.
If we assume that local reaction rates are proportional to enzyme concentration,
that is:
vi D ki Ei
Then:
dx Ei
dx ki Ei
D
d.ki Ei / x
dEi x
294
16
First-Order Systems
16.1 Introduction
We now have all the machinery in place to make a thorough study of systems. The
simplest model, with the exception of zero-order systems, are first-order systems.
As introduced in Chapter 11, the transfer function for a first-order system is given
by:
H.s/ D
K
s C 1
(16.1)
Note that the order of a system is defined by the order of the characteristic polynomial. That is, the highest power of s in the denominator. Figure 16.1 shows three
typical first-order systems that includes a simple water tank, an RC circuit, and a
gene expression cassette.
Simple Tank Model
We have already seen the simply tank model from Chapter 1. The differential
equation that describes the change in volume of the tank is given by:
dV
D Q1
dt
Q2 D Q1
.K=A/V
295
296
b) RC Circuit
V1
Q1
R
h1
Vo
K1
Vr
Q2
v1
x
v2
dVo
dt
dVo
R C Vo
dt
297
Rearranging yields:
dVo
V1
D
dt
RC
Vo
RC
kx
What is common with the three examples is that the differential equation describing
their dynamics is a linear differential equation:
dx
D u.t/
dt
kx
or in standard form:
dx
C P .t/x D f .t/
dt
Taking Laplace transforms on both sides, setting x.0/ D 0, and assuming the input
u.t / is a unit impulse we obtain the first-order transfer function:
H.s/ D
1
sCk
This is slightly different from equation 16.1 where we also include the input gain
K in the numerator. We will now summarize the dynamical properties of a first
order system. All the properties we describe apply equally to all three systems in
Figure 16.1.
298
Operationally is the time it takes for the system to decay to 36.8% of its original
value. Therefore if equals 2 minutes and x.0/ is 10 mM, then after 2 minutes,
only 3.68 mM will be left. Anther way to look at this is to note that the time
constant is related to the half-life by way of:
t1=2 D ln 2
ko
s C 1=
t =
We see that the impulse response is almost identical to the free response but with a
different scaling factor.
1
1
s s C 1=
t =
This result tells us that initially x.t/ is zero but over time x.t/ tends to ko . The
time constant, , tells us that at time , the value of x.t/ is 63.2% of its maximum
value. Another useful property is that at t D 0, the slope of the tangent line at t D 0
is given by 1=.
299
x.t/
0:8
0:6
D1
Rise Time D 2:2
Settling Time D 4
Completion Time D 5
0:4
0:2
0
Time
Figure 16.2 Transient response of a first-order system when a unit-step is applied.
D1
The transient step response for a first-order system is shown in Figure 16.2. There
are a number of transient metrics that are commonly used in the literature to characterize such a response. We have already encountered one which is the time constant.
Two others are also frequently used:
Rise Time, Tr The raise time is defined as the time for the response to go from
10% to 90% of its final value. If we set the transient equation to 0.1 and 0.9 and
compare the difference the raise time can be computed from:
Tr D 2:2
For a system with a time constant of 2 minutes, this means it will take 4.4 minutes
for the system to go from 0.1% to 0.9% of its maximum value.
Settling Time, Ts The settling time is defined as the time for the response to reach
within 2% of the final value. By looking at the transient equation and setting the
final value to 0.98 it is possible to show that the setting time is given by:
Ts D 4
For a system with a time constant of 2 minutes, this means it will take 8 minutes
for the system to reach within 2% of the final value. Often it is considered that the
system has completed its response when t 5 . If the case of our example, we
consider the response has completed after 10 minutes.
300
1
s C 1=
From previous work we know that the frequency amplitude for this system is given by:
1
Amplitude D p
.1= /2 C ! 2
p
We now wish to solve for ! in terms of 1= at the bandwidth point, that is 1= 2 of the
DC level. It is important to recall that the DC level is given by the transfer function at zero
301
Zero Frequency (! D 0)
k1 =k2
Cut-off or bandwidth
1=
Amplitude Loss beyond cut-off -20 dB/Decade
Table 16.1 Summary of Amplitude Properties
frequency which in this case is . Therefore we must solve the following equation in terms
of :
1
p Dp
2
.1= /2 C ! 2
Inverting and squaring both sides yields:
2
1
D 2 C !2
2
From which we conclude that:
!D
1
10
Magnitude (dB)
Bandwidth:
1
D!
0
10
20
30
10
10
10
100
101
102
103
104
tan
!
k2
Phase (Degrees)
302
50
90o line
100
10
10
10
10
Further Reading
1. Edelstein-Keshet L (2005) Mathematical Model sin Biology. SIAM Classical
In Applied Mathematics. ISBN-10: 0-89871-554-7
2. Fall CP, Marland ES, Wagner JM, Tyson JJ (2000) Computational Cell Biology. Springer: Interdisciplinary Applied Mathematics. ISBN 0-387-95369-8
3. Steuer R and Junker BH (2009). Computational models of metabolism: stability and regulation in metabolic networks. Advances in chemical physics,
142, 105.
Exercises
1. Show that for a unit step response applied to a first-order system, the slope at
t D 0 is given by 1=tau.
2. Show that the raise time can be computed from the expression Tr D 2:2.
3. Show that the settling time can be computed from the expression Ts D 4 .
4. Determine the behavior of a first-order system in response to a unit-ramp.
5. Using Figure 17.23, show graphically that the toggle switch model has two
stable and one unstable steady states.
17
Stability
17.1 Introduction
Biological systems are continually subject to external and internal momentary perturbations. Unless adequately checked, such perturbation could be highly disruptive and prevent normal functioning of the cell. Of particular interest is whether a
given momentary perturbation leads to the system diverging from the steady state
or whether the system relaxes back. In this chapter we will look more closely at the
stability of systems.
k2 x
303
304
Xo k1
1
k2
k2 t
C x.0/e
k2 t
(17.1)
As described before the first part of the equation is the steady state term and the
second part the transient term. The dynamics of the system is therefore determined
by the second part of the equation. This part corresponds to the solution of the
homogeneous system, i.e. when there is no input to the system. That is the solution
to:
dx
D Ax
dt
We know that the general solution to the homogeneous system is a weighted sum
of exponentials, c1 e i t , where i are an eigenvalues of the A or Jacobian matrix.
Recall that the eigenvalues of a matrix A are defined by:
Av D v
where v is an eigenvector. The above definition can be rearranged to give a set of
linear homogeneous equations:
I/v D 0
.A
I/ D 0
I is given by:
2
0
2
.1 C /
/.1 C / D 0
305
1 and 2.
It will be recalled that the Laplace transform with non-zero initial conditions is
given by:
X .s/ D .sI A/ 1 xo C .sI A/ 1 BU .s//
The first part represents the transient response under non-zero initial conditions
and the second part the zero-state response. Other than a change of sign, the term
.sI A/, is the same as the expression for computing the eigenvalues for the
homogeneous system. We therefore conclude that terms for the homogeneous
solution can be obtained by looking at polynomial that results from expanding the
determinant of .sI A/. When written out in the transfer form, the expansion of the
determinant appears as the polynomial in the denominator of the transfer function.
We conclude that the dynamics of a system can be determined by examining the
solutions to the polynomial in the denominator of the transfer function, H.s/.
Given the importance of this result, control engineers have named the various components of the transfer function. For example the denominator is often called the
characteristic equation. Likewise the equation formed by setting the numerator
to zero is often called the characteristic polynomial. The roots to these polynomials also have terms attached to them. The roots of the characteristics equation
(denominator) are called the poles, while the roots of the numerator are called the
zeros.
Consider a system with the transfer function:
H.s/ D
s2
s
C 6s C 8
s2 D
These values correspond to the terms in the solution to the homogeneous system.
That is the homogeneous solution must be of the form:
C1 e
4t
C C2 e
2t
306
Given that both exponentials are negative, we conclude that an initial condition, x
will return back to the steady state solution. Our criterion for looking at the internal
stability is to examine how internal perturbations in x relax or expand and this
The steady state of a system is stable if all the eigenvalues of the Jacobian
matrix have negative real parts. The steady state is unstable if at least one of
the eigenvalues has a positive real part.
Complex Eigenvalues
In general the eigenvalues computed from the Jacobian matrix can be complex, that
is D aCi b. Since entries in the jacobian are always real, any complex eigenvalue
will be automatically accompanied by its complex conjugate. For example:
3
2
4
1
has the characteristic equation: 2 2 C 5 D 0. Since one of the roots is the
complex number D 1 C 2i, the second root must the complex conjugate: D
1 2i . If the n n matrix A has real entries, its complex eigenvalues will always
occur in complex conjugate pairs. Why is this important? Consider a solution to a
homogeneous system:
x.t/ D c1 z1 e .Ci/t C c2 z2 e .
i/t
bi, we get:
Multiply out and separate the real and imaginary parts yields:
x.t / D e t c1 .a cos.t/
C c2 .a cos.t/
307
The complex terms cancel leaving only the real parts. If we set c1 C c2 D 2k1 and
.c1 c2 /i D 2k2 then:
x.t/ D e t k1 .a cos.t/ b sin.t//
k2 .a sin.t/ C b cos.t//
This result shows that the appearance of complex numbers in the eigenvalues results
in periodic solutions.
Example 17.1
The following system:
! S1 ! S2 !
is governed by the set of differential equations:
dS1
D3
dt
2S1
dS2
D 2S1 4S2
dt
The Jacobian matrix is computed by differentiating the equations with respect to the steady
state values of S1 and S2 :
2
0
J D
2
4
The eigenvalues for this matrix are: 2 and 4, respectively. Since both eigenvalues are
negative, the system is stable to small perturbations in S1 and S2 .
Example 17.2
Consider the system:
Xo ! S1
vo
S1 ! X1
S1
S1 ! S 2
S1 .1 C S23 /
S2 ! X2
5S2
where Xo ; X1 and X2 are fixed species. At steady state S1 D 2:295 and S2 D 1:14
with parameter values vo D 8. Determine whether this steady state is stable or not. The
differential equations for the system are given by:
dS1
D vo
dt
S1
S1 .1 C S23 /
dS2
q
D S1 .1 C S2 /
dt
5S2
308
The Jacobian matrix is computed by differentiating the equations with respect to the steady
state values of S1 and S2 :
#
# "
"
3S1 S22
2 S23
3:4815
8:948
D
J D
2:482
3:948
1 C S3
5 C 3S1 S22
The eigenvalues for this matrix are: 0:2333 C 2:9i and 0:2332 2:9i, respectively. Since
the real parts of the eigenvalues are positive, the system is unstable to small perturbations
in S1 and S2 .
! y1
k1 y 1
! y2
k2 y 2
309
with ODEs:
dy1
D vo
dt
k1 y1
dy2
D k1 y1
dt
k2 y2
We can assign particular values to the parameters, set up some initial conditions
and plot in phase space the evolution of x1 and x2 . If we replot the solution using
many different initial conditions we get something that looks like the plots shown
in Figure 17.3 to 17.8.
.
0. 8
S2
0. 5
0
0
S1
S2
2
Phase Plot
0. 6
0. 4
0. 2
10
0
0
0.2
0. 4
0. 6
Time
0. 8
1. 2
1. 4
S1
Figure 17.1 Time course simulation plot and corresponding phase plot.
2.0
1.8
1.6
1.4
S2
1.2
1.0
0.8
0.6
0.4
0.2
0.0
0.0
0.2
0.4
0.6
0.8
1.0
1.2
1.4
1.6
1.8
2.0
S1
Figure 17.2 Multiple trajectories plotted on the phase plot Jarnac listing: 17.2.
A visual representation of the phase space is often called a phase portrait or phase
310
plane. To illustrate a phase portrait consider the following simple reaction network:
vo
! S1
k1 S 1
! S2
k2 S 2
k1 S1
k2 S2
We can assign particular values to the parameters, set up some initial conditions,
and plot the evolution of S1 and S2 in phase space. If we replot the solution using
many different initial conditions, we get something that looks like the plots shown
in Figures 17.2 to 17.8.
The plots illustrate a variety of transient behaviors around the steady state. These
particular transient behaviors apply specifically to linear differential equations. If
we have a nonlinear system and we linearize the system around the steady state, the
linearized system will also behave in a way suggested by these plots.
Consider the general two dimensional linear set of differential equations:
dS1
D a11 S1 C a12 S2
dt
dS2
D a21 S1 C a22 S2
dt
As weve seen already, such a two dimensional linear system of differential equations has solutions of the form:
S1 D c1 k1 e 1 t C c2 k2 e 2 t
S2 D c3 k3 e 3 t C c4 k4 e 4 t
That is, a sum of exponential terms. The ci and ki terms are constants related to
the initial conditions and eigenvectors, respectively, but the i terms or eigenvalues
determine the qualitative pattern that a given behavior might have. It should be
noted that the eigenvalues can be complex or real numbers. In applied mathematics,
e raised to a complex number immediately suggests some kind of periodic behavior.
Let us consider different possibilities fort he various eigenvalues.
311
Concentration
S1
S2
0.5
0
0.5
1
0
Time
Figure 17.3 Trajectories for a two species reaction network. On the left is the phase
plot and on the right, a single transient as a function of time. This system illustrates
a stable node corresponding to Negative Eigenvalues in the Jacobian. Matrix A:
a11 D 2; a12 D 0; a21 D 0:15; a22 D 2. Corresponding eigenvalues: 1 D
2; 2 D 2. The symmetry in the trajectories is due to eigenvalues of the same
magnitude.
Both Eigenvalues have the same sign, different magnitude but are real. If
both eigenvalues are negative, the equations describe a system known as a stable
node. All trajectories move towards the steady state point. If the eigenvalues have
the same magnitude and the ci terms have the same magnitude, the trajectories
move to the steady state in a symmetric manner as shown in Figure 17.3. If the ki
values differ, the trajectories will tend to twist as seen in Figure 17.4. They move
towards the steady state rather than away.
If the two eigenvalues are both positive, the trajectories move out from the steady
state reflecting the fact that the system is unstable. Such a point is called an unstable node. If the two eigenvalues have different magnitudes but are still positive,
the trajectories twist as shown in Figure 17.4.
Real Eigenvalues but of opposite sign. If the two eigenvalues are real but of
opposite sign, we see behavior called a saddle-node shown in Figure 17.5. This
is where the trajectories move towards the steady state in one direction, called the
stable manifold, and form a stable ridge. In all other directions trajectories move
away, resulting in an unstable manifold. Since trajectories can only move towards
the steady state if they are exactly on the stable ridge, the saddle nodes are generally
considered unstable.
Complex Eigenvalues. Sometimes the eigenvalues can be complex, that is of the
form a C i b where i is the imaginary number. It may seem strange that the solution to a differential equation that describes a physical system can admit complex
312
Variables
S1
S2
0
1
2
Time
Figure 17.4 Trajectories for a two species reaction network. On the left is the phase
plot and on the right a single transient as a function of time. This system illustrates a
unstable node, also called an improper node corresponding to Positive Eigenvalues.
Matrix A: a11 D 1:2; a12 D 2; a21 D 0:05; a22 D 1:35. Corresponding
eigenvalues: 1 D 1:6; 2 D 0:95.
eigenvalues. To understand what this means we must recall Eulers formula:
e i D cos./ C i sin./
Extended to:
e .aCbi /t D e at cos.bt/ C ie bt sin.bt/
(17.2)
When the solutions are expressed in sums of sine and cosine terms, the imaginary
parts cancel out, leaving just trigonometric terms with real parts (The proof is provided at the end of the chapter in an appendix). This means that systems with
complex eigenvalues show periodic behavior.
Description
Eigenvalues
Behavior
Both Positive
Both Negative
Positive and Negative
Complex Conjugate
Complex Conjugate
Pure Imaginary
r1
r1
r1
r1
r1
r1
Unstable
Stable
Saddle point
Unstable spiral
Stable spiral
Center
> r2 > 0
< r2 < 0
< 0 < r1
> r1 > 0
< r1 < 0
D r1 D 0
313
S1
S2
Variables
0
2
4
0
0.5
1.5
Time
2.5
Figure 17.5 Trajectories for a two species reaction network. On the left is the phase
plot and on the right a single transient as a function of time. This system illustrates
a saddle node corresponding to One Positive and One Negative Eigenvalue. Matrix
A: a11 D 2; a12 D 1; a21 D 1; a22 D 2. Corresponding eigenvalues: 1 D
1:73; 2 D 1:73.
Figures 17.6, 17.7, and 17.8 show typical trajectories when the system admits complex eigenvalues. If the real parts are positive, the spiral trajectories move outwards
away from the steady state. Such systems are unstable. In a pure linear system,
the trajectories will expand out forever. They will only stop and converge to a stable oscillation if the system has nonlinear elements which limits the expansion. In
these cases we observe limit cycle behavior.
If the real parts of the eigenvalues are negative, the spiral trajectory moves into the
steady state and is therefore considered stable.
314
S1
S2
Variables
2
0
2
4
0
4
Time
Figure 17.6 Trajectories for a two species reaction network. On the left is the phase
plot and on the right, a single transient as a function of time. This system illustrates
a stable spiral node corresponding to Negative Complex Eigenvalues. Matrix A:
a11 D 0:5; a12 D 1; a21 D 1; a22 D 1. Corresponding eigenvalues: 1 D
0:75 C 0:97i; 2 D 0:75 0:97i .
Conjugate Pair
A complex conjugate pair is a complex number of the form: a bi. The
eigenvalues for a two variable linear system with matrix A can be computed
directly using the relation:
p
tr .A/ tr2 .A/ 4 det .A/
D
2
where tr .A/ D a Cd , and det .A/ D ad bc. If the term in the square root is
negative, the eigenvalues will always come out as a conjugate pair owning to
the term. If tr2 .A/ 4 det.A/ < 0, then the solution will be the conjugate
pair:
p
tr .A/
tr2 .A/ 4 det.A/
D
2
2
Therefore a complex eigenvalue will always be accompanied by its conjugate
partner.
315
10
Variables
S1
S2
0
5
10
10
Time
15
20
25
Figure 17.7 Phase portrait for the two species reaction network. Unstable spiral
node. Positive Complex Eigenvalues. Matrix A: a11 D 0; a12 D 1:0; a21 D
1:2; a22 D 0:2. Corresponding eigenvalues: 1 D 0:1 C 1:09i; 2 D 0:1 1:09i .
k2 S1
(17.3)
(17.4)
We can plot the steady state value of S1 as a function of k2 as shown in Figure 17.10.
This isnt a particularly interesting bifurcation plot however and misses one of the
most important characteristics.
Equation (17.4) shows that the simple system (17.3) only has one steady state for
a given set of parameters. That is, if we set values to Xo ; k1 , and k2 , we find there
is only one value of S1 that satisfies these parameter settings. This is what Figure 17.10 also demoonstrates. What is more interesting is when a system admits
multiple possible steady state values for a given set of parameter values. To illustrate this behavior, let us look at a common system that can admit three possible
steady states. It is in these cases that bifurcation plots become particularly useful
and more interesting.
316
S1
S2
Variables
1
0
1
2
0
10
Time
15
20
25
Figure 17.8 Phase portrait for the two species reaction network. Center node.
Complex Eigenvalues, Zero Real Part. Matrix A: a11 D 1; a12 D 2:0; a21 D
2; a22 D 1. Corresponding eigenvalues: 1 D 0 C 1:76i; 2 D 0 1:76i .
Bistable Systems
Bifurcation plots can be useful for identifying changes in qualitative behavior, particularly systems that have multiple steady states. Consider the system shown Figure (17.11). This shows a gene circuit with a positive feedback loop. As the transcription factor x accumulates, it binds to an operator site upstream of the gene
which increases its synthesis. The more transcription factor made, the higher the
rate of expression.
At first glance this would seem to be a very unstable situation. One might imagine that the transcription factor concentration would continue to increase without
limit. However, the physical limits, in this case the saturation of the translation,
transcription, and degradation machinery, ultimately limits the upper value for the
concentration of transcription factor. To investigate the properties of this networks
we will construct a simple model. This model uses the following kinetic laws for
the synthesis and degradation steps:
v1 D b C k1
x4
k2 C x 4
v2 D k3 x
v1 is a Hill like equation with a Hill coefficient of four and a basal rate of b. v2 is a
simple irreversible mass-action rate law. The differential equation for the model is:
dx
D v1
dt
v2
A: stable node
317
B: stable focus
Im
Im
Im
Re
C: saddle point
Re
D: unstable focus
Im
Im
Re
E: unstable node
Re
Re
Dynamics
Eigenvalues
Unstable States
Stable States
318
Concentration of S1
1
0:8
0:6
0:4
0:2
0
10
15
20
k2
25
30
35
40
v2
v1
Figure 17.11 System with Positive Feedback.
v1 > v2 which means that the rate of change of x is positive. Since it is positive, the
perturbation, instead of falling back, continues to grow until x reaches the upper
steady state. We conclude that the middle steady state is unstable. This system
possess three steady states, one unstable and two stable. Such a system is known as
a bistable system because it can rest in one of two stable states but not the third.
Another way to observe the different steady states is to run a time course simulation
at many different starting points. Figure 19.3 shows the plots generated using the
script in Listing 19.1. The plots show two steady states, a high state at around 40,
and a low state at around 3. Notice that there is no third state observed. As we have
already discussed, the middle steady state is unstable, and all trajectories diverge
from this point. It is therefore not possible when doing a time course simulation to
observe an unstable steady state since there is no way to reach it.1
1
The author has been reliably informed that running time backwards in a time course simulation
will cause the simulation to converge on the unstable steady state. The author has not tried this
himself, however.
319
1
v1
v1 and v2
0:8
0:6
v2
0:4
0:2
0
0:2
0:4
0:6
0:8
x
1:2
1:4
320
k3 D 0:26
2
x
k3 D 0:7
4 0
0:5
1:5 0
k3 D 1:2
0:5
1:5
p.ss.eval;
321
v1 > v2
dx=dt > 0
v1 and v2
0:8
v1
v2
0:6
v2 > v1
dx=dt < 0
0:4
0:2
0
x
0:2 0:4
0:6
0:8
x
1:2
1:4
Figure 17.14 A graphical understanding of the stability of the steady states. See
text for details. Computed using the SBW rate law plotter. k1 D 0:9I k2 D
0:3I k3 D 0:7I b D 0:1.
Concentration, x
1:5
0:5
8
Time
10
12
14
Figure 17.15 Time course data generated from Jarnac model 19.1. Each line represents a different initial concentration for x. Some trajectories transition to the low
state while others to the upper state.
322
Kinetic Order
Elasticity
First-Order
Zero-Order
Sigmoidal
1.0
0.0
> 1:0
Table 17.2
ential equation for this system is:
dx
D v1 .x/
dt
v2 .x/
where we have explicitly shown that each reaction rate is a function of x. To determine whether the system is stable to small perturbations we can differentiate the
equation with respect to x to form the Jacobian. Notice there is only one element
in the Jacobian because we only have one state variable:
@v1
dx=dt
D
dx
@x
@v2
@x
The terms on the right are unscaled elasticities (??). If the expression is positive,
the system is unstable because it means that dx=dt is increasing if we increase x.
We can scale both sides to yield:
Js D "1x
"2x
where the right-hand term now includes the scaled elasticities (??). The criteria for
stability is again that "1x > "2x . Therefore, if the positive feedback is stronger than
the effect of x on the degradation step v2 , the system will be unstable.
Recall that the elasticities are a measure of the kinetic order of the reaction. Thus an
elasticity of one means the reaction is first-order. A saturable irreversible MichaelisMenten reaction will have a variable kinetic order between one and zero (near saturation). A Hill equation can, depending on the Hill coefficient, have kinetic orders
greater than one (Table 17.2). Knowing this information, there are at least two
ways to make sure that the elasticity for the feedback elasticity, v1 , is greater than
the elasticity for the degradation step, v2 :
1. v1 is modeled using a Hill equation with a Hill coefficient > 1 and v2 is firstorder or less.
323
Jacobian: .dx=dt/=dx
Elasticity, "vx1
0.145
0.683
1.309
-0.664
0.585
-0.47
0.052
1.835
0.33
Table 17.3 Table of steady state values of x and corresponding values for the
Jacobian element. Negative Jacobian values indicate a stable steady state, positive
elements indicate an unstable steady state. The table shows one stable and two
unstable steady states.
The unstable steady state at x D 0:683 has an elasticity for v1 of 1.835. Note this
value is greater than the elasticity of the first-order degradation reaction, v2 , which
equals one. Therefore this state is unstable.
Bifurcation Plot
Lets now return to the question of plotting a bifurcation graph for the bistable system in Figure 17.11. Figure 17.13 shows both reaction rates, v1 and v2 plotted
as a function of the intermediate species x. In this figure we see three intersection points, marking the three possible steady states. By varying the degradation
constant k3 , we can change the behavior of the system so that it exhibits a single
high steady state, three separate steady states, or a single low steady state (See
Figure 17.12).
If we track the intersection points as we vary the value of the rate constant k3 , we
obtain the bifurcation plot shown in Figure 17.16.
Figure 17.16 shows that at some value of the parameter k3 , the system has three
possible steady states, outside this range only a single steady state persists. Bifurcation diagrams are extremely useful for uncovering and displaying such information.
Drawing bifurcation diagrams is not easy, however. and there are some software
tools that can help. Figure 17.16 for example was generated using the SBW Auto
324
Concentration, x
3
4
1
3
0
0:2
0:4
2
0:6
0:8
k3
1:2
1:4
http://jdesigner.sourceforge.net/Site/Auto_C.html
http://oscill8.sourceforge.net/
Concentration, x
325
0:5
1
Parameter, k3
1:5
0:5
1
Parameter, k3
1:5
Irreversible Bistability
It is possible to design an irreversible bistable switch. Figure 17.19 shows the
bifurcation plot for such a system. This is modified from the Mutual activation
model in the review by Tyson [52], Figure 1e. In this example increasing the signal
results in the system switching to the high state at around 2.0. If we reduce the
signal from a high level, we traverse the top arc. If we assume the signal can never
be negative, we will remain at the high steady state even if the signal is reduced
to zero. The bifurcation plot in the negative quadrant of the graph is physically
inaccessible. This means it is not possible to transition to the low steady state by
decreasing signal. As a result, the bistable system is irreversible, that is, once it is
switched on, it will always remain on.
p = defn
$X
R1
EP
E
end;
cell
-> R1;
-> $w;
-> E;
-> EP;
k1*EP + k2*Signal;
k3*R1;
Vm1*EP/(Km + EP);
((Vm2+R1)*E)/(Km + E);
326
Signal
v1
R1
v2
v3
EP
v4
Figure 17.18 System with Positive Feedback using a covalent modification cycle,
E; EP .
The Jarnac script for the model is shown in listing 17.3. To create Figure 17.19,
first install Oscill84 , then launch Jarnac. Load the script into Jarnac. Run the script
(press green button in toolbar) to put the model into memory. Go to the SBW menu
and select Oscill8. Once in Oscill8, select Run; 1 Parameter. In the new dialog
box select continuation and the parameter Signal. Then select run to view the
bifurcation plot.
Toggle Switch
The final example we will consider is the toggle switch. This example illustrates
the idea of an attractor basin in a phase plot. The system in question is shown in
Figure 17.20 and is comprised of just two nodes, S1 and S2 . Each node inhibits
the other. This is a high level diagram, but mechanistic realizations can be made
using protein or gene regulatory networks. In fact, one of the first synthetic biology constructs was the toggle switch made from an engineered gene regulatory
network [18]. Intuitively, one can imagine the type of behavior this system might
4
http://oscill8.sourceforge.net/
327
Concentration, R1
20
15
10
5
0
Signal
Figure 17.19 Bifurcation diagram for species R1 with respect to the signal. Signal from the model shown in Jarnac script 17.3. The continuous line represents
stable steady state points, the dotted line the unstable steady states. Plotted using
Oscil8 http://oscill8.sourceforge.net/.
S1
S2
Figure 17.20 The toggle switch. Two mutually inhibited nodes, S1 and S2 .
exhibit. For example, if S1 has a high concentration then this will repress S2 . Since
S2 is now low, repression of S1 is negligible. This state appears stable. Alternatively we could imagine that S2 is at a high concentration. This will repress S1 and
thus we have another state that appears stable.
We can better study the behavior of the toggle switch by building a computer simulation. Using the set of differential equations shown in equations (17.5), we describe the model as:
dS1
k1
D
k2 S1
dt
1 C S2n1
(17.5)
dS2
k3
D
k4 S2
dt
1 C S1n2
A suitable set of parameter values are given by k1 D 6I k3 D 6I k2 D 2I k4 D 2.
Figure 17.22 shows the phase plot for this system. The three points marked by
328
S1
S2
Figure 17.21 The toggle switch constructed from a gene regulatory network.
round circles represent the steady state locations. Note that there is a high S1 /low
S2 state, and a low S1 /high S2 state. These correspond to the states we intuitively
described before. There is also a third point which represents a medium level of
S1 and S2 . The arrows in the diagram represent the direction of change of S1 and
S2 in time. The diagram has been subdivided into four basins of attraction. For
example, the phase plot highlights one of the direction arrows in the basin labelled
two. The arrows initially point right to left and down. This tells us that if we started
a time course simulation at the first arrow, the concentration of S1 and S2 would
both decline. Initially, the arrow points roughly in the direction of the unstable state
but as it moves, its direction changes and eventually moves towards the upper stable
state to the left.
The phase plot gives us a condensed view of how any initial condition will evolve.
Points that start in basin one will converge to the third state, basin two points converge on the first state, basin three to the third state, and finally basin four to the
first state.
Nullclines
The toggle switch model from the previous section also allows us to introduce the
concept of nullclines. Weve seen various mechanisms for visualizing a system,
especially related to its steady state stability. For example, in Figure ?? we show
how the stability of a one dimensional system can be visualized, likewise for Figure 17.12, the bifurcation plot in Figure 17.16, and the phase plot in Figure 17.22.
There is yet another plot that can help us understand the instability of a two dimensional system and that has to do with plotting the nullclines. The nullcline is the
solution to a differential equation when set to zero as a function of the two system
variables. For example, the toggle switch differential equation for S1 is:
k1
dS1
D
dt
1 C S2n1
k2 S1
329
10
S2
3
1
S1
10
Figure 17.22 Phase portrait for a toggle switch, equation (17.5). The four numbered areas mark the four basins of attraction. The line of thicker arrows illustrates
one possible time course trajectory given an initial starting point. Diagram generated by PPlane http://math.rice.edu/~dfield/dfpp.html, k1 D 6I k3 D
6I k2 D 2I k4 D 2.
We can set this to zero and find all combinations of S1 and S2 that satisfy this
equation. These points form a line on a two dimensional plane. Such a line is
called the nullcline. Figure 17.23 shows two nullclines corresponding to the two
differential equations for the toggle model. Note that where the nullclines intersect,
we find the steady state because at these points, both equations yield the same
values for S1 and S2 .
Although useful, many of the plots we have reviewed are limited to systems of two
or three variables. The nullclines in an n-dimensional system cannot easily be visualized. Likewise, a two dimensional phase plot can only take a slice through the
dynamics of a system with many variables. Nevertheless, these techniques, in particular bifurcation plots, are extremely useful in delineating the potential behavioral
modes of networks. See [8, 9] for examples of bifurcation plots in studying protein
signaling pathways.
330
S2
dS2/dt=0
dS1/dt=0
S1
10
Figure 17.23 Phase portrait for a toggle switch. The two nullclines are superimposed on the phase plot and labelled with arrows. Diagram generated by
PPlane http://math.rice.edu/~dfield/dfpp.html. See Figure 17.22 for parameter values.
Routh-Hurwitz
The first point to make is that for a polynomial such as:
a0 s n C a1 s n
C : : : C an 1 s C an D 0
if any of the coefficients is zero or negative then there is at least one root with
331
a2
a3
b2
c2
a4
a5
b3
c3
a1 a2
a0 a3
a1
b 1 a3
b2 a1
b1
b2 D
c2 D
a1 a4
a0 a5
etc.
b3 a1
etc:
a1
b 1 a5
b1
Subsequent rows are computed in a similar manner from the immediate rows above.
The process is continued until the nth row is completed. rows Stability is then
332
determined by the number of sign changes in the 1st column which is equal to the
number of poles with real parts greater than zero.
A system is stable if and only if all the entries of the first column of the Routh
table are of the same sign.
Consider a quadratic polynomial, ax 2 C bx C c D 0.
Table (17.5) shows the Routh table for the characteristic equation s 2 C 3s C 1 D 0.
a0
a1
.a1 a2
a2
0
a0 a3 /=a1
1
3
1
1
0
1
16
0
Further Reading
1. Edelstein-Keshet L (2005) Mathematical Model sin Biology. SIAM Classical
In Applied Mathematics. ISBN-10: 0-89871-554-7
333
2. Fall CP, Marland ES, Wagner JM, Tyson JJ (2000) Computational Cell Biology. Springer: Interdisciplinary Applied Mathematics. ISBN 0-387-95369-8
3. Steuer R and Junker BH (2009). Computational models of metabolism: stability and regulation in metabolic networks. Advances in chemical physics,
142, 105.
Exercises
1. Determine the Jacobian matrix for the following systems:
a)
dx
D x2
dt
b)
dx
Dy
dt
y2
xy
dy
D x.1
dt
y/
dy
D xy
dt
2. Compute the steady state solutions to the two systems in the previous question.
3. Determine the stability of the solutions from the previous question.
4. Determine the Jacobian in terms of the elasticities and stoichiometry matrix
for the following systems:
a) Xo ! S1 I S1 ! S2 I S2 ! S3 I S3 ! X1
b) Xo ! S1 I S1 ! S2 I S2 ! S1 I S2 ! X1
c) S1 ! S2 I S2 ! S3
Assume all reactions are product insensitive, Xi species are fixed, and in c)
S3 regulates the first step.
5. Show that the following system is stable to perturbations in S1 and S2 by
computing the eigenvalues at steady state (See Listing ??):
Xo ! S1 ! S2 ! X1
334
Vm1 Xo
Km1 C Xo C S1 =Ki
v2 D
Vm2 S2
Km2 C S1 C S2 =Kj
v2 D
Vm3 S3
Km3 C S2
7. Show that the following system is unstable. What kind of unstable dynamics
does it have?
p = defn cell
var R, X, E, EP;
ext src, S, Waste;
J1: $src -> X;
k1*S;
J2:
J3:
J4:
J5:
X
R
E
EP
->
->
->
->
335
R;
(kop + ko*EP)*X;
$waste; k2*R;
EP;
Vmax_1*R*E/(Km_1 + E);
E;
Vmax_2*EP/(Km_2 + EP);
end;
p.src = 0;
p.kop = 0.01;
p.ko = 0.4; p.k1 = 1;
p.k2 = 1;
p.R = 1;
p.EP = 1;
p.S = 0.2;
p.Km_1 = 0.05; p.Km_2 = 0.05;
p.Vmax_2 = 0.3; p.Vmax_1 = 1;
p.KS4 = 0.5;
8. Using Figure 17.23, show graphically that the toggle switch model has two
stable and one unstable steady states.
336
18
Negative Feedback
18.1 Introduction
Biological networks are riddled with regulatory interactions. These include positive and negative, forward and feedback. Chapter 3 described some of the basic
properties of negative feedback. Properties such as disturbance rejection and improvements to performances. In this chapter we will consider more closely the
properties of systems with negative feedback regulation. Negative feedback can
occur in many different ways in biochemical networks. Some is quite overt, for
example an allosteric regulator. Others can be more subtle, for example by simple
sequestration in a signaling network. Some of the more important problems in the
analysis of negative feedback include:
The Transient Response
Stability
Accuracy
Sensitivity
Linearization
337
338
Error (e)
Kp
eK P
Process
Output
Feedback
Integral Control
In integral control the signal that feeds the process block, A, is the integral of the
error signal, e. In practice that this means is that a running total of the error signal
is maintained and it is this total that is fed to the process block. Mathematically we
can express this as:
Z
u D KI
e.t/dt
Integral control is a very powerful control strategy for removing any difference
between the set point and the output. So long as there is some difference between
the set point and the output, the integral control will accumulate the error forcing the
process, A to match the set point. Once the set point and output are equal, the error
goes to zero and the integration stops. This behavior is in contrast to proportional
control where there will always be a small difference between the output and the
set point. It is possible to close the gap by increasing the Kp constant but, as we
will see, at the expense of instability.
339
Derivative Control
The third type of control is called derivative control. What this does is feed that
rate of change of the error into the process block A. In practice this can be simply
achieved by sampling the error at two different times and passing on the difference
to the process block. Mathematically we can express this using the equation:
u D KD
de.t/
dt
The advantage of using derivative control is that it can slow down fast changes and
so help prevent overshoot and undershoot in the output. For example, assume the
set point is set to 5 units. If the feedback returns 10 units indicating that the output
is above the set point, then the error will be 5 10 D 5, ie the output should be
reduced. Assume that a moment later, the output rises to 25 units, as a result the
error changes to 5 25 D 20. The change in the error per unit time is therefore
-20 - (-5) = -15 units per time. The negative sign indicates that the process should
slow down the rate at which the output is rising. One disadvantage of derivative
control is that computing derivatives tends to amplify any noise in the signals so
that is the noise and gain of the derivative control is sufficiently high, the output
of the system can be swamped by noise and in the worse case become unstable.
Another disadvantage is that a derivative controller puts a damper on the response
time so that the system will responds slower to changes.
Kp e(t)
Proportional
Set Point
+
{
Ki
t
0
e()d
Integral
Process
Kd de(t)
dt
Derivative
340
x
v2
v1
k2 x
where the function v1 indicates that the rate is a function of x1 as well as the input,
I . We can linearize this equation using 12.2 to obtain:
d x
@v1 @v2
@v1
D
x C
I
dt
@x
@x
@I
The solution to this linear differential equation is:
x.t/ D xe t C
@v1
@I k2
I
1
@v1 =@x
e t
341
where:
@v1 @v2
@x
@x
Note that is net negative because @v1 =@x, the feedback term, is always negative.
The exponential terms represent the transient response and depend on the feedback
term, @v1 =@x. The large the absolute value of @v1 =@x the faster the decay rate.
The conclusion therefore is that negative feedback can increase the response of the
system.
D
v1
x
v2
@v
@p
342
Entering these terms into the transfer function equation gives us:
H.s/ D
s
@v1
@I
@v2
@v1
C
@x
@x
This equation has the same structure as equation 13.3. Equation 13.4 is the amplitude frequency response. In this case the k2 term in equation 13.4 is now the sum
of two unscaled elasticities. Since @v1 =@x is negative the sum is overall positive.
k1
Amplitude D jH .j!/j D s
@v1
@v2 2
C !2
.
C
@x
@x
(18.1)
This equation is plotted as a Bode plot in Figure 18.4. Two curves are shown,
the continuous line represents the system with weak negative feedback and the
dotted line is the same system with strong negative feedback. The input signal
has been adjusted in each case so that both systems have the same amplitude at
zero frequency. If this werent done, the strong negative feedback line would be
lower due to attenuation and it would be more difficult to see the extension in
the bandwidth. The result however is that with negative feedback, the frequency
response, or bandwidth, improves. This is consistent with the previous result which
showed an increase in the transient response.
Negative feedback improves the frequency response of the system.
18.4 Stability
Another key property of negative feedback systems is their propensity to become
unstable. In terms of the frequency response is it straight forward to understand
the origins of this instability. In a negative feedback system, most disturbances are
damped due to the action of the feedback. However what if the feedback mechanism takes too long to respond so that by the time the negative feedback acts, the
disturbance has already abated. In such a situation, the feedback would now attempt to restore a disturbance that is not longer present resulting in an incorrect
action. Imagine that the feedback system acts in the opposite direction to the disturbance because of the delay, this would cause the disturbance to grow rather then
18.4. STABILITY
343
Magnitude (dB)
10
0
10
20
30
10
10
10
100
101
102
103
104
344
Log Magnitude
of Loop Gain
unity at the frequency where the phase angle is 180o and the phase margin is
the difference between the phase of the response and 180o when the loop gain is
1.0, see Figure 18.5. Both terms can be used to measure the relative stability of
a negative feedback system. Both however must be measured with respect to the
loop gain and not the closed loop gain.
Positive Gain
Margin
Gain of one
crossover point
Phase
180
Log Magnitude
of Loop Gain
Stable System
Negative gain
margin
{
180 degree
crossover point
Phase
180
Unstable System
Negative phase
margin
18.4. STABILITY
345
To reduce the chance of unsustained oscillations a rule of thumb is that there should
be a 45 to 60 phase margin. When the loop gain is one (crosses the y axis at zero),
the phase must be less than 180 . If the gain is one or larger than one and at the
same time the phase is at 180 this is a recipe for instability. The reason for this
is that the inverting feedback also contributes a 180 shift, this means that the total
shift is 360 . A 360 shift turns the negative feedback into a positive feedback
response. That is, a perturbation that increases with a frequency component of
180 will cause the response of the system to further increase. This is of course
an unstable situation. Most systems will see signals with a frequency component
of 180 because they are subject to random noise which is a wide spectrum signal.
Note that without noise, even an unstable system can sit precariously at the unstable
focus.
The phase margin is the amount the phase angle exceeds 180 degrees when
the loop gain is one.
The gain margin is the size of the loop gain when the phase angle is
degrees.
180
x1
v1
v2
x2
v3
v4
346
To determine the stability of the system we only need to look at the characteristic
equation which can be computed from determinant of:
@v 1
sI N
@x
We will assume that the degradation steps, v2 and v3 , are first order. In addition
because the pathway is not branched, the unscaled elasticities translate directly to
the scaled elasticities. This means that the degradation scaled elasticities are one.
We can write down N and the elasticity matrix as:
1
1
0
0
ND
0
0
1
1
Note that entry .1; 3/ is zero because v2 does not consume x1 . @v=@x is a .4; 1/
matrix corresponding to 4 reactions rates and 2 species, x1 and x2 . Here we use the
symbol
mat hcalEji to represent an unscaled elasticity, that is Eji D @vi =xj , for example
Exv21 is the feedback elasticity.
2
3
0 Exv21
6
7
6 v2
7
7
6Ex
0
1
7
@v 6
6
7
D 6 v3
7
@x 6Ex1 0 7
6
7
6
7
v
4 0 E 45
x2
We will now replace each unscaled elasticity, E with the equivalent term involving
the scaled elasticity, ". We do this by substituting each unscaled elasticity with the
term:
vi
Eji D "ji
xi
If we assume that the following reactions are first-order, v2 , v3 and v4 then the
corresponding elasticities, "21 , "31 , and "42 can be set to one. This means that the
@v=@x matrix can be written as:
2
3
0 "vx12 F1
6
7
6
7
6F2
0 7
6
7
@v 6
7
D6
7
0 7
@x 6F3
6
7
6
7
40
F4 5
18.4. STABILITY
347
where Fi are the scaling factors, vi =xi . From these matrices the determinant and
therefore the characteristic equation can be shown to be:
s 2 C s.F2 C F4 / C F2 F4
F1 F3 "12 D 0
The criterion for negative roots (and hence stability) in a quadratic equation is that
all the coefficients have the same sign. We should note that all Fi factors are positive therefore is should be clear that the first two coefficients (attached to s 2 and
s) are positive. The right most coefficient is F1 F3 "32 where the elasticity "vx32 is
negative because it represents the negative feedback. The expression, F1 F3 "32 is
therefore positive overall. Since all the coefficients are positive and referring to the
rule for quadratic equations (Appendix E) the system must be stable. In fact, unless
the negative feedback changes to a positive feedback loop, there is no possibility of
this system ever showing instability.
Now let us look at a slightly modified system shown in Figure 20.2 where we have
added a second step in the feedback loop.
v1
x1
v2
v3
x2
v4
v5
v6
x3
348
vi
xi
Multiplying out the matrices and assuming first-order kinetics except for the feedback mechanism (meaning all elasticities other than the feedback elasticity equal
one) and computing the determinant yields the following characteristic equation:
s 3 C s 2 .F2 C F4 C F6 / C s.F2 F4 C F2 F6 C F4 F6 /
F1 F3 F5 "vx13 D 0
C F2 F4 F6
Since reactions rates v1 and v2 are equal at steady state it is also true that F1 D F2 .
The same arguments apply to F3 D F4 and F5 F6 . This changes the characteristic
equation to:
s 3 C s 2 .F2 C F4 C F6 / C s.F2 F4 C F2 F6 C F4 F6 /
C F2 F4 F6 .1
"vx13 / D 0
As expected this is a cubic equation. For a general cubic equation such as:
a0 x 3 C a1 x 2 C a2 x C a3 D 0
the condition for all negative roots is that:
a0 > 0; a1 > 0; a3 > 0 and
a1 a2 > a0 a3
Negative roots of course indicate stability. We note that the first three conditions are
satisfied because the first two coefficients are positive (All Fi terms are positive)
and the expression .1 "vx13 / is net positive (Note "vx13 < 0). With these conditions
satisfied the requirement for stability (a1 a2 > a0 a3 ) becomes:
.F2 C F4 C F6 /.F2 F4 C F2 F6 C F4 F6 / > F2 F4 F6 .1
"vx13 /
"vx13
(18.2)
18.4. STABILITY
349
Frequency
600
400
200
10
15 20 25 30
Stability Threshold
35
40
Figure 18.8 Stability Threshold for equation 18.2. Only feedback elasticities
greater than eight result in instability.
This is the condition for stability. So long as the left hand side is greater than
the feedback elasticity, "vx13 , the system will be stable. The question now is what
values can the left-hand side have? One way to pin down the inequality further is
to try all combinations of Fi to discover what range of values the left-hand side
can yield. Figure 18.8 shows combinations of Fi terms and what is clear from the
graph is that no combination falls below 8. Eight is therefore a lower bound for
the left-hand term. This means that the feedback elasticity must be less than 8 in
order to achieve stability. Values above 8 (or more precisely less than -8) will cause
instability.
Another way to look at equation 18.2 is to recall the relation:
1
Ca >2
a
Looking more closely at 18.2 we see that each pair of terms is in the form .1=a/Ca.
Therefore each pair of terms must be > 2. The lowest value each term can have
is 2, therefore the minimum value for the sum of three of the paired terms must be
6. With the addition of the remaining 2 in the equation means that the minimum
value for the left-hand side of equation 18.2 is 8. This matches the result from the
numerical study 18.8. We can therefore summarize that if
"vx13 < 8
then the system is stable. For example if "vx13 D 2:0 then the system will be stable.
If the strength of the negative feedback is more than j 8j then the system will be
350
Length of Gene
Cascade
Instability
Threshold
"feedback
1
2
3
4
5
6
7
::
:
stable
stable
8.0
4.0
2.9
2.4
2.1
::
:
1.0
Table 18.1 Relationship between the gene cascade length and the degree of feedback inhibition on the threshold for stability. "feedback is the elasticity of the feedback inhibition. The analysis assumes first order kinetics for the degradation steps
and the effect of other transcription factors on gene expression.
unstable. For example if "vx13 D 10 then the system will be instable. This means
that instability in this system requires a fairly strong degree of feedback. A similar
analysis can be done on pathways of other lengths. Table 18.1 summarizes the
threshold values for the feedback elasticity for various lengths of gene cascade.
From this analysis we can also make a few more observations. We have assumed
that the elasticities for the effect of x1 on v3 and x2 on v5 was first order, It is more
likely however that these values are, if anything, greater than one since we know
that many transcription factors show cooperativity with respect to gene expression.
If we reintroduce these factors we find that the stability expression becomes:
8>
If all three elasticities are greater than one, then the threshold for instability is in
fact lower. For example, if "vx31 and "vx52 have values of two each, then the threshold
equation becomes:
8 > 4"vx13
That is:
8 < 4"vx13
2 < "vx13
18.5. ACCURACY
351
The threshold for instability is now much lower. The opposite is also true if the
transcription factor binding is non-cooperative (elasticities < 1) or that the steps
within the transcriptional and translational machinery results in the elasticities having values less than one.
In designing a stable negative feedback system we must enure two things:
The feedback elasticity should not too strong (Threshold dependent on the
length of the pathway)
The number of steps between the feedback signal and its site of action should
be as small as is possible
18.5 Accuracy
The accuracy of a feedback system is the difference between the set point and
output of the system at steady state. The closer the output is to the set point the
more accurate the system is. Figure ?? shows a block diagram of a generic feedback
system with components expressed as transforms.
R(s)
E(s)
KG(s)
Y(s)
F(s)
KG.s/F .s/
1 C KG.s/F .s/
Y.s/F .s/
352
ess D lim
The term, KG.0/F .0/ is the loop gain and is a constant which we can designate,
Kp , so that:
1
ess D
1 C Kp
It should be clear from this final result that the error can be made smaller by increasing the loop gain although the error can only be eliminated if the gain is infinite.
The higher the gain the smaller the error. However as we saw in the last section,
the stronger the feedback the more likely the system will become unstable. There
is therefore a trade off between accuracy and stability when designing feedback
systems.
18.6 Linearization
We will consider only the steady-state behavior of the system. We take the input u,
the output y, and the error e to be constant scalars. Assume (for now), that both the
amplifier A and the feedback F act by multiplication (take A and F as non-negative
scalars). Then without feedback (i.e. F D 0), the system behavior is described by
y D Au, which is an amplifier (with gain A) provided that A > 1. Assuming for
now that the disturbance d is zero, and if we now include feedback in our analysis,
the behavior of the system is as follows. From the diagram, we have
y D Ae
eDu
Au
1 C AF
or simply y D Gu;
F y:
Eliminating e, we find
yD
18.6. LINEARIZATION
353
A
where G D 1CAF
is the system (or closed loop) gain. Comparing G with A, it is
immediate that the feedback does indeed reduce the gain of the amplifier. Further,
if the loop gain AF is large (AF 1), then
G
1
A
D :
AF
F
That is, as the gain AF increases, the system behavior becomes more dependent on
the feedback loop and less dependent on the rest of the system. We next indicate
three specific consequences of this key insight.
Improved fidelity of response Consider now the case where the amplifier A is
nonlinear. For example a cascade pathway exhibiting a sigmoid response. Then the
behavior of the system G (now also nonlinear) is described by
G.u/ D y D A.e/
eDu
Fy D u
F G.u/:
Differentiating we find
G 0 .u/ D A0 .u/
Eliminating
de
,
du
de
du
de
D1
du
F G 0 .u/:
we find
G 0 .u/ D
A0 .u/
:
1 C A0 .u/F
1
;
F
354
e
A
-
Ae
Summation Point
A = Amplifier
F = Feedback
18.7 Sensitivity
Resistance to internal parameter variation In all real amplifiers, both manmade and natural, there will be variation in the amplifier (A) characteristics, either
as a result of the manufacturing process or internally generated thermal noise. We
can study the effect of variation in the amplifier characteristics by investigating how
A causes variation in the gain G.
Considering the sensitivity of the system gain G to variation in the parameter A,
we find
@G
@
A
1
D
D
:
@A
@A 1 C AF
.1 C AF /2
Clearly, this sensitivity decreases as AF increases. It may be more telling to consider the relative sensitivity, in which case we find
@G A
1
D
;
@A G
1 C AF
355
so that for a small change A in the gain of the amplifier, we find the resulting
change G in the system gain satisfies
G
A
1
:
G
1 C AF A
As the strength of the feedback (F ) increases the influence of variation in A decreases.
Resistance to disturbances in the output Suppose now that a nonzero disturbance d affects the output as in Figure 18.10. The system behavior is then described
by
y D Ae
eDu
F y:
Eliminating e, we find
yD
Au d
:
1 C AF
1
:
1 C AF
Again, we see that the sensitivity decreases as the loop gain AF is increased. In
practical terms, this means that the imposition of a load on the output, for example a
current drain in an electronic circuit or protein sequestration on a signaling network,
will have less of an effect on the amplifier as the feedback strength increases. In
electronics this property essentially modularizes the network into functional modules.
356
X0
E1
x1
E2
x2
E3
X1
Figure 18.11 Three step pathway with negative feedback, Ei represent enzymes.
following transfer functions at zero frequency:
CEJ1 D
"21 "32
"21 "32 "21 "12
"21
"21 "32
"21 "12
CES22 D 0
CEJ2 D 0
CEJ3 D
CES12 D
"21 "12
"21 "32 "21 "12
CES32 D
"21
"21 "32 "21 "12
We have assumed that the reactions catalyzed by E1 and E2 are product insensitive
so that their elasticities are zero. From these equations we can make some interesting observations. Let us make the feedback strength, "12 strong, that is 0. In this
situation the transfer functions tend to the following limits:
CEJ1 !
"32
"12
CES12 !
1
"12
CEJ2 D 0
CES22 D 0
CEJ3 ! 1
CES32 !
1
"12
357
attempt to disturb these targets will be resisted by the feedback loop. Conversely,
targets up stream and particularly downstream are very susceptible to disturbance.
Figure 18.12 illustrates the effect of a 20 fold decrease in enzyme activity at two
points in a simple reaction chain. In the first case both disruption at the center or
end of the network has a significant effect on the concentration of the last species,
S3 . In the second panel, the same pathway is shown but with a negative feedback
loop from S3 to the first enzyme. The same activity reductions are also shown, but
this time note the almost insignificant effect that modulating a step inside the loop
has compared to modulating a step outside the loop.
Thus the take-home message to pharmaceutical companies who are looking for
suitable targets is to avoid targeting reaction steps inside feedback loops!
358
Exercises
1. A controller has the PI function given by:
t
Z
c.t/ D Kp e.t/ C Kp
e./d
0
where e.t / is the error between the set point and the feedback:
e.t/ D r.t/
y.t/
19
Positive Feedback
19.1 Basic Properties
Many biological systems show positive feedback regulation. Figure 19.1 shows a
simple gene regulatory network where the gene product activates its own production. We can examine the stability of this system by looking at the characteristic
equation of the system.
v2
v1
Figure 19.1 System with Positive Feedback
Let us look at a specific model:
v1
v2
k1 Xo .x C 1/3
D 5C
.x C 1/3 C 10000
D k2 x
359
360
The first equation, v1 , mimics a reaction that is activated by its product. In this case
the activation also shows a degree of sigmoidicity in the response. The steady state
of this simple model is computed at dx=dt D v1 v2 D 0 or v1 D v2 . If v1 and v2
are plotted against x (Figure 19.2), the points where the curves intersect correspond
to the steady states of the system. Inspection of Figure 19.2 shows three intersection
points and therefore this systems can display three possible steady states.
100
80
v
60
40
20
0
20
40
60
x1
Figure 19.2 v1 and v2 plotted against x concentration. Intersection points on curve
mark the steady state points.
We can also run this system use Jarnac. The script in listing 19.1 shows multiple
simulations carried out at different initial conditions.
p = defn cell
J1: $X -> x1; 5 + (k1*Xo*(x1+1)^3)/(((x1+1)^3)+10000);
x1 -> $w; k2*x1;
end;
p.k1 = 87.5;
p.k2 = 2;
p.Xo = 1;
p.x1 = 2;
m = p.sim.eval (0, 5, 100);
for i =1 to 10 do
begin
p.x1 = i*6;
mm = p.sim.eval (0, 5, 100, [<p.x1>]);
m = augc (m, mm);
end;
361
Concentration, x
60
40
20
Time
Figure 19.3 Time course data generated from Jarnac model 19.1. Each line represents a different initial concentration for x. Some trajectories transition to the low
state while others to the upper state.
graph (m);
Jacobian: .dx=dt/=dx
Elasticity, "vx1
2.72
15
41.07
-1.64
1.363
-1.349
0.18
1.681
0.326
Table 19.1 Table of steady state x and corresponding value for the Jacobian element. Negative Jacobian values indicate a stable steady state, positive elements
indicate an unstable steady state. The table shows one stable and two unstable
steady states.
This system possess three steady states, one unstable and two stable. Such a system
is known as a bistable system because it can rest in one of two stable states.
The stoichiometry matrix and elasticity matrix are now given by:
1
ND 1
362
" v #
"x1
@v
D v
@x
"x2
Inserting these terms into:
sI
@v
N
@x
Exv1 D 0
In order for the system to be stable the root, s, must be negative. Let us first convert
the unscaled elasticities (E), in to scaled versions. We do this by swapping each
unscaled elasticity by the term:
Eji D "ji
This gives us:
s C "vx2
v2
x
vi
xi
"vx1
v1
D0
x
Since at steady state v1 D v2 we can replace the terms v1 =x, v2 =x with the same
factor, F and the root is then (note the change in signs):
s D F "vx1 "vx2
For stability the root must be negative. Since F > 0, this is equivalent to stating
that:
"vx1 "vx2 < 0
for stability. Noting that "vx1 is positive, it must be that case that
"vx1 < "vx2
for the system to be stable. Conversely, if the positive feedback elasticity is larger
than the degradation elasticity then the system will be unstable. If we assume that
the degradation step, v2 is first-order then "vx2 D 1 and the condition for instability
will simply be:
Condition for instability: "vx1 > 1
363
This is easily achieved if the positive feedback displays cooperativity. Table 19.1
shows the elasticity for each of the steady states. Note that for the unstable state,
the elasticity is 1.68, a value greater than 1.
Other designs, eg covalent cycles, toggle switch, sequestration (http://www.nature.
com/msb/journal/v8/n1/full/msb201252.html),
"Self-perpetuating states in signal transduction: positive feedback, double-negative
feedback and bistability"
LacI bistability
Hysteresis
Real examples
364
20
Oscillators
20.1 Introduction
The study of oscillatory systems in biochemistry has a long history dating back to
at least to the 1950s. Until recently however, there was little interest in the topic
from mainstream molecular biology. In fact, one suspects that the concept of oscillatory behavior in cellular networks was considered more a curiosity than anything
serious. With the advent of new measurement technologies, particulary high quality microscopy, and the ability to monitor specific protein levels using GFP and
other fluorescence techniques a whole new world has opened up to many experimentalists. Of particular note is the recent discovery of oscillatory dynamics in the
p53/Mdm2 couple [30, 20] and Nf-B [25] signaling; thus rather than being a mere
curiosity, oscillatory behavior is in fact an important, though largely unexplained,
phenomenon in cells.
20.1.1
There are two basic kinds of oscillatory designs, one based on negative feedback
and a second based on a combination of negative and positive feedback. Both kinds
of oscillatory design have been found in biological systems. An excellent review of
these oscillators and specific biological examples can be found in [52, 13]. A more
365
366
X0
x1
x2
X1
x3
Figure 20.1 Simple negative feedback model with three variables, S1 , S2 and S3 .
This network can oscillate.
v1
x1
v2
v3
x2
v4
v5
v6
x3
Figure 20.2 Simple negative feedback model constructed from three genes. This
network can oscillate.
A critical factor that determines the onset of oscillations, apart from the number of
variables is the strength of the feedback. Savageau [45] showed that if the substrate
elasticities were equal (e.g. all first-order kinetics) then the ratio of the feedback
elasticity ("i nh ) to the output elasticity ("sub ; "43 ) determined the onset of oscillations (Table 20.1). Table 20.1 shows that as the pathway becomes longer less
feedback inhibition is required to destabilize the pathway. This highlights the other
factor that contributes to instability, the delay in routing the signal around the network. All feedback oscillators require some device to provide amplification of a
Length of
Pathway
367
Instability
Threshold
"i nh ="sub
1
2
3
4
5
6
7
::
:
stable
stable
8.0
4.0
2.9
2.4
2.1
::
:
1.0
Table 20.1 Relationship between the pathway length and the degree of feedback
inhibition on the threshold for stability. "i nh is the elasticity of the feedback inhibition and "sub is the elasticity of the distal step with respect to the signal.
signal combined with a suitable time delay so that the signal response can go out
of phase. In metabolic pathways amplification is often provided by a cooperative
enzyme while the delay is provided by the intermediate steps in the pathway. In signalling pathways, amplification can be generated by covalent modification cycles.
Amplification can also be provided by another means. The criterion for instability is the ratio of the inhibition elasticity to the substrate elasticity. If the output
reaction of the pathway is governed by a saturable enzyme then it is possible to
have "sub less than unity. This means that it is possible to trade cooperativity at the
inhibition site with saturation at the output reaction. The modified Goodwin model
of Bliss [3] illustrates the model with no cooperativity at the inhibition site but with
some saturation at the output reaction by using a simple Michaelis-Menten rate law.
A second property uncovered by BCT is that stability is enhanced if the kinetic
parameters of the participating reactions are widely separated, that is a mixture of
fast and slow reactions. The presence of fast reactions effectively shortens the
pathway and thus it requires higher feedback strength to destabilize the pathway
since the delay is now less.
All feedback oscillators require some device to provide amplification of a signal
combined with a suitable time delay so that the signal response can go out of phase.
In metabolic pathways amplification is often provided by a cooperative enzyme
while the delay is provided by the intermediate steps in the pathway. In signalling
368
Concentration of S1
1:5
0:5
6
Time
10
12
Figure 20.3 Plot of S3 versus time for the model shown in Figure 20.1. Note that
the profile of the oscillation is relatively smooth. See Table 20.2 for Jarnac listing.
p = defn feedback
J0:
J1:
J2:
J3:
J4:
end;
369
p.X0 = 10;
p.X1 = 0;
p.S1 = 0.973182; p.S2 = 1.15274;
p.S3 = 1.22721; p.S4 = 1.5635;
p.VM1 = 10; p.Keq1 = 10;
p.h = 10; p.Vm4 = 2.5;
p.KS4 = 0.5;
m = p.sim.eval (0, 12, 250, [<p.time>, <p.S1>]);
graph (m);
Figure 20.4 Three ring oscillators, one stage, three stage and five stage oscillators.
All ring oscillators require an odd number of gating elements. Even rings behave
as toggle switches.
370
Relaxation oscillators are commonly used in electronics where non-sinusoidal waveforms are required, for example, sawtooth or square waveforms[28]. Electronic
relaxation oscillators repeatedly alternate between two states at a frequency that
depends on a charging device such as a capacitor or an inductor. As a result, relaxation oscillators are characterized by two time-scales which is why they can be
used to generate non-sinusoidal waveforms.
Relaxation oscillators are favorite amongst theorists [31, 36, 41, 14] as well as biological evolution [53, 15, 21, 38, 7] is the relaxation oscillator. This kind of oscillator operates by charging a species concentration which, upon reaching a threshold,
changes the state of a bistable switch. When the switch changes state it causes the
species to discharge. Once the species has discharged the bistable switch returns
to the original state and the sequence begins again. Positive feedback or a twostep ring oscillator forming a toggle switch is used to generate the bistability and a
negative feedback loop provides the signal to switch the bistable switch.
In biological networks, a relaxation oscillator operates by charging a species concentration which, upon reaching a threshold, changes the state of a bistable switch.
When the switch changes state it causes the species to discharge. Once the species
has discharged the bistable switch returns to the original state and the sequence
begins again. Figure 20.5 shows a simple genetic relaxation oscillator with the
bistable switch on the left (formed from P1 ) and the negative feedback originating
from the gene product P2 .
Bistable Switch
P1
P2
Negative Feedback
371
Concentration
0:5
50
100
150
200
250
Time
Figure 20.6 Typical spiky appearance of oscillatory behavior from a relaxation
oscillator, from Tyson et. al, 2003, model 2(c). See Table 20.2 for Jarnac listing.
var R, X, E, EP;
ext src, S, Waste;
end;
p.src = 0;
p.ko = 0.4;
p.k2 = 1;
p.EP = 1;
p.kop = 0.01;
p.k1 = 1;
p.R = 1;
p.S = 0.2;
Listing 20.2 Jarnac script used to generate Figure 20.6. From Tyson et al [52],
Substrate-depletion oscillator 2c
372
373
374
Appendices
375
Symbols
A.1 List Of Symbols
x
y
p
vi
s
H.s/
"ji
Ejj
L./
A
B
C
D
N
ki
377
378
APPENDIX A. SYMBOLS
.t/
u.t/
!
Delta function
Unit step function
Frequency
Phase angle
Numerical Methods
B.1 Introduction
In this chapter we will look at ways to solve the differential equations that we
generate when we build a model of a system. For example, consider the simplest
possible model, the first-order irreversible degradation of a molecular species, S
into product P :
S !P
The differential equation for this simple reaction is given by the familiar form:
dS
D
dt
k1 S
(B.1)
Our aim is to solve this equation so that we can describe how S changes in time.
There are at least two ways to do this, we can either solve the equation mathematically or use a computer to obtain a numerical solution. Mathematically we can
either solve the system using traditional methods from calculus or we can use the
Laplace transform method.
To use the traditional method we move the dependent variables onto one side and
in the independent variables onto the other, this is called the separation of variables.
In the above equation one can easily do this by dividing both sides by S to give:
379
380
dS 1
D
dt S
(B.2)
k1
k1
ln Sdt D
k1
ln S D
dt
k1 t C C
S
ln
So
k1 t C ln So
D
k1 t
Taking anti-natural logarithms on both sides and multiply both sides by So gives:
S D So e
k1 t
For simple systems such as this, it is possible to obtain analytical solutions but very
rapidly one is confronted with the fact that for 99.99% of problems, no mathematical solution exists. In such cases, we must carry out numerical simulations.
381
10
8
S
6
4
2
0
10
Time, t
15
20
382
B.2.1
Euler Method
The Euler method is the simplest possible way solve a set of ordinary differential
equations. The idea is very simple. Consider the following differential equation
that describes the degradation rate of a species, S :
dS
D
dt
k1 S
The Euler method uses the rate of change of S to predict the concentration at some
future point in time. Figure B.2 describes the method in detail. At time t1 , the
rate of change in S is computed from the differential equation using the known
concentration of S at t1 . The rate of change is used to compute the change in S over
a time interval h, using the relation, h dS=dt . The current time, t1 is incremented
383
by the time step, h and the procedure repeated again, this time starting at t2 . The
method can be summarized by the following two equations which represent one
step in an iteration that repeats until the final time point is reached:
y.t C h/ D y.t / C h
tnC1 D tn C h
dy.t /
dt
(B.3)
Figure B.2 also highlights a problem with the Euler method. At every iteration,
there will be an error between the change in S we predict and what the change in
S should have been. This error is called the truncation error and will accumulate at each iteration. If the step size is too large, this error can make the method
numerically unstable resulting in wild swings in the solution.
Figure B.2 also suggests that the larger the step size the larger the truncation error.
This would seem to suggest that the smaller the step size the more accurate the
solution will be. This is indeed the case, up to a point. If the step size becomes too
small there is the risk that roundoff error which will propagate at each step into the
solution. In addition, if the step size is too small it will require a large number of
iterations to simulate even a small time period. The final choice for the step size
is therefore a compromise between accuracy and effort. A theoretical analysis of
error propagation in the Euler method indicates that the error accumulated over the
entire integration period (called the global error) is proportional to the step size.
Therefore halving the step size will reduce the global error by half. This means that
to achieve even modest accuracy, small step sizes are necessary. As a result, the
method is rarely used in practice. The advantage of the Euler method is that it is
very easy to implement in computer code or even on a spreadsheet.
The Euler method can also be used to solve systems of differential equations. In
this case all the rates of change are computed first followed by the application of the
Euler equation B.3. As in all numerical integration methods, the computation must
start with an initial condition for the state variables at time zero. The algorithm is
described using pseudo-code in Algorithm 1.
Example 9
Solve the decay differential equation B.1 using the Euler method. Assume k1 D 0:2 and
the concentration of So and P are time D 0 is 10 and 0 respectively. Assume a step size,
h, of 0.4. Form a table of four columns, write out the solution to two three decimal places.
384
S
Slope:
dS/dt
True Solution
A)
Truncation Error
S = h . dS/dt
Time
t1
B)
t1
t2
Time
Figure B.2 Euler Method. Starting at t1 , the slope dS=dt at t1 is computed (Panel
A). The slope is used to project forward to the next solution in time step, h, to t2
(Panel B). The new solution at t2 is indicated by P . However the solution is given
by point R, located on the solution curve at t2 . Reducing the step size h will reduce
the error between the exact and the projected solution, but will simultaneously increase the number of slope projections necessary to compute the solution over a
given time period.
The 4t h column should include the exact solution for comparison.
Time
Exact Solution
0
0.4
0.8
1.2
...
10
9.2
8.464
7.787
10
9.23
8.52
7.87
2
1.84
1.6928
0.01
Figure B.3 shows the effect of different steps sizes on the Euler method. Four
cases are shown, in the worse case the solution is unbounded and the computer
will eventually crash with an overflow error. The second case is where the result
is bounded but the solution bears no resemblance at all to the actual solution. The
third case shows that the solution is beginning to resemble the actual solution but
irregularities appear near the beginning of the integration. The final case shows the
actual solution generated from a specialized integrator.
104
385
h D 0:55, unbounded
104
5
0
10
20
h D 0:5, bounded
10
h D 0:00625, convergent
20
15
10
5
0
10
20
t
20
Best solution
20
15
10
5
0
10
20
t
Figure B.3 Effect of different step sizes on the Euler method using a simple linear chain of reactions where each reaction follows reversible mass-action kinetics:
k1
k3
k5
k7
Xo
S1 , S1
S2 , S2
S3 , S3
X1 where k1 D 0:45; k2 D 0:23; k3 D
k2
k4
k6
386
B.2.2
As indicated in the last section, the Euler method, though simple to implement,
tends not to be used in practice because it requires small step sizes to achieve reasonable accuracy. In addition the small step size makes the Euler method computationally slow. A simple modification however can be made to the Euler method to
significantly improve its performance. This approach can be found under a number
of headings, including the modified Euler method, the Heun or the improved Euler
method.
The modification involves improving the estimate of the slope by averaging two
derivatives, one at the initial point and another at the end point. In order to calculate
the derivative at the end point, the first derivative must be used to predict the end
point which is then corrected by using the averaged slope (Figure B.4). This method
is a very simple example of a predictor-corrector method. The method can be
summarized by the following equations:
387
dy.t /
dt
h dy.t/ dy.t C h/
Corrector: y.t C h/ D y.t/ C
C
2
dt
dt
Predictor: y.t C h/ D y.t/ C h
(B.4)
(B.5)
(B.6)
tnC1 D tn C h
S
P
B
P
A
C
Average Slope of A and B
A)
t1 h t 2
Time
B)
t1 h t 2
Time
Figure B.4 Heun Method. Starting at t1 , the slope A at T is computed. The slope
is used to predict the solution at point P using the Euler method. From point P the
new slope, B is computed (Panel A). Slopes A and B are now averaged to form a
new slope C (Panel B). The averaged slope is used to compute the final prediction.
A theoretical analysis of error propagation in the Heun method shows that it is
a second order method, that is if the step size is reduced by a factor of 2, the
global error reduced by a factor of 4. However, to achieve this improvement, two
evaluations of the derivatives is required per iteration, compared to only one for
the Euler method. Like the Euler method the Heun method is also quite easy to
implement.
388
h
.ai C bi /
2
currentTime D currentTime C h
end while
Algorithm 2: Heun Integration Method. fi .y/ is the i t h ordinary differential
equation
B.2.3
Runge-Kutta
The Heun method described in the previous section is sometimes called the RK2
method where RK2 stands for 2nd order Runge-Kutta method. The Runge-Kutta
methods are a family of methods developed around the 1900s by the German mathematicians, Runge and Kutta. In addition to the 2nd order Heun method there are
also 3rd, 4th and even 5th order Runge-Kutta methods. For hand coded numerical
methods, the 4th order Runge-Kutta algorithm (often called RK4) is probably the
most popular among modelers. The algorithm is a little more complicated in that it
involves the evaluation and weighted averaging of four slopes. In terms of global
error, however, RK4 is considerably better than Euler or the Heun method and has
a global error of the order of four. This means that halving the step size will reduce
the global error by a factor or 1/16. Another way of looking at this is that the step
size can be increased 16 fold over the Euler method and still have the same global
error. The method can be summarized by the following equations which have been
simplified by removing the dependence on time:
389
k1 D h f yn
k1
k2 D h f yn C
2
k2
k3 D h f yn C
2
k4 D h f yn C k3
1
y.t C h/ D y.t/ C
k1 C 2 k2 C 2 k3 C k4
6
tnC1 D tn C h
Figure B.5 shows a comparison of the three methods, Euler, Heun and RK4 in
solving the Van der Pol equations. The Van der Pol equations are a classic problem
set that is often used when comparing numerical methods. The equations model
an oscillating system, inspired originally from modeling vacuum tubes but also
later formed the basis for developments in modeling action potentials in neurons.
The Figure shows that the Heun and RK4 methods are very similar, at least for
the Van der Pol equations, though this is not always be the case. For this particular
model the solution generated by the RK4 method is very similar to the best possible
solution that can be obtained by numerical solution.
B.2.4
390
h
.k1i C 2 k2i C 2 k3i C k4i /
6
currentTime D currentTime C h
end while
Algorithm 3: 4th Order Runge-Kutta Integration Method
391
25
20
Euler
15
10
5
0
-5 0
-10
10
Time
15
20
Figure B.5 Comparison of Euler, Heun and RK4 numerical methods at integrating
the Van der Pol dynamic system: dy1 =dt D y2 and dy2 =dt D y1 C.1 y1 y1 /y2 .
The plots show the evolution of y1 in time. The RK4 solution is almost indistinguishable from solutions generated by much more sophisticated integrators. Step
size was set to 0.35.
ing the step size is called the Dormand-Prince method. This method carries out two
trials based on the fourth and fifth order Runge-Kutta. Any difference between the
two trials is used to adjust the step size. Matlabs ode45 implements the DormandPrince method. Similar methods to Dormand-Prince include the Fehlberg and more
recently the Cash-Karp method.
Many of these simple adjustable step size solvers are quite effective. Sometimes
they can be slow however especially for the kinds of problem we find in biochemical models. In particular there is a class of problem called stiff problems which generally plagues the biochemical modeling community. Stiff models require highly
specialized solvers which have been developed in the last four decades.
B.2.5
Stiff Models
392
compared to other components. This means that the step size has to be very small
to accommodate the fast processes even though the rest of the system could be accurately solved using a much larger step size. The overall result is the need for very
small steps sizes and therefore significant computational cost and the possibility of
roundoff error which will tend to be amplified by the large time constants in the fast
system. The net result are solutions which bear no resemblance to the true solution.
Most modern simulators will employ specific stiff algorithms for solving stiff differential equations. Of particular importance is the sundials suite and odepack.
Sundials includes a number of very useful, well written and documented solvers.
In particular the CVODE solver is very well suited to finding solutions to stiff differential equations. As a result sundials is widely used in the biochemical modeling
community (for example by Jarnac and roadRunner). Before the advent of sundials,
the main workhorse for solving stiff systems was the suite of routines in odepack.
Of particular importance was LSODA which in the 1990s was very popular and is
still a valuable set of software (currently used in COPASI). The original stiff differential equation solver was developed by Gear in the 1970s and is still used in
Matlab in the form of ode15s. We will return to a fuller discusion of software tools
in chapter ??.
393
k1 y1
k2 y2
function dy = myModel(t, y, p)
dy = zeros (2,1);
vo = p(1);
k1 = p(2);
k2 = p(3);
dy(1) = vo - k1 y(1);
dy(2) = k1 y(1) - k2 y(2);
We would then call the solver as follows:
394
Further Reading
Unfortunately there are very few reasonably priced books on numerical analysis.
The two most popular expensive books are by Press and Burden and are included
here for reference. Both books can be bought second-hand at reasonable prices and
the content has not changed significantly between editions. One could even argue
that the code examples in the latest edition of Press are actually worse than the
previous editions.
1. Press, Teulolsky, Vetterling and Flannery (2007) Numerical Recipes. Cambridge University Press, 3rd Edition ISBN-10: 0521880688
2. Burden and Faires (2010) Numerical Analysis. Brooks Cole, 9th Edition.
395
ISBN-10: 0538733519
For the budget conscious buyer I can highly recommend the Dover edition:
1. Dahlquist and Bjrck (2003) Numerical Methods. Dover Publications ISBN10: 0486428079
Exercises
1. Implement the Euler method in your favorite computer language and use
the code to solve the following two problems. Set initial conditions, S1 D
10; S2 D 0. Set the rate constants to k1 D 0:1I k2 D 0:25. Investigate the
effect of different steps sizes, h, on the simulation results.
a) dS1 =dt D
k1 S1
b) dS1 =dt D
k1 S1 I dS2 =dt D k1 S1
k2 S2
2. The following model shows oscillations at a step of h D 0:3 when using the
Euler method to solve the differential equations. By using simulation, show
that these oscillations are in fact an artifact of the simulation.
3. Solve the following problem using the Euler method. What step size do you
need to secure a reliable solution? Run the same problem on a specialist
application such as Matlab or SBW. Compare the time take to carry out both
simulations.
4.
396
Enzyme Kinetics
Enzyme kinetics is a branch of science that deals with the many factors that can
affect the rate of an enzyme-catalysed reaction. The most important factors include the concentration of enzyme, reactants, products, and the concentration of
any modifiers such as specific activators, inhibitors, pH, ionic strength, and temperature. When the action of these factors is studied, we can deduce the kinetic
mechanism of the reaction. That is, the order in which substrates and products bind
and unbind, and the mechanism by which modifiers alter the reaction rate.
397
398
formation, releasing product and free enzyme. The free enzyme is then available
for another round of binding to new reactant.
k1
k2
E C S )* ES ! E C P
k
(C.1)
Vm
1
0:8
0:6
1
v
2
0:4
0:2
0
5
Km
10
15
20
25
30
Substrate Concentration (S)
Figure C.1 Relationship between the initial rate of reaction and substrate concentration for a simple Michaelis-Menten rate law. The reaction rate reaches a limiting
value called the Vm . Km is set to 4.0 and Vm to 1.0. The Km value is the substrate
concentration that gives half the maximal rate.
399
to be some degree of product inhibition because the product can bind to the active
site and compete with the substrate.
k2
E C S )* ES )* E C P
k
(C.3)
The aggregate rate law for the reversible form of the mechanism can also be derived
and is given by:
Vf S=KS Vr P =KP
vD
(C.4)
1 C S=KS C P =KP
Peq
Vf KP
D
Seq
Vr KS
(C.5)
This equation shows that the four kinetic constants, Vf ; Vr ; KP and KS are not
independent. Haldane relationships can be used to eliminate one of the kinetic
constants by substituting the equilibrium constant in its place. This is useful because equilibrium constants tend to be known compared to kinetic constants which
may be unknown. By incorporating the Haldane relationship, we can eliminate the
reverse maximal velocity (Vr ) from C.4 to yield the equation:
vD
Vf =KS .S P =Keq /
1 C S=KS C P =KP
(C.6)
Separating out the terms makes it easier to see that the above equation can be partitioned into a number of distinct parts:
v D Vf .1
=Keq /
S=Ks
1 C S=KS C P =KP
(C.7)
where
D P =S. The first term, Vf , is the maximal velocity; the second term,
.1
=Keq /, indicates the direction of the reaction according to thermodynamic
considerations. The last term refers to the fractional saturation with respect to substrate. Thus we have a maximal velocity, a thermodynamic and a saturation term.
400
ES
E+P
EI
b) Uncompetitive Inhibition
E
ES
E+P
ESI
(C.9)
C.6. COOPERATIVITY
401
where Vm is the forward maximal velocity, and Ks and Kp are the substrate and
product half saturation constants.
Sometimes reactions appear irreversible, where no discernable reverse rate is detected, and yet the forward reaction is influenced by the accumulation of product. This effect is caused by the product competing with substrate for binding to
the active site and is often called product inhibition. Given that product inhibition is a type of competitive inhibition, we will briefly discuss it. An important industrial example of this is the conversion of lactose to galactose by the enzyme galactosidase where galactose competes with lactose, slowing the forward
rate [19].
To describe simple product inhibition with rate irreversibility, we can set the P =Keq
term in the reversible Michaelis-Menten rate law (C.4) to zero. This yields:
vD
Vm S
P
S C Km 1 C
Kp
(C.10)
It is not surprising to discover that equation (C.10) has exactly the same form as the
equation for competitive inhibition (C.8). As the product increases, it out competes
the substrate and therefore slows down the reaction rate.
We can also derive the equation by using the following mechanism and the rapidequilibrium assumption:
E CS *
) ES ! EP *
)E CP
(C.11)
C.6 Cooperativity
Many proteins are known to be oligomeric, meaning they are composed of more
than one identical protein subunit where each subunit has one or more binding
sites. Often the individual subunits are identical.
If the binding of a ligand (a small molecule that binds to a larger macromolecule)
to one site alters the affinity at other sites on the same oligomer, it is called cooperativity. If ligand binding increases the affinity of subsequent binding events, it is
termed positive cooperativity whereas if the affinity decreases, it is termed negative cooperativity. One characteristic of positive cooperativity is that it results in a
sigmoidal response instead of the usual hyperbolic response.
402
The simplest equation that displays sigmoid like behavior is the Hill equation:
vD
Vm S n
Kd C S n
(C.12)
One striking feature of many oligomeric proteins is the way individual monomers
are physically arranged. Often one will find at least one axis of symmetry. The
individual protein monomers are not arranged in a haphazard fashion. This level
of symmetry may imply that the gradual change in the binding constants as ligands
bind, as suggested by the Adair model, might be physically implausible. Instead,
one might envision transitions to an alternative binding state that occurs within the
entire oligomer complex. This model was originally suggested by Monod, Wyman
and Changeux [35], abbreviated as the MWC model. The original authors laid out
the following criteria for the MWC model:
1. The protein is an oligomer.
2. Oligomers can exist in two states: R (relaxed) and T (tense). In each state,
symmetry is preserved and all subunits must be in the same state for a given
R or T state.
3. The R state has a higher ligand affinity than the T state.
4. The T state predominates in the absence of ligand S .
5. The ligand binding microscopic association constants are all identical. This
is in complete contrast to the Adair model.
Given these criteria, the MWC model assumes that an oligomeric enzyme may exist
in two conformations, designated T (tensed, square) and R (relaxed, circle). The
equilibrium between the two states has an equilibrium constant L D T =R, which
is also called the allosteric constant. If the binding constants of ligand to the two
states are different, the distribution of the R and T forms can be displaced towards
either one form or the other. By this mechanism, the enzyme displays sigmoid
behavior. A minimal example of this model is shown in Figure C.3.
In the exclusive model (Figure C.3) the ligand can only bind to the relaxed form
(circle). The mechanism that generates sigmoidicity in this model works as follows. When ligand binds to the relaxed form, it displaces the equilibrium from the
tense form to the relaxed form. In doing so, additional ligand binding sites become
available. Thus, one ligand binding may generate four or more new binding sites.
Eventually there are no more tense states remaining, at which point the system is
saturated with ligand. The overall binding curve will therefore be sigmoidal and
C.6. COOPERATIVITY
403
L
L=
Figure C.3 A minimal MWC model, also known as the exclusive model, showing
alternative microscopic states in the circle (relaxed) form. L is called the allosteric
constant. The square form is called the tense state.
will show positive cooperativity. Given the nature of this model, it is not possible
to generate negative cooperativity. By assuming equilibrium between the various
states, it is possible to derive an aggregate equation for the dimer case of the exclusive MWC model:
S
S
1C
kR
kR
v D Vm
2
S
1C
CL
kR
This also generalizes to n subunits as follows:
S
S n 1
1C
kR
kR
Y D
(C.13)
n
S
CL
1C
kR
For more generalized reversible rate laws that exhibit sigmoid behavior, the reversible Hill equation is a good option. Invoking the rapid-equilibrium assumption,
we can form a reversible rate law that shows cooperativity:
vD
Vf .1
/ . C /
1 C . C /2
where D =Keq and and are the ratio of reactant and product to their
respective equilibrium constant, =KS and =KP . For an enzyme with h (using
the authors original notation) binding sites, the general form of the reversible Hill
equation is given by:
Vf .1 / . C /h 1
vD
(C.14)
1 C . C /h
404
C.7 Allostery
An allosteric effect is where the activity of an enzyme or other protein is affected
by the binding of an effector molecule at a site on the proteins surface, other than
the active site. The MWC model described previously can be easily modified to
accommodate allosteric action.
X
T2
R2
R 2X
X X R 2 X2
L
L=
T2
R2
R 2X
Figure C.4 Exclusive MWC model based on a dimer showing alternative microscopic states in the form of T and R states. The model is exclusive because the
ligand, X, only binds to the R form.
The key to including allosteric effectors is to influence the equilibrium between
the tense (T) and relaxed (R) states (See Figure C.4). To influence the sigmoid
curve, an allosteric effector need only displace the equilibrium between the tense
and relaxed forms. For example, to behave as an activator, an allosteric effector
needs to preferentially bind to the R form and shift the equilibrium away from
the less active T form. An allosteric inhibitor would do the opposite, that is bind
preferentially to the T form so that the equilibrium shifts towards the less active T
form. In both cases the Vm of the enzyme is unaffected.
The net result of this is to modify the normal MWC aggregate rate law to the following if the effector is an inhibitor:
v D Vm
.1 C /n 1
.1 C /n C L.1 C /n
(C.15)
where D S=Ks , D I =KI , and Ks and KI are kinetic constants related to each
ligand. A MWC model that is regulated by an inhibitor or an activator is described
C.7. ALLOSTERY
by the equation:
405
.1 C /n 1
v D Vm
.1 C /n
.1 C /n C L
.1 C
/n
There are also reversible forms of the allosteric MWC model but they are fairly
complex. Instead, it is possible to modify the reversible Hill rate law to include
allosteric ligands.
Vf 1
vD
Keq
. C /h
1 C h
C . C /h
1 C h
(C.16)
where:
<1
>1
inhibitor
activator
Vm S n
K C Sn
Repression: v D
Vm
K C Sn
Dual: v D
1 C K1 S1n1
Vm S1n1
C K2 S2n2 C K3 S1n1 S2n2
Further Reading
1. Sauro HM (2012) Enzyme Kinetics for Systems Biology. 2nd Edition, Ambrosius Publishing ISBN: 978-0982477335
406
Visualizing Simulations
D.1 Time Course Simulation
The simplest way to visualize a simulation is to plot the time courses that describe
the evolution of the state variables. Although useful, such plots are not the only
way to display a behavior of a model. In this chapter we will briefly look at three
different ways to visualize a model.
D.2 NullClines
Nullclines are very useful for looking at systems of two variables. If we have
a variable xi with differential equation dfi =dt , then the nullcline is the line that
satisfies, dfi =dt D 0. If two nullclines are plotted, then their intersection makes
the steady state points.
For example, the system:
vo
! y1
k1 y1
! y2
k2 y2
407
408
with ODEs:
dy1
D vo
dt
k1 y1
dy2
D k1 y1
dt
k2 y2
has two nullclines. Assuming y1 is plotted on the x axis and y2 on the y axis, then
the first nullcline is simply y1 D vo =k1 , that is a vertical straight line at y1 . The
second equation is given by:
y2 D k1 y1 =k2
and represents a straight line with slope k1 =k2 running through the zero point.
6
y2
4
2
0
10
y1
Figure D.1 Nullclines for the system: dy1 =dt D vo k1 y1 I dy2 =dt D k1 y1
k2 y2 where vo D 6; k1 D 0:15; k2 D 3. The intersection point marks the steady
state when both equations equal zero.
! S1
kS1
409
kS1
The steady state solution can be found by setting the differential equation to zero
and solving for S1 . If we do this we obtain:
S1 D
vo
k
To plot a bifurcation plot we must choose a parameter, for example k and plot the
steady state solution with respect to k.
8
S1
6
4
2
0
6
k
10
vo
12
kS1
410
Math Notes
E.1 Polynomials
Some useful results can be generalized from the Routh criterion for small polynomials.
Quadratic: a0 x 2 C a1 x C a2 D 0
If all coefficients are positive (1) then both roots will have negative real parts.
Cubic: a0 x 3 C a1 x 2 C a2 x C a3 D 0
If coefficients, ai are positive and:
a1 a2 > a0 a3
then the system will be stable.
aj D jaj
jabj D jajjbj
ja=bj D jaj=jbj
411
412
Z
Z
(E.1)
1
x dx D x 2
2
(E.2)
1
x
1
dx D
x2
k dx D kx
(E.3)
1
dx D ln jxj
x
(E.4)
1 ax
e
a
(E.5)
e ax dx D
1
cos.ax/
a
sin.ax/ dx D
f .x/ dx D F .b/
F .a/
Example:
1
st
1
e
s
1 0
e
s
dt D
st
1 1
1
e
D
s
s
(E.6)
Complex Numbers
F.1 Review of Complex Numbers
p
Imaginary numbers are solutions to equations
such
as
x and are usually repp
resented
for convenience
by the symbol, xi . Thus the imaginary number of
p
p
1 D i and for
9 D 3i (We ignore the fact that there are two solution,
C3i and 3i.
Although i is often used to represent the imaginary unit number, in engineering j
is often used instead to avoid confusion with electrical current, i .
Imaginary numbers can also be paired up with real numbers to form complex numbers. Such number have the form:
a C bj
where a represents the real part and b the imaginary part. This notation is actually a short-hand for the more general statement:
.a; 0j / C .0; bj /
that is vector addition. For convenience the 0 values are omitted and the notation
shortened to a Cbj . It is therefore important to realize that the C notation doesnt
mean that a is added to b, this is not possible because a and b are quite different
objects. The rules for the addition of complex numbers is different and will be
described later.
413
414
Conjugate Pairs
A conjugate complex pair is given by the pair of complex numbers:
a C bj
bj
That is both numbers have the same real part, a, but the imaginary part, b, has the
same magnitude but opposite sign.
Polar form
We can express a complex number on a two dimension plane where the horizontal
axis represents the real part and the vertical axis the imaginary part. A complex
number can therefore represented as a point on the plane.
Im
a + bj
r
q
Re
b
r
q
q
r
-b
a
Re
a - bj
415
we can express the real and imaginary parts using trigonometric functions:
b D r sin
a D r cos
where r is the length of the line from the origin to the point and the angle. The
following two representations are therefore equivalent.
a C bj D r.cos C j sin /
When written like this, r is also known as the magnitude or modulus of the complex
number, A, that is:
p
jAj D r D a2 C b 2
The notation jAj is often used to denote the magnitude of a complex number. Examples:
p
p
j3 4i j D 32 C . 4/2 D 25 D 5
p
p
j 4ij D . 4/2 D 16 D 4
The angle, is known as the argument or phase and is given by:
1 b
D tan
a
In calculating the angle we must be careful about the sign of b=a. Figure F.3
illustrates the four possible situations. If the point is in the second quadrant, 180o
should be added to the tan 1 result. If the point is in the third quadrant, then 180o
should be subtracted from the tan 1 result. The 1st and 4th quadrants need no
adjustments.
The rules for computing the angle are summarized in the list below. The atan2
function often found in software such as Matlab will usually automatically take
into consideration the signs.
8
x>0
tan 1 . yx /
C tan 1 . x /
y 0; x < 0
< C tan 1 . y /
y < 0; x < 0
x
atan2.y; x/ D
y > 0; x D 0
y < 0; x D 0
: 2
undefined
y D 0; x D 0
Examples
416
a)
b)
1st Quadrant
2nd Quadrant
Unit Circle
Unit Circle
45
135
180
-180
-1
-90
c)
3rd Quadrant
Unit Circle
d)
4th Quadrant
Unit Circle
-1
-135
1
o
-1
-1
-45
Figure F.3 a) Both axes are positive, arctan .1=1/ D 45o ; b) Horizontal axis is
negative -1, arctan .1= 1/ D 135o ; c) Vertical axes is negative, arctan . 1=1/ D
45o ; d) Both axes are negative, arctan . 1= 1/ D 135o
atan2.2=5/ D 21:8
atan2. 2=5/ D
atan2. 2=
atan2.2=
5/ D
21:8
158:18
5/ D 158:18
417
.c C dj / D .a
4. .a C bj / .c C dj / D .ac
5. .a C bj / .a
6.
c/ C .b
d /j
bd / C .ad C bc/j
bj / D a2 C b 2
.ac C bd / C .bc
a C bj
D
c C dj
c2 C d 2
ad /j
Division is accomplished by multiplying the top and bottom by the conjugate. Note
that the product of a complex number and its conjugate gives a real number, this
allows us to eliminate the imaginary part from the denominator.
a C bj c
a C bj
D
c C dj
c C dj c
D
.ac
dj
dj
b. d // C .a. d / C bc/j
c2 C d 2
.ac C bd / C .bc
c2 C d 2
bc ad
ac C bd
C 2
2
2
c Cd
c C d2
ad /j
Exponential Form
By Eulers formula:
e j D cos./ C j sin./
we can substitute the sine/cosine terms in the polar representation to give:
a C bj D re j
This gives us three equivalent ways to represent a complex number:
418
a C bj D r.cos./ C j sin./ D re j
Common Integrals
Computing a definite integral:
Z
f .x/ dx D F .b/
F .a/
Example:
1
st
1
e
s
1 0
e
s
dt D
st
1 1
1
e
D
s
s
419
420
Z
Z
Z
Z
Z
Z
k dx D kx
(G.1)
1
x dx D x 2
2
(G.2)
1
dx D
x2
1
x
(G.3)
1
dx D ln jxj
x
(G.4)
1 ax
e
a
(G.5)
e ax dx D
sin.ax/ dx D
1
cos.ax/
a
(G.6)
Lf .t/ D F .s/
Scaling
af .t/
aF .s/
Unit Step
Delay
u.t
Ramp
Impulse
.t/
Delayed Impulse
.t
Exponential
e at f .t/
Delayed Function f .t
Power
1
s
e
a/
as
s
1
s2
1
t0 /
a/u.t
t f .t/
a/
st0
F .s
a/
as
F .s/
d n F .s/
. 1/
ds n
n
421
422
f .t/
Lf .t/ D F .s/
f 0 .t/
sF .s/
f .0/
f n .t/
s n F .s/
s .n
f .n
sin kt
k
s2 C k2
cos kt
s
s2 C k2
.0/
e at
a
k
sinh kt
s2
a
.1
b
e at
a
e bt
b
k2
s
cosh kt
s2
bt
k2
a
s.s C b/
ae at be bt
a b
Z t
f .x/g.t
x/dx
.s
1
a/.s
b/
.s
s
a/.s
b/
F .s/G.s/
t n (n D 0; 1; 2; : : : )
t x (x
te at
1/
1/
n
s nC1
.x C 1/
s xC1
1 2 R)
1
.s
a/2
f .0/
423
f .t/
Lf .t/ D F .s/
t n e at
n
.s a/nC1
e at sin kt
k
a/2 C k 2
.s
e at cos kt
s a
.s a/2 C k 2
e at sinh kt
k
a/2
.s
s a
.s a/2 k 2
e at cosh kt
1
.a cos.at/
a
t sin kt
sin.at//
1
s 2 C a2
.s 2
2ks
C k 2 /2
s2 k2
.s 2 C k 2 /2
t cos kt
t sinh kt
.s 2
2ks
k 2 /2
t cosh kt
s2 k2
.s 2 k 2 /2
sin at
t
arctan
1
p e
t
a2 =4t
a
2
e a =4t
p
2 t3
a
erfc p
2 t
k2
e
e
p
a s
p
a s
p
a s
a
s
424
References
[1] Aris, R. 1965. Arch. Rational Mech. Anal 19:8199.
[2] Black, H. 1977. IEEE spectrum 14 (12):5561.
[3] Bliss, R. D., P. R. Painter, and A. G. Marr. 1982. J Theor Biol 97 (2):177193.
[4] Boahen, K. 2005. Scientific American 292 (5):5663.
[5] Bode, A. M., and Zigang Dong. 2004. Nat Rev Cancer 4 (10):793805.
[6] Brilli, M., Marco Fondi, Renato Fani, Alessio Mengoni, Lorenzo Ferri, Marco
Bazzicalupo, and Emanuele Biondi. 2010. BMC Systems Biology 4 (1):52.
[7] Chen, K. C., L Calzone, A Csikasz-Nagy, F R Cross, B Novak, and J J Tyson.
2004. Mol Biol Cell 15 (8):38413862.
[8] Chickarmane, V., B. N. Kholodenko, and H. M. Sauro. 2007. J Theor Biol 244
(1):6876.
[9] Chickarmane, V., C. Troein, U. A. Nuber, H. M. Sauro, and C. Peterson. 2006.
PLoS Comput Biol 2 (9).
[10] Cohen, P. 2000. Trends in Biochemical Sciences 25 (12):596601.
[11] Deckard, A., and H. M. Sauro. 2004. Chembiochem 5:142331.
[12] Elowitz, M. B., and S Leibler. 2000. Nature 403:335338.
[13] Fall, C. P., E. S. Marland, J. M. Wagner, and J. J. Tyson. 2002. Computational
Cell Biology. Springer-Verlag, New York.
[14] Field, R. J., and R. M. Noyes. 1974. J. Chemical Physics 60(5):18771884.
[15] FitzHugh, R. 1955. Bull. Math. Biophysics 17:257278.
425
426
REFERENCES
REFERENCES
427
428
REFERENCES
REFERENCES
429
430
REFERENCES
History
1. VERSION: 0.1
Date: 2011-04-2
Author(s): Herbert M. Sauro
Title: Introduction to Control Theory for Biologists
Modification(s): First Edition Release
2. VERSION: 0.47
Date: 2014-12-8
Author(s): Herbert M. Sauro
Title: Introduction to Control Theory for Biologists
Modification(s): See correction list at front of book
3. VERSION: 0.50
Date: 2014-12-21
Author(s): Herbert M. Sauro
Title: Introduction to Control Theory for Biologists
Modification(s): Major update across all sections
431
432
REFERENCES
INDEX
433
Index
Symbols
convolution . . . . . . . . . . . . . . . . . . . . . . 233
Km . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 398 cooperativity . . . . . . . . . . . . . . . . . . . . . 401
Vm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 398 cut-off point . . . . . . . . . . . . . . . . . . . . . 275
bandwidth . . . . . . . . . . . . . . . . . . . . . . . 275
bifurcation plot . . . . . . . . . . . . . . 315, 323
biochemical model . . . . . . . . . . . . . . . 193
biochemical systems theory . . . . . . . . . 81
bistable . . . . . . . . . . . . . . . . . . . . . 316, 318
block diagrams . . . . . . . . . . . . . . . . . . . 226
Bode plots . . . . . . . . . . . . . . . . . . . . . . . 270
Briggs-Haldane . . . . . . . . . . . . . . . . . . 398
competitive . . . . . . . . . . . . . . . . . . . . . . 400
competitive inhibition . . . . . . . . . . . . . 400
complete solution . . . . . . . . . . . . . . . . 198
complex exponential . . . . . . . . . . . . . . 162
complex number . . . . . . . . . . . . . . . . . 314
computer exchange format . . . . . . . . . 76
conjugate Pair . . . . . . . . . . . . . . . . . . . . 314
control matrix . . . . . . . . . . . . . . . . . . . . 185
convergence, Laplace . . . . . . . . . . . . . 212
elasticity . . . . . . . . . . . . . . . . . . . . . . . . 323
elasticity coefficient . . . . . . . . . . . . . . . 81
elasticity rules . . . . . . . . . . . . . . . . . . . . . 87
electrical model . . . . . . . . . . . . . . . . . . 190
enzyme action . . . . . . . . . . . . . . . . . . . 397
enzyme elasticity . . . . . . . . . . . . . . . . . . 86
enzyme inhibitor complex . . . . . . . . . 400
enzyme kinetics . . . . . . . . . . . . . . . . . . 397
enzyme-reactant complex . . . . . . . . . 397
Eulers theorem . . . . . . . . . . . . . . . . . . 162
exclusive model . . . . . . . . . . . . . . . . . . 403
exponential . . . . . . . . . . . . . . . . . . . . . . 160
exponential series . . . . . . . . . . . . . . . . 205
extensive property . . . . . . . . . . . . . . . . . 79
extent of reaction . . . . . . . . . . . . . . . . . . 79
434
INDEX
mass-action kinetics . . . . . . . . . . . . . . . 84
Mathematica . . . . . . . . . . . . . . . . . . . . . . 89
maximal velocity . . . . . . . . . . . . . . . . . 398
mechanical model . . . . . . . . . . . . . . . . 188
Michaelis-Menten kinetics . . . . . . . . 397
I
multiple steady states . . . . . . . . . . . . . 316
impulse . . . . . . . . . . . . . . . . . . . . . . . . . 152 MWC model . . . . . . . . . . . . . . . . . . . . . 402
initial rate . . . . . . . . . . . . . . . . . . . . . . . 398
inputs . . . . . . . . . . . . . . . . . . . . . . . . . . . 143
N
intensive property . . . . . . . . . . . . . . . . . 80 negative cooperativity . . . . . . . . . . . . . 401
inverse transform . . . . . . . . . . . . . . . . . 212 Newton algorithm . . . . . . . . . . . . . . . . 123
inverse transforms . . . . . . . . . . . . . . . . 216 non-Homogeneous system . . . . . . . . 203
irreversible bistability . . . . . . . . . . . . . 325
INDEX
435
R state . . . . . . . . . . . . . . . . . . . . . . . . . . 402
ramp . . . . . . . . . . . . . . . . . . . . . . . . . . . . 150
rate of change . . . . . . . . . . . . . . . . . . . . . 77
rate of reaction . . . . . . . . . . . . . . . . . . . . 79
rational fractions . . . . . . . . . . . . . . . . . . 77
reaction kinetics . . . . . . . . . . . . . . . . . . . 77
reaction order . . . . . . . . . . . . . . . . . . . . . 81
reaction rate. . . . . . . . . . . . . . . . . . . . . . .78
relaxed . . . . . . . . . . . . . . . . . . . . . . . . . . 402
response of linear systems . . . . . . . . . 197
reversible rate law . . . . . . . . . . . . . . . . 399
T state . . . . . . . . . . . . . . . . . . . . . . . . . . . 402
Taylor series . . . . . . . . . . . . . . . . . . . . . 161
tense . . . . . . . . . . . . . . . . . . . . . . . . . . . . 402
thermodynamic equilibrium . . . . . . . 115
time invariance . . . . . . . . . . . . . . . . . . . 171
total response . . . . . . . . . . . . . . . . . . . . 204
total solution . . . . . . . . . . . . . . . . . . . . . 198
transcription factor . . . . . . . . . . . . . . . 316
truncation error . . . . . . . . . . . . . . . . . . 383
Tyson . . . . . . . . . . . . . . . . . . . . . . . . . . . 325
volume . . . . . . . . . . . . . . . . . . . . . . . . . . . 80