Académique Documents
Professionnel Documents
Culture Documents
rodov
Pr Fakulta
edecka
Brno 2009
Jakub Cupera
Declaration
I would like to thank my advisor doc. RNDr. Petr Lansky CSc. for his
friendly and pleasant attitude, enthusiasm and care, and my parents for
their love and support.
Abstract
Introduction 2
Notation 3
1 Models of dissolution 4
1.1 Deterministic models . . . . . . . . . . . . . . . . . . . . . . . 6
1.1.1 Homogenous model . . . . . . . . . . . . . . . . . . . . 6
1.1.2 Weibull model . . . . . . . . . . . . . . . . . . . . . . . 8
1.1.3 Hixson-Crowell model . . . . . . . . . . . . . . . . . . 9
1.2 Stochastic models . . . . . . . . . . . . . . . . . . . . . . . . . 10
1.2.1 Gaussian model . . . . . . . . . . . . . . . . . . . . . . 11
1.2.2 Log-normal model . . . . . . . . . . . . . . . . . . . . 14
2 Theoretical background 21
2.1 Fisher information and its lower bound . . . . . . . . . . . . . 21
2.2 Maximum likelihood estimation . . . . . . . . . . . . . . . . . 26
Summary 59
Bibliography 60
1
Introduction
2
Notation
Random vectors
Functions
3
Models of dissolution
1
The dissolution is defined as a process of attraction and association of molecu-
les of a solvent with molecules of a solute. Number of these associated parti-
cles is given in moles, where 1 mole contains approximately 6.022045 1023
particles. The concentration C is defined as a number of solvent molecules in
unit volume of solute, mathematically C = n/V , where n is the total number
of dissolved particles and V is the total volume of solute.
The mathematical models of dissolution can be divided into two basic
groups - deterministic and stochastic. Both of them investigate the whole
population of particles and describe the time course of concentration C(t).
While the deterministic models work with given function C(t), the stochastic
ones describe the dissolution as a random process. It is natural to assume
that the concentration C(t) is a nondecreasing function of time. Furthermore,
we assume the function C(t) is continuous and smooth (with derivatives of
all orders), C(0) = 0 and C(t) CS for increasing t, where CS is a limit
concentration after the solvent is dissolved or the solute is saturated. For
comparison among the dissolution profiles, we normalize the profile C(t) to
the form
C(t)
F (t) = , (1.1)
CS
where the function F (t) expresses the dissolved fraction of solvent at the time
instant t. Note that F (t) 1 for increasing t.
Function F (t) satisfies all the conditions to be a cumulative distribution
function (cdf) of a random variable and thus it can be seen as a cdf of a
random variable T representing the time until a randomly selected molecule
enters solution. Due to the assumptions made on C(t), function F (t) is
continuous and smooth with corresponding probability density dF (t)/dt of
the random variable T .
4
An alternative way how to characterize the dissolution is by defining the
fractional dissolution rate
dF (t)
dt
k(t) = . (1.2)
1 F (t)
1
2
(A)
(B)
0.8
1.5
0.6
F (t)
k(t)
1
0.4
0.5
0.2
0 0
0 1 2 3 0 1 2 3
t t
Fig. 1.1: Different profiles of fractional dissolution rate and corresponding cdfs (A)
Weibull functions (see Section 1.1.2) with scale parameter a = 1 and shape parameters
b = 0.9 (red) and b = 1.1 (black), (B) corresponding fractional dissolution rates.
If the form of the fractional dissolution rate k(t) is known, then the dissolu-
tion profile can be evaluated from equation
Z t
F (t) = 1 exp k(s)ds , (1.3)
0
5
1.1. DETERMINISTIC MODELS
The parametric methods fit the measured data to certain functions, where
its estimated parameters could give us an information about properties of
the dissolution process and consequently about the random variable T .
dC(t)
= a(CS C(t)), C(0) = 0, (1.5)
dt
where a > 0 is constant and CS is the limit concentration achieved after the
solvent is dissolved or the solute is saturated. In this model the rate of disso-
lution, dC(t)/dt, is proportional to the difference between the instantaneous
concentration and the final concentration of solvent. This model was intro-
duced by chemists Noyes and Whitney in [23]. In generalizations of (1.5) the
constant a is often considered to depend on temperature, surface of solvent
etc., for example see section 1.1.3. Solution of differential equation (1.5) is
6
1.1. DETERMINISTIC MODELS
thus
F (t) = 1 eat . (1.7)
The fractional dissolution rate defined by equation (1.2) is constant, k(t) a,
for model (1.7). Thus the probability that a selected particle of the solvent
undissolved at the time t will be dissolved in the time interval [t, t + dt), is
constant. This property makes the homogenous model specific among the
whole spectrum of dissolution models.
As seen from equation (1.7), the dissolution never ends in this model, and
thus there is always a small amount of undissolved substance. Practically,
as an end of the dissolution we could consider, for example, the time instant
when 99% of the solvent is dissolved. Examples of the homogenous model
with different values of parameter a are shown in Fig. 1.2.
0.9
0.8
0.7
0.6
F (t)
0.5
0.4
0.3
0.2
0.1
0
0 0.5 1 1.5 2 2.5 3
t
Fig. 1.2: Time course of dissolution profile F (t) of homogenous model (1.7) with param-
eter a = 0.5 (blue), a = 1 (red) and a = 2 (black).
7
1.1. DETERMINISTIC MODELS
1 1
(A) (B)
0.8 0.8
0.6 0.6
F (t)
F (t)
0.4 0.4
0.2 0.2
0 0
0 0.5 1 1.5 2 0 0.5 1 1.5 2
t t
Fig. 1.3: Time course of dissolution profile F (t) of Weibull model (1.8). (A) Parameter
b fixed, b = 3 and parameter a = 4 (black), a = 1 (red), a = 0.4 (blue). (B) Parameter a
is fixed, a = 1 and parameter b = 0.4 (black), b = 1 (red) and b = 4 (blue).
8
1.1. DETERMINISTIC MODELS
9
1.2. STOCHASTIC MODELS
where a > 0, b > 0 are constants. These models are called the root laws
because of the form of equation (1.14). How the parameters a, b affect the
shape of F (t) can be seen in Fig. 1.4.
1 1
(A) (B)
0.8 0.8
0.6 0.6
F (t)
F (t)
0.4 0.4
0.2 0.2
0 0
0 0.5 1 1.5 2 0 0.5 1 1.5 2
t t
Fig. 1.4: Time course of dissolution profile F (t) of Hixson-Crowell model (1.16) with (A)
fixed parameter b = 0.5, parameter a = 0.3 (black), a = 0.6 (red) and a = 1.2 (blue), (B)
fixed parameter a = 2, parameter b = 0.5 (black), b = 1 (red) and b = 5 (blue).
10
1.2. STOCHASTIC MODELS
11
1.2. STOCHASTIC MODELS
where (s) is a white noise given by (1.19). Sample paths of random processes
(1.22) and (1.24) are shown in Fig. 1.6 and details about their numerical
simulation are given in Chapter 4.
Our aim is to analyze measured data, thus we assume that the variance
s2 (t) of the data obtained from stochastic model (1.18), resp. (1.22), has
form
s2 (t) = 2 (t) + (t), (1.25)
where 2 (t) is variance of the dissolution and (t) 0 reflects the measure-
ment error. The function 2 (t) is unknown in general. We should assume
1.5
f (x, t)
0.5
0
1
3
0.5 2
1
x 0 0
t
Fig. 1.5: Gaussian probability density (1.21) with mean described by homogenous model
(1.7), parameter a = 1, and constant variance of the data s2 (t) 0.04.
12
1.2. STOCHASTIC MODELS
3
(A) 1 (B)
2
0.8
1
0.6
W (t)
(t)
0
0.4
1
0.2
2
0
3
0 1 2 3 0 1 2 3
t t
Fig. 1.6: Sample paths of random processes. (A) Sample paths of Wiener process W (t),
(B) sample paths of random process (1.22) with mean F (t) = 1 exp(t2 ) and variance
s2 (t) = 0.04 exp(t2 )(1 exp(t2 )).
that 2 (0) = 0 (nothing is dissolved) and 2 (t) tends to zero for increasing t
(everything is dissolved). As an example of such a function we can take
13
1.2. STOCHASTIC MODELS
where > 0 is an amplitude of the noise. From (1.2), after replacing F (t) by
(t) to distinguish between deterministic and stochastic models, we obtain
stochastic differential equation
dU(t)
= k(t)dt dW (t), U(0) = 1, (1.29)
U(t)
thus Z t Z t
dU(s)
= k(s)ds W (t). (1.30)
0 U(s) 0
To evaluate the integral on the left hand side we have to introduce the It
o
calculus.
or equivalently
dX(t) = u(t)dt + v(t)dW (t),
where u(t), v(t) are random processes satisfying
Z t
P |u(s)|ds < , t 0 = 1,
0
and Z t
2
P v(s) ds < , t 0 = 1.
0
14
1.2. STOCHASTIC MODELS
Proof of the theorem can be found in [25]. To evaluate the integral at the
left hand side of (1.30) we use Ito formula (1.32) for the function
g(t, x) = ln x; x>0
and obtain
1 1 1
d(ln U(t)) = dU(t) 2
[dU(t)]2
U(t) 2 U (t)
dU(t) 1
= 2 U 2 (t)dt
U(t) 2U 2 (t)
dU(t) 1 2
= dt.
U(t) 2
Hence
dU(t) 1
= d(ln U(t)) + 2 dt
U(t) 2
so from (1.30) we conclude
Z t
U(t) 1
ln = k(s)ds 2 t W (t)
U(0) 0 2
where U(0) = 1, so
Z t
1
U(t) = exp( k(s)ds 2 t W (t)),
0 2
15
1.2. STOCHASTIC MODELS
and thus Z t
1
(t) = 1 exp( k(s)ds 2 t W (t)). (1.34)
0 2
We wish to obtain probability distribution of the random variable (t).
Due to (1.23), the expression in exponent in (1.34) has Gaussian probability
distribution Z t
1
N k(s)ds 2 t, 2 t ,
0 2
thus random variable 1 (t) has log-normal distribution (see ref. [21])
1 (t) logN (t), 2 (t) , (1.35)
1
(A) 1 (B)
0.8
0.8
0.6
0.6
(t)
(t)
0.4
0.4
0.2
0.2
0
0
0 1 2 3 0 1 2 3
t t
Fig. 1.7: Sample paths of a random process (t) with indicated Weibull dissolution
profile F (t) given by (1.8) with parameters a = 1, b = 2, (A) random process (t) defined
by equation (1.34), = 0.2, (B) random process (t) defined by equation (1.42), = 0.05.
16
1.2. STOCHASTIC MODELS
It holds
E(t) = F (t),
var(t) = (1 F (t))2 exp( 2 t) 1 .
As the function f (x, t) for fixed t gives probability density of the random
variable 1 (t), the random variable (t) has probability density f (1x, t).
Model (1.34) suffers of certain defects. In Fig. 1.7 (A) it can be seen that
this model completely prevents the concentration to take values above 100%,
but it allows them to drop below zero, thus the model does not satisfy our
requirements. In equation (1.28) the random component dW (t) influences
the process of dissolution at its beginning only, because term 1 (t) is large
for t small. To obtain a more precise model one should assume, for example,
stochastic differential equation
or
d(t) = k(t)(1 (t))dt + (t)(1 (t))dW (t), (1.38)
1
1 (A) (B)
0.8
0.8
0.6
0.6
(t)
(t)
0.4
0.4
0.2 0.2
0 0
0 1 2 3 0 1 2 3
t t
Fig. 1.8: Sample paths of dissolution profile (t), k(t) = 2t. (A) model (1.37), = 0.1,
(B) model (1.38), = 0.4.
17
1.2. STOCHASTIC MODELS
with initial condition (0) = 0, where the random factor dW (t) does not
influence the process of dissolution at its begin and end. In this case we are
not able to find solution of stochastic differential equations (1.37) and (1.38)
in analytic form. Their numerical approximations (see Section 4.1) can be
seen in Fig. 1.8.
Let us study another model with solution in analytic form. Stochastic dif-
ferential equation (1.28) with solution (1.34) can be rewritten after dividing
by (1 (t))dt to form
d d
dt
(t) dt
F (t)
= + (t), (0) = 0.
1 (t) 1 F (t)
As we have seen, this model does not satisfy our assumptions and thus let
us consider stochastic differential equation
d d
dt
(t) dt
F (t)
= + (t), (0) = 0. (1.39)
(t) F (t)
Let > 0 be small. Firstly we solve equation (1.39) with initial condition
() = F (). This equation can be solved similarly as the equation (1.28),
where after application of Ito theorem we obtain
(t) F (t) 1 2
ln = ln t + W (t)
() F () 2
which can be rewritten as
() 1 2
(t) = exp ln F (t) t + W (t) . (1.40)
F () 2
Our purpose is to find limit for 0. As () is a random process, we have
to use stochastic approach.
Definition 1.3. Let X(t) be a random process. We say that X(t) has limit
in the mean X(t0 ) for t t0 if
lim E(X(t) X(t0 ))2 = 0.
tt0
We write
l. i. m. X(t) = X(t0 ).
tt0
Theorem 1.2. Let F (t) be a dissolution profile and (t) be a random process
with properties E(t) = F (t) and var(t) = o(F 2 (t)) for t small. Then
(t)
l. i. m. = 1. (1.41)
t0 F (t)
18
1.2. STOCHASTIC MODELS
Proof. We have
F (t)
2 z }| {
(t) E(t)2 2 E(t)
lim E 1 = lim 2 + 1
t0 F (t) t0 F (t) F (t)
var(t) + (E(t))2
= lim 1
t0 F 2 (t)
var(t)
= lim 2 = 0.
t0 F (t)
with coefficients
1
(t) = ln F (t) 2 t,
2
2 (t) = 2 t,
E(t) = F (t),
var(t) = F 2 (t) exp( 2 t) 1 .
Here we can see that random process (t) fulfill conditions mentioned in
Theorem 1.2. Model described by stochastic differential equation (1.39) with
analytical solution (1.42) was deduced from model described by (1.28) on
which additional conditions were imposed. Disadvantage of this model is
that its variance, var(t), tends to infinity for large t. On the other hand,
this model prevents the dissolution data to drop bellow zero and permits the
data fluctuation around 100%, as we can see in Fig. 1.7. Furthermore, in
Fig. 1.8 (A) can be seen that sample paths of model (1.37) are similar to
sample paths of solution to stochastic differential equation (1.42) depicted in
Fig. 1.7 (B).
19
1.2. STOCHASTIC MODELS
Model (1.42) satisfies our assumptions and thus we deal in the next with
the generalized model with log-normal probability distribution (1.43). Its
parameters (t), 2 (t) we obtain from equations for mean and variance of
log-normal distribution (see ref.[21])
1 2 (t)
E(t) = e(t)+ 2 = F (t), (1.44)
2
2(t)+2 (t)
var(t) = e e (t) 1 = s2 (t), (1.45)
20
Theoretical background
2
Theoretical results needed in the text are presented in this chapter. It con-
tains basic theory of regular densities, Fisher information and maximum
likelihood estimation, including formulae for its numeric solution. Informa-
tion were taken mostly from ref. [1],[2],[20],[27] where more details can be
found.
21
2.1. FISHER INFORMATION AND ITS LOWER BOUND
Theorem 2.1. Let f (x) = f (x; ) be a regular density. Then score vector
U = U() of the density f (x) has mean EU = 0 and variance varU = J.
Theorem 2.2. Let X and Y be two independent random vectors with regular
probability densities fX (x), fY (y) F , and corresponding Fisher information
matrices JX , JY . Then random vector Z = (XT , Y T )T has regular probability
density
fZ (z) = fX (x)fY (y), z = (xT , yT )T , (2.4)
and Fisher information matrix takes form JZ = JX + JY .
Proof. Conditions of regularity (a) (d) can be easily verified for joint prob-
ability density (2.4). For every component JijZ of the Fisher information
matrix JZ in i-th column and j-th row it holds
22
2.1. FISHER INFORMATION AND ITS LOWER BOUND
ZZ
fZ (z; ) fZ (z; ) 1
JijZ = dz
i j fZ (z; )
M M
ZZ
fX (x)fY (y) fX (x)fY (y) 1
= dxdy
i j fX (x)fY (y)
M M
Z Z
fX (x) fX (x) 1 fY (y) fY (y) 1
= dx + dy
M i j fX (x) M i j fY (y)
X 2 ZZ
fX (x) fY (y)
+ dxdy
i,j=1
i j
M M
i6=j
2 Z
X Z
fX (x) fY (y)
= JijX + JijY + dx dy
i,j=1 M i M i
i6=j
= JijX + JijY .
23
2.1. FISHER INFORMATION AND ITS LOWER BOUND
24
2.1. FISHER INFORMATION AND ITS LOWER BOUND
Fisher information (2.2) can be equal to its lower bound (2.11) under
certain conditions. Equality in Cauchy-Schwarz inequality (2.9) appears if
S(x) g() = 0 or if exists such function K() independent of x that
d
d
f (x; )
= K()[S(x) g()], x M.
f (x; )
d ln f (x; )
= K()S(x) g()K(). (2.12)
d
Denote Q() and R() functions satisfying
where H(x) is independent. Denote u(x) = eH(x) and v() = eR() . Then
This density belongs to the exponential family with one parameter. Only
in this case can Rao-Cramer inequality (2.11) became equality. Note that
functions Q() and R() must fulfil conditions mentioned above.
Theorem 2.3 can be generalized for the case of parameter vector .
Theorem 2.4. Let F = {f (x; ) : Rn } be regular system of den-
sities. Let S = S(X) = (S1 , . . . , Sk )T be an unbiased estimator of the
parametric function g() = (g1 (), . . . , gk ())T , where ESi2 < for every
i = 1, . . . , k and every . Let the partial derivative
gi ()
gij0 () = , i = 1, . . . , k; j = 1, . . . , n
j
exists and let
Z
f ()
Si (x) f ()dx = gij0 (), i = 1, . . . , k; j = 1, . . . , n.
M j
25
2.2. MAXIMUM LIKELIHOOD ESTIMATION
Proof of the theorem can be found in ref. [2], [29]. Inequality A B for
two symmetric matrices A and B means the difference A B is positive-
semidefinite matrix. If n = k > 1 and S = b = (b1 , . . . , bn )T is an unbiased
estimate of g() = (1 , . . . , n )T , then H = In , where In is the identity matrix
of size n, and formula (2.15) takes form
b J1 ,
var (2.16)
L(; x) = f (x, ),
bM LE ; X) l(; X)
l( (2.18)
for every .
26
2.2. MAXIMUM LIKELIHOOD ESTIMATION
U() = 0 (2.21)
due to Definition 2.2 of score vector U(). It can be shown (see ref.[1]) that
if equation Z
2 f (x, )
dx = 0, i, j = 1, . . . , n (2.22)
Rm i j
is satisfied for every , then for U0 () (matrix of second partial deriva-
tives of l(; X) with respect to every component of parameter vector ) holds
EU0 () = J,
Iterative methods
The likelihood equations U() = 0 are generally nonlinear with respect to
an unknown parameter , so we can solve them with iterative methods. The
most common methods are Newton-Raphson method
1
bk =
bk1 U0 (
bk1 ) bk1), k = 1, 2, . . .
U( (2.23)
27
Parameter estimation in stochastic
3
models of dissolution
Gaussian model
At first we study stochastic model described by equation (1.22) with Gaussian
probability distribution N(F (t), s2 (t)) of the dissolved fraction data, where
F (t) = F (t; a) is a dissolution profile and s2 (t) = s2 (t; a) is given variance.
Both of them contain only a single unknown parameter a characterizing scale
of dissolution. Probability density f (x, t) = f (x, t; a) is given by (1.21) and
28
3.1. FISHER INFORMATION IN THEORY OF DISSOLUTION
hence
!
d d d 2
da
f (x, t) 1 d s2 (t) 4(x F (t))( da F (t))s2 (t) 2 s (t) (x F (t))2
= da 2 da
f (x, t) 2 s (t) 4s4 (t)
! !
d 2 d d 2
1 s (t) F (t) s (t)
= x2 da
+x da 2 F (t) da 4
2 s4 (t) s (t) s (t)
| {z } | {z }
=A =B
!
d 2 d d 2
1 2 da s (t) F (t) 1 s (t)
+ F (t) 4 F (t) da 2 da 2
2 s (t) s (t) 2 s (t)
| {z }
=C
= Ax2 + Bx + C
and thus
!2
d
f (x, t)
da
= A2 x4 + 2ABx3 + (2AC + B 2 )x2 + 2BCx + C 2 . (3.1)
f (x, t)
Inserting (3.1) into (2.2) gives, for = a and f (x, t) given by (1.21), equation
Z Z
2 4
J(t) = A x f (x, t)dx + 2AB x3 f (x, t)dx
Z Z
2 2
+(2AC + B ) x f (x, t)dx + 2BC xf (x, t)dx + C 2
= A F (t) + 6s (t)F (t) + 3s (t) + 2AB F 3 (t) + 3s2 (t)F (t)
2 4 2 2 4
+(2AC + B 2 ) F 2 (t) + s2 (t) + 2BCF (t) + C 2 , (3.2)
where the last equation follows evaluation of moments for Gaussian distri-
bution with mean F (t) and variance s2 (t), see ref. [22]. After a tedious
calculation we obtain formula describing Fisher information
!2 2
d 2 d
1 da s (t) da
F (t)
J(t) = + . (3.3)
2 s2 (t) s2 (t)
29
3.1. FISHER INFORMATION IN THEORY OF DISSOLUTION
After inserting specific functions F (t) and s2 (t) into (3.3) we can find the
optimal time topt , where
topt = argmax J(t). (3.4)
t(0,)
Log-normal model
Let us investigate stochastic model of dissolution described by formula (1.42)
with log-normal distribution logN((t), 2 (t)) of the dissolved fraction data.
Parametric functions (t) = (t; a) and 2 (t) = 2 (t; a) are defined by (1.46)
and (1.47), and contain only a single unknown parameter a characterizing
scale of dissolution. Inserting probability density (1.36) into formula (2.2),
= a, gives with similar approach as in the previous case
!2 2
d 2 d
1 da (t) (t)
J(t) = + da 2 , (3.5)
2 2 (t) (t)
After inserting specific functions F (t) and s2 (t) into (1.46), (1.47) and conse-
quently into (3.5) we can find the optimal time topt as a solution of (3.4). One
can see that formula (3.3) with parameters of Gaussian distribution (1.20)
is similar to formula (3.5) with parameters of log-normal distribution (1.43)
given by (1.46) and (1.47).
Multiple measurements
If we can measure at m 2 time instants given by vector t = (t1 , . . . , tm )T ,
where ti > 0 for i = 1, . . . , m, and the measurements can be considered
to be mutually independent, then the Fisher information J(t) has due to
Theorem 2.2 form m
X
J(t) = J(ti ). (3.6)
i=1
Thus if the optimal time topt exists, and such index i {1, . . . , m} that
ti 6= topt exists, then it holds
m
X m
X
J(t) = J(ti ) J(topt ) = mJ(topt ).
i=1 i=1
30
3.1. FISHER INFORMATION IN THEORY OF DISSOLUTION
31
3.1. FISHER INFORMATION IN THEORY OF DISSOLUTION
At the time instant topt.a we should obtain the best estimate of parameter a,
but not the best estimate of parameter b and vice versa.
Determinant (3.9) of Fisher information matrix (3.8) can be zero, espe-
cially if the variance s2 (t) is parameter independent. The case of a singular
Fisher information matrix represents a significant complication for the the-
ory of the Rao-Cramer lower bound and is usually handled by resorting to
the pseudoinverse of the Fisher matrix, see ref. [29]. It has been shown
there, that the Rao-Cramer lower bound does not exist for estimation prob-
lems with singular Fisher information matrix, and thus we can not find the
optimal times.
Dissolution profile F (t) is given, thus singularity of Fisher information
matrix (3.8) depends on form of the function s2 (t). For example, the deter-
minant is zero if variance of measured data has form
N
X
s2 (t) = i F i (t), (3.12)
i=0
32
3.1. FISHER INFORMATION IN THEORY OF DISSOLUTION
s2a 0 (t) = Fa0 (t)P (t), s2b 0 (t) = Fb0 (t)P (t),
where
N
X
P (t) = ii F i1 (t), (3.13)
i=1
thus
s2a 0 (t)Fb0 (t) s2b 0 (t)Fa0 (t) 0 (3.14)
which gives |J(t)| 0 for variance s2 (t) given by (3.12). Hence, we are not
able to find the optimal times especially if the variance takes form
with determinant
2
[ 2 0 (t)0b (t) b2 0 (t)0a (t)]
|J(t)| = a 0. (3.17)
2 6 (t)
If the determinant is nonzero, the inverse Fisher information matrix has form
2 0 2
1 b (t) (0b (t))2 20 20
1 a (t)b (t) 0a (t)0b (t)
1 2 2 (t) + 2 (t) 2 4 (t) 2 (t)
.
J1 (t) = 2 2
2
|J(t)| 1 a (t)b (t) a (t)b (t)
0 0 0 0
1 a2 0 (t)
+ (0a (t))2
2 4 (t) 2 (t) 2 2 (t) 2 (t)
(3.18)
The optimal times topt.a and topt.b correspond to the argument minima of the
diagonal functions of matrix J1 (t) as given by (3.11). Again, if the Fisher
information matrix is singular, we can not find the Rao-Cramer bounds nor
33
3.1. FISHER INFORMATION IN THEORY OF DISSOLUTION
the optimal times. Similarly as in the previous case it can be shown that
determinant (3.17) is zero if
N
X
2
(t) = i i (t), (3.19)
i=0
a2 0 (t) = s2a 0 (t)P (t) Fa0 (t)Q(t), b2 0 (t) = s2b 0 (t)P (t) Fb0 (t)Q(t)
1 1
0a (t) = Fa0 (t)R(t) s2a 0 (t)P (t), 0b (t) = Fb0 (t)R(t) s2b 0 (t)P (t),
2 2
where
F (t) 2s2 (t)
P (t) = , Q(t) = ,
F (t)(F (t) + s2 (t))
2 F (t)(F 2(t) + s2 (t))
1 1
R(t) = Q(t) + .
2 F (t)
It holds
1 20 1
sa (t)s2b 0 (t)P 2 (t) + Fa0 s2b 0 (t)P (t)Q(t),
2 2
and
1
a2 0 (t)0b (t)b2 0 (t)0a (t)
= Fa0 (t)s2b 0 (t)
Fb0 (t)s2a 0 (t)
P (t)R(t) P (t)Q(t) .
2
(3.20)
2
For variance s (t) given by (3.12) holds (3.14), hence |J(t)| 0 for Fisher
information matrix (3.16).
34
3.1. FISHER INFORMATION IN THEORY OF DISSOLUTION
Multiple measurements
If the measurements can be taken at m 2 time instants given by vector
t = (t1 , . . . , tm )T , ti > 0 for i = 1, . . . , m, and the dissolution data are mu-
tually independent, then the Fisher information matrix of parameter vector
= (a, b)T takes due to Theorem 2.2 form
m
X
J(t) = J(ti ), (3.21)
i=1
|J(t)| = (J11 (t1 ) + J11 (t2 ))(J22 (t1 ) + J22 (t2 )) (J12 (t1 ) + J12 (t2 ))2
2
2
= J11 (t1 )J22 (t1 ) J12 (t1 ) + J11 (t2 )J22 (t2 ) J12 (t2 ) +
+J11 (t1 )J22 (t2 ) + J11 (t2 )J22 (t1 ) 2J12 (t1 )J12 (t2 )
= |J(t1 )| + |J(t2 )| +
+J11 (t1 )J22 (t2 ) + J11 (t2 )J22 (t1 ) 2J12 (t1 )J12 (t2 ). (3.24)
If the determinant is not zero, then the inverse Fisher information matrix
has form
1 1 J22 (t1 ) + J22 (t2 ) J12 (t1 ) J12 (t2 )
J (t) = (3.25)
|J(t)| J12 (t1 ) J12 (t2 ) J11 (t1 ) + J11 (t2 )
35
3.1. FISHER INFORMATION IN THEORY OF DISSOLUTION
(1) (2)
thus if the optimal time (topt , topt )T of any of the parameter exists, then
(2) (1)
(topt , topt )T is the optimal time too.
If Fisher information matrix J(t) for a single measurement is singular,
then the Fisher information matrix (3.23) does not necessarily need to be
singular too. We show it for Gaussian model (1.22) with variance s2 (t) given
by (3.12). For log-normal model (1.43) the approach is similar. Components
of Fisher information matrix J(t) has form
2
1 Fa0 (t)P (t) [Fa0 (t)]2
J11 (t) = + ,
2 s2 (t) s2 (t)
2
1 Fb0 (t)P (t) [Fb0 (t)]2
J22 (t) = + ,
2 s2 (t) s2 (t)
1 Fa0 (t)Fb0 (t)P 2 (t) Fa0 (t)Fb0 (t)
J12 (t) = + ,
2 s4 (t) s2 (t)
Here we can see that Fisher information matrix (3.23) of Gaussian model
(1.22) with variance (3.12) is singular if
36
3.2. PARAMETERS OF DISSOLUTION MODELS AND THEIR
MAXIMUM LIKELIHOOD ESTIMATION
It holds
1 (Ci F (ti ))2
ln fi () = ln 2s2 (ti )
2 2s2 (ti )
hence
F (ti ) 2 s (ti ) 2
2 2
1 j s (ti ) 2 j s (ti )(Ci F (ti )) + j (Ci F (ti ))
ln fi () = +
j 2 s2 (ti ) 2s4 (ti )
2
1 j s (ti ) (Ci F (ti ))2 F (ti ) Ci F (ti )
= 2 2
1 +
2 s (ti ) s (ti ) j s2 (ti )
37
3.2. PARAMETERS OF DISSOLUTION MODELS AND THEIR
MAXIMUM LIKELIHOOD ESTIMATION
38
3.2. PARAMETERS OF DISSOLUTION MODELS AND THEIR
MAXIMUM LIKELIHOOD ESTIMATION
39
3.3. EXAMPLES
3.3 Examples
3.3.1 Stochastic homogenous model
As an example of stochastic model with single parameter = a we take
stochastic homogenous model with mean specified by (1.7). Form of its
variance s2 (t) influences substantially time course of the Fisher information,
thus we introduce several examples to illustrate this fact. Analytical form of
the Fisher information for log-normal model (3.5) is complicated in all cases
and is not given here.
Example 1
The simplest example we can take is stochastic homogenous model with
constant variance,
s2 (t) r, r > 0. (3.35)
Fisher information of Gaussian model (1.22) takes, after substitution (1.7)
and (3.35) into formula (3.3), form
t2 e2at
J(t) = J2 (t) = . (3.36)
r
For variance given by (3.35), Fisher information (3.36) satisfies J(0) = 0 and
limt J(t) = 0. Optimal time for this model is obtained from (3.4) and has
analytic form topt = 1/a. Time courses of Fisher information J(t) given by
(3.3) and (3.5) for Gaussian and log-normal models are shown in Fig. 3.1.
Example 2
The stochastic homogenous model with nonconstant variance
s2 (t) = peat 1 eat , p > 0, (3.37)
40
3.3. EXAMPLES
These functions are plotted together with Fisher information (3.5) of log-
normal model in Fig. 3.1. There it can be seen that functions J(t) tend to
infinity for increasing t in this example. This is caused by formal continuation
of dissolution in model (1.7) for large t while variance s2 (t) tends to zero. In
this case Fisher information J(t) gives no optimal time according to (3.4).
We have a natural requirement that no information about the dissolution
process can be obtained when it is finished. Example of Fisher information
with variance (3.37) contradicts that assumption, hence the model cannot be
accepted.
Example 3
In the last example we take into account the measurement error. Variance
(3.37) of the stochastic homogenous model can be, for example, modified to
form
s2 (t) = peat 1 eat + q(1 eat ), p > 0, q > 0. (3.40)
Analytic form of the Fisher information is complicated and is not given here.
Lower bound of the Fisher information has form
t2 e2at
J2 (t) = . (3.41)
(1 eat )(peat + q)
where p > 0, r > 0 are constants, then the course of the Fisher information is
very similar to the one with variance (3.40) depicted in Fig. 3.1. We can see
that for behavior of the Fisher information of stochastic homogenous model
is crucial variance which does not tend to zero for increasing t.
41
3.3. EXAMPLES
15 (B)
0.03 (A)
0.025
10
0.02
J(t)
s2 (t)
0.015
0.01 5
0.005
0 0
0 1 2 3 4 0 1 2 3 4
t t
12 7
(C) (D)
6
10
5
8
4
J(t)
J(t)
6
3
4
2
2 1
0 0
0 1 2 3 4 0 1 2 3 4
t t
Fig. 3.1: Dependence of the Fisher information on different forms of variance s2 (t)
for stochastic homogenous model with mean (1.7), a = 1. (A) s2 (t) given by (3.35),
r = 0.015 (black), s2 (t) given by (3.37), p = 0.1 (red) and s2 (t) given by (3.40), p = 0.1
and q = 0.01, (B) Fisher information of homogenous Gaussian model (3.36) (black) and
log-normal model (3.5) (red) corresponding to variance (3.35). The optimal times are
topt = 1 for Gaussian model and topt = 0.935 for log-normal model, (C) Fisher information
of Gaussian model (3.38) (black), log-normal model (3.5) (red) and lower bound (3.39)
(blue) corresponding to variance (3.37). Lower bound J2 (t) has maximum at t = 1.593
(D) Fisher information of Gaussian model (3.3) (black), log-normal model (red) and lower
bound (3.41) (blue) corresponding to variance (3.40). The optimal times are topt = 1.266
for Gaussian model, topt = 1.163 for log-normal model. Lower bound has maximum at
t = 1.186.
42
3.3. EXAMPLES
0.06
(A) (B)
0.05 0.04
0.04 0.03
s2 (t)
s2 (t)
0.03
0.02
0.02
0.01
0.01
0 0
0 1 2 3 0 1 2 3
t t
Fig. 3.2: Different functions of variance s2 (t). (A) s2 (t) corresponding to Weibull model
(1.8) with parameters a = 1, b = 2 of the form (3.43), p = 0.1 (black), variance (3.44),
p = 0.1, r = 0.02 (blue) and variance (3.45), p = 0.1, q = 0.02 (red). (B) Variance
corresponding to Hixson-Crowell model (1.16), a = 0.5, b = 2, of the form (3.35), r = 0.02
(black), variance (3.46), p = 0.1 (blue) and variance (3.47), p = 0.1, r = 0.015 (red).
Example 1
As the first example of stochastic Weibull model with variance that gives
nonzero determinant of the Fisher information matrices described in Section
3.1.2 we take function
b
s2 (t) = ptb (1 F (t)) = ptb eat , p>0 (3.43)
43
3.3. EXAMPLES
stochastic models are nonzero. It can be easily verified that (3.43) satis-
fies our assumptions about the variance of the dissolution given in Section
1.2.1. Inserting F (t) and s2 (t) into (3.10) and (3.18) gives the inverse Fisher
information matrices of Gaussian and log-normal model. Time courses of
1 1
diagonal functions J11 (t) and J22 (t) are shown in Fig. 3.3.
1
There the discontinuity of function J22 (t) at the time instant t = 1 can be
seen, which is due to presence of logarithm in denominator of the function.
Despite the one-parameter model had increasing Fisher information, resp.
decreasing Rao-Cramer bound if its variance tended to zero for increasing t,
this does not hold for the two-parameter model.
For model with Gaussian distribution, the optimal time to measure for
parameter a can be evaluated from equation (3.11) and has analytic form
topt.a = a1/b . The parameter b has optimal time to measure topt.b = 0 which
implies that for the best estimate of parameter b we have to measure when
nothing, or only very small amount of a solvent is dissolved. This is analogous
to previous examples of single-parameter model, where the models without
measurement errors had the optimal time at the infinity, i.e. when everything
is dissolved.
For model with log-normal distribution, the optimal time of the param-
eter a is close to the optimal time of the Gaussian model, as one can see in
Fig.3.3. In contrast to the Gaussian model, the optimal time for estimation of
10 100
(A) (B)
80
8
60
(t)
(t)
6
1
1
J11
J22
40
4
20
2 0
0 0.5 1 1.5 2 0 0.5 1 1.5 2 2.5
t t
Fig. 3.3: Rao-Cramer bounds for stochastic Weibull model with mean (1.8) and variance
(3.43), a = 1, b = 2, p = 0.1. Rao-Cramer bound for (A) parameter a and (B) parameter
b of Gaussian (black) and log-normal (red) stochastic models. The optimal times are
topt.a = 1, topt.b = 0 (Gaussian model) and topt.a = 1.042, topt.b = 0.259 (log-normal
model).
44
3.3. EXAMPLES
the parameter b is positive. In the example depicted in the Fig.3.3 (B) the op-
timal time topt.b = 0.259. But for given parameters it holds F (topt.b ) = 0.065,
which can be interpreted as the best time for measure is when only 6.5% of
the solvent is dissolved.
Example 2
As the second example we take stochastic Weibull model with variance (3.43)
with added constant measurement error,
b
s2 (t) = ptb eat + r, p > 0, r > 0, (3.44)
where a, b are parameters of Weibull model (1.8). Inserting F (t) and s2 (t)
into (3.10) and (3.18) gives the inverse Fisher information matrices of Gaus-
1
sian and log-normal model. Time courses of their diagonal functions J11 (t)
1
and J22 (t) are shown in Fig. 3.3. Discontinuity of the Rao-Cramer bound of
the parameter b at the time instant t = 1 can be seen there due to presence of
1
logarithm in denominator of function J22 (t). The optimal times to measure
can be obtained from (3.11). For a model with Gaussian distribution, adding
a constant measurement error to variance (3.43) has no effect to value of the
optimal time topt.a , but the optimal time topt.b is no longer zero.
40 200
(A) (B)
35
30 150
(t)
(t)
25
1
1
100
J22
J11
20
15
50
10
5
0 0.5 1 1.5 2 2.5 0 0.5 1 1.5 2 2.5
t t
Fig. 3.4: Rao-Cramer bounds for stochastic Weibull model with mean (1.8) and variance
(3.44), a = 1, b = 2, p = 0.1, r = 0.02. Rao-Cramer bound for (A) parameter a and (B)
parameter b of Gaussian (black) and log-normal (red) stochastic models. The optimal
times are topt.a = 1, topt.b = 0.382 (Gaussian model) and topt.a = 1.077, topt.b = 0.533
(log-normal model).
45
3.3. EXAMPLES
Example 3
As the last example we take stochastic Weibull model with variance
b
b
s2 (t) = ptb eat + q 1 eat , p > 0, q > 0, (3.45)
40 200
(A) (B)
35
30 150
25
(t)
(t)
100
1
1
J22
J11
20
15
50
10
5
0
0 0.5 1 1.5 2 2.5 0 0.5 1 1.5 2 2.5
t t
Fig. 3.5: Rao-Cramer bounds for stochastic Weibull model with mean (1.8) and variance
(3.45), a = 1, b = 2, p = 0.1, q = 0.02. Rao-Cramer bound for (A) parameter a and (B)
parameter b of Gaussian (black) and log-normal (red) stochastic models. The optimal
times are topt.a = 0.719, topt.b = 0.056 (Gaussian model) and topt.a = 0.862, topt.b = 0.25
(log-normal model).
46
3.3. EXAMPLES
Example 1
The simplest example of stochastic Hixson-Crowell model is with a constant
variance (3.35). Inverse diagonal functions of the Fisher information matrix
1 1
of Gaussian model (1.22), 1/J11 (t) and 1/J22 (t), are shown in Fig 3.6.
We use this approach for better view as the diagonal functions have large
number of discontinuities. We can see that despite variance (3.35) gives no
single optimal time due to singularity of the Fisher information matrix, for
measurements in two different time instants it gives optimal times for both
parameters.
(A) (B)
5
0.15
(t1 , t2 )
(t1 , t2 )
4
3 0.1
1
2
1/J11
1/J22
0.05
1
0 0
2 2
2 2
1 1
1 1
t1 0 0 t2 t1 0 0 t2
Fig. 3.6: Inverse Rao-Cramer bounds for Gaussian model (1.22) with mean described
by (1.16) of Hixson-Crowell model and variance given by (3.35), a = 0.5, b = 2, r = 0.02.
(A) inverse Rao-Cramer bound of scale parameter a and (B) inverse Rao-Cramer bound of
shape parameter b. The optimal times are topt.a = (0.398, 1.6)T , topt.b = (0.467, 1.647)T .
47
3.3. EXAMPLES
Example 2
As the second example we take stochastic Hixson-Crowell model with vari-
ance without measurement error
2 p(1 at)b 1 (1 at)b for t [0, a1 ]
s (t) = , p > 0. (3.46)
0 for t > a1
(A) (B)
500 1
(t1 , t2 )
(t1 , t2 )
400
300
0.5
1
200
1/J11
1/J22
100
0 0
2 2
2 2
1 1
1 1
t1 0 0 t2 t1 0 0 t2
Fig. 3.7: Inverse Rao-Cramer bounds for Gaussian model (1.22) with mean described
by (1.16) of Hixson-Crowell model and variance given by (3.46), a = 0.5, b = 2, p = 0.1.
(A) inverse Rao-Cramer bound of scale parameter a and (B) inverse Rao-Cramer bound
of shape parameter b. The optimal times are topt.a = (1.235, 2)T , topt.b = (1.284, 2)T .
48
3.3. EXAMPLES
Example 3
As the last example we take
2 p(1 at)b 1 (1 at)b + r for t [0, a1 ]
s (t) = , p > 0, r > 0.
0 for t > a1
(3.47)
Inverse diagonal elements of the Fisher information matrix of Gaussian model
1 1
(1.22), 1/J11 (t) and 1/J22 (t), are shown in in Fig 3.8. We can see that for
behavior of the Fisher information of stochastic Hixson-Crowell model is
crucial variance which does not tend to zero for t 1/a.
(A) (B)
6 0.15
(t1 , t2 )
(t1 , t2 )
4 0.1
1
1
1/J11
1/J22
2 0.05
0 0
2 2
2 2
1 1
1 1
t1 0 0 t2 t1 0 0 t2
Fig. 3.8: Inverse Rao-Cramer bounds for Gaussian model (1.22) with mean described by
(1.16) of Hixson-Crowell model and variance given by (3.47), a = 0.5, b = 2, p = 0.1, r =
0.015. (A) inverse Rao-Cramer bound of scale parameter a and (B) inverse Rao-Cramer
bound of the shape parameter b. The optimal times are topt.a = (0.431, 1.697)T , topt.b =
(0.533, 1.744)T .
49
Computational procedures and
4
examples
50
4.2. MAXIMUM LIKELIHOOD ESTIMATION
>> t=0:0.1:5;
>> plot(t, wiener(t));
where and are given functions, and W (t) is a standard Wiener process
given by Definition 1.1. With similar notation as in the previous case we
can obtain numerical approximation n (t) at n + 1 equidistant time instants
ti = it, i = 0, . . . , n, from recursive equation (see ref.[5],[13],[30])
n (ti ) = n (ti1 ) + (ti1 , n (ti1 ))t + (ti1 , n (ti1 ))Ni t (4.2)
Example 1
In the first example we investigate error behavior of the maximum likelihood
estimate of a single parameter a of stochastic homogenous model with mean
51
4.2. MAXIMUM LIKELIHOOD ESTIMATION
52
4.2. MAXIMUM LIKELIHOOD ESTIMATION
53
4.2. MAXIMUM LIKELIHOOD ESTIMATION
0.02
(A) (B)
0.018 0.016
0.016
a)2
a)2
0.014
(a b
(a b
0.014
0.012
0.012
0.01
0.01
0.008 0.008
0.5 1 1.5 2 2.5 0.5 1 1.5 2 2.5
t t
Example 2
In the second example we investigate error behavior of the maximum likeli-
hood estimate of parameter vector (a, b)T of stochastic Weibull model with
mean described by (1.8). The dissolved fraction data are Monte-Carlo sim-
ulated in Matlab with function generate_weibull_data(m,t,a,b,type),
where m is required number of observations at the time instants given by
vector t, a and b are parameters of model (1.8) and type determines proba-
bility distribution of simulated data. This parameter can take value g for
Gaussian distribution or l for log-normal distribution. Variance s2 (t) is
given by function
b b
s2 (t) = 0.02eat (1 eat ) + 0.0005, (4.4)
54
4.2. MAXIMUM LIKELIHOOD ESTIMATION
given by (3.29) and assumes Gaussian distribution of the data. The func-
tion mp_lmle(data,a0,b0) uses approach given by (3.34) and assumes log-
normal distribution of the data in the problem. For numeric solution of
given equations we use score method (2.24) again. Matrix data=[t C] is
output of function generate_weibull_data and variables a0 and b0 are ini-
tial approximations of the parameters a and b. This initial approximation
can be obtained
P from function mp_fit(data), which fits sample averages
Cj = m1 m C
i=1 ij of the data simulated at the time instant tj , j = 1, 2, with
Weibull dissolution profile (1.8), i.e.
1 exp(atb1 ) = C1 , 1 exp(atb2 ) = C2 ,
and returns an initial approximation ab0 and bb0 obtained from equations
! ln tln1 ln
t2
t2
ln 1 C2
ab0 = ln 1 C2 ,
ln 1 C1
ln(1C2 )
ln ln(1C1 )
bb0 = .
ln t2 ln t1
Let us select a = 1, b = 2 in arbitrary time units. We investigate sample
variance of estimates a , b based on m = 4 dissolved fraction data simulated
at two time instant ti , ti + 0.5 for i = 1, . . . , 11, given by vector
For example, at the time instant t1 = 0.3 we obtain random sample from
Gaussian distribution
>>data = generate_weibull_data(4, [0.3 0.8], 1, 2, g)
data =
0.3000 0.0924
0.3000 0.1178
0.3000 0.0970
0.3000 0.0670
0.8000 0.3936
0.8000 0.5604
0.8000 0.4801
0.8000 0.4704
At first we need intitial approximations of the parameters. We use approach
described above:
55
4.2. MAXIMUM LIKELIHOOD ESTIMATION
>>theta_0=mp_fit(data)
theta_0 =
0.9926
1.9212
, b of the parame-
At each of the time instants now we can obtain estimates a
ters a and b from the functions mp_gmle and mp_lmle. For example, for the
data given above we obtain
This estimation procedure is done for every ti given by vector t and for each
time instant we obtain matrix of two sample variance in two columns, first
for parameter a and second for b, where rows of the matrix correspond to
specific estimation method. These vectors vary in dependency on simulated
data. To gain a more realistic picture, we take average of these sample
variances. Matlab function mp_point_estimate(r,m,t_1,t_2,a,b,type)
simulate r times m dissolved fraction data at the time instant t_1 and t_2
and returns average matrix of sample variances of estimates a and b. They
are plotted in Fig. 4.2 and Fig. 4.3.
56
4.2. MAXIMUM LIKELIHOOD ESTIMATION
0.25
0.04 (A) (B)
0.2
0.03
a)2
(b bb)2
(a b
0.15
0.02
0.01 0.1
0 0.05
0.2 0.4 0.6 0.8 1 0.2 0.4 0.6 0.8 1
t t
0.025
(A) (B)
0.15
0.02
a)2
(b bb)2
(a b
0.015
0.1
0.01
0.005
0.05
0.2 0.4 0.6 0.8 1 0.2 0.4 0.6 0.8 1
t t
57
Summary
58
Bibliography
[12] Hixson A. W., Crowell J.H. (1931) Dependence of reaction velocity upon
surface and agitation, Ind. Eng. Chem. 23: 923-931.
59
BIBLIOGRAPHY
[15] Lansky P., Lanska V., Weiss M. (2004) A stochastic differential equation
model for drug dissolution and its parameters, J. Control. Release 100:
267-274.
[18] Lansky P., Weiss, M. (2003) Does the dose solubility ratio affect the
mean dissolution time of drugs?, Pharm. Res. 8: 1632-1647.
[19] Lee C.J. et al. (1999) Analysis of drug dissolution data, Statistics in
medicine 18: 799-814.
[24] OHara T., Dunne A., Butler J., Devane J. (1998) A review of methods
used to compare dissolution profile data, Pharm. Sci. & Tech. Today 5:
214-223.
[26] Polli J. E., Rekhi G. S., Augsburger L. L., Shah V. P. (1997) Meth-
ods to compare dissolution profiles and a rationale for wide dissolution
specification for metoprolol tablets, J. Pharm. Sci. 6: 690-700.
60
BIBLIOGRAPHY
[27] Rao C.R. (1962) Linear Statistical Inference and its Applications, Wiley,
New-York.
[34] Zeev, E. (1997) On the variability of dissolution data, Pharm. Res. 10:
1355-1362
61