Académique Documents
Professionnel Documents
Culture Documents
1. Introduction. Many practical problems can be transformed into the computations of multivariate integrals:
f (x) dx.
Is (f ) =
[0,1]s
Typical examples are the valuations of nancial derivatives. In principle, any stochastic simulation whose purpose is to estimate an expectation ts this framework. Highdimensional integrals are usually approximated by Monte Carlo (MC) or quasi-Monte
Carlo (QMC) algorithms:
Qn,s (f ) := Qn,s (f ; P ) :=
n1
1
f (xk ),
n
k=0
Received
by the editors January 24, 2006; accepted for publication (in revised form) October
11, 2006; published electronically March 30, 2007. This work was supported by the National Science
Foundation of China.
http://www.siam.org/journals/sisc/29-2/65071.html
Department of Mathematical Sciences, Tsinghua University, Beijing 100084, China (xwang@
math.tsinghua.edu.cn).
598
599
where P := {x0 , x1 , . . . , xn1 } is a set of random points in MC or deterministic quasirandom numbers in QMC. The MC methods converge as O(n1/2 ), independently of
s. QMC methods try to take the points more uniformly distributed and have the
potential to improve the convergence rate for some classes of functions. Digital nets
and good lattice points are two important classes of QMC; see [18, 20]. For example,
a rank-1 lattice rule has the form
n1
1
kz
Qn,s (f ) =
,
f
n
n
k=0
s
j=1
fj (xj ) +
fij (xi , xj ),
1i<js
can usually provide a quite satisfactory approximation, where fj (xj ) and fij (xi , xj )
represent the individual and the cooperative eects. For numerical integration of such
functions the quality of low-order (especially the order-1 and order-2) projections of
the node set is crucial. QMC point sets should be constructed to be as well distributed
as possible in their low-order projections. We believe that good quality in order-2
projections of a lattice point set is necessary and sucient for many problems in
nance. Point sets with optimal order-1 projections are not dicult to obtain (an
arbitrary lattice point set has perfect order-1 projections). However, it is not so easy
(or even impossible) to obtain point sets with optimal projections of various orders
simultaneously.
It was shown in [29] that in high dimensions (say, s 50) common quasi-random
numbers, such as the Sobols [25], have excellent order-1 projections (as expected),
but often exhibit patterns in order-2 and higher-order projections and that these projections are no more uniform on the average than those of random points for practical
n (which may contradict common intuition). In fact, this nonuniformity is typical for
quasi-random numbers [16]. It is interesting to know whether good lattice points could
provide better order- 2 and higher-order projections than digital nets do. What is the
optimal situation (i.e., the logical limit) which can be achieved by a construction, say,
the Korobov construction [12]? How does one construct lattice points with optimal
low-order (or some specic order) projections? How ecient and robust are they for
practical applications? The method in [6] provides a theoretical basis for answering
these questions (but no generators are published there).
Confronted with such issues, the objectives of this paper are to explore the potential possibility of good lattice points in achieving good low-order projections (aiming
at constructing robust lattice rules) and to study their applications to nance. Lattice rules are usually found by minimizing some quality measures; thus the choice of
measure is important [7, 14, 8, 9, 28, 30]. To construct lattice points with low-order
600
XIAOQUN WANG
projections as well distributed as possible, one needs a suitable measure. The classical
quality measure P (see [20]) was found to be not suitable in high dimensions for this
purpose (though it may not be bad in small dimensions, say s 8), since it treats
implicitly higher-order projections as more important than lower-order ones and does
not put special emphasis on low-order projections, which may cause serious problems
in high dimensions; see [28]. It was shown in [28] that the lattice rules constructed
based on P may not give good results for high-dimensional problems (in fact, they
can be even worse than MC, and this is why lattice rules are not so popular for nance
problems). The reason will become clear in section 4.
In [28] an attempt was made to choose suitable weights involved in the weighted
version of P by relating the weights to the sensitivity indices of a particular problem.
The weighted version of P is related to the worst-case error of certain weighted function spaces. We refer to [23] for the theory of weighted function spaces. A principal
diculty of using the theory of weighted space for practical applications is in choosing
appropriate weights. The lattice rules constructed in [28] are extremely eective, but
they are problem-dependent. The high eciency of those rules is achieved by focusing
on important dimensions and low-order projections. Based on a similar idea, in this
paper we attempt to construct robust lattice points, which are of acceptable quality
for a wide range of problems.
This paper is organized as follows. In section 2 we introduce weighted Sobolev
spaces and discuss the relationship of the uniformity of the projections with the
quadrature error. In section 3 we discuss quality measures for lattice rules and present
methods to construct lattice rules with optimal low-order (or some specic order)
projections. Numerical comparisons on the quality of projections of various orders are
presented in section 4. The potential possibility and the limitations of lattice rules
(i.e., what lattice rules can do and cannot do) are investigated. Applications to the
valuation of path-dependent options and American options (based on the least-square
MC [15]) are demonstrated in section 5. Numerical experiments in sections 4 and 5
show the robustness property of our lattice rules. Concluding remarks are presented
in section 6. Some generators of lattice rules with optimal order-2 projections are
provided in the appendix.
2. Weighted Sobolev space and quadrature error. In this section we present
some background on weighted function spaces and discuss the factors that aect QMC
quadrature errors.
2.1. Background on reproducing kernel Hilbert space. Let H(K) be a
reproducing kernel Hilbert space (RKHS) with reproducing kernel K(x, y). The kernel
K(x, y) has the following properties:
K(x, y) = K(y, x),
K(, y) H(K)
x, y [0, 1]s
and
f (y) = f, K(, y)
where , is the inner product of H(K) and ||f ||H(K) = f, f 1/2 . Let
K(x, y) dy
x [0, 1]s .
hs (x) =
[0,1]s
f H(K).
601
The worst-case error of the algorithm Qn,s (f ; P ) over the unit ball of H(K), or the
discrepancy of the set P with respect to the kernel K, is dened by
D(P ; H(K)) = sup{|Is (f ) Qn,s (f ; P )| : f H(K), ||f ||H(K) 1}.
Since H(K) is an RKHS, the square worst-case error can be expressed in terms of the
reproducing kernel (see [9, 23]):
(2)
n1
n1
2
1
D (P ; H(K)) =
K(x, y) dxdy
K(xi , x) dx + 2
K(xi , xk ).
n i=0 [0,1]s
n
[0,1]2s
i,k=0
f H(K).
(4)
Ku (xu , yu ),
=uA
(xj , yj ),
ju
with
(x, y) =
1
1
1
B2 ({x y}) + x
y
,
2
2
2
|u|
|u|
g
1
f, gH(Ks, ) =
s,u
dxu
dxu dxu ,
[0,1]|u|
[0,1]s|u| xu
[0,1]s|u| xu
uA
where xu denotes the vector of the coordinates of x with indices in u, u denotes the
complementary set A u, and |u| denotes the cardinality. The
term corresponding to
u = in the inner product is interpreted as [0,1]s f (x)dx [0,1]s g(x)dx. (We assume
that s, = 1.) The weight s,u measures the importance of the group of variables in
u. If ||f ||H(Ks, ) 1 and s,u is small, then
|u| f
xu
is small.
602
XIAOQUN WANG
n1
1
n2
i,k=0 =uA
s,u
(xi,j , xk,j ).
ju
s
j=1
i,k=0
and
||f ||2H(Ks, ) =
||fu ||2H(Ks, ) .
uA
Note that for any f H(Ku ), we have ||f ||H(Ku ) = ||f ||H(Ks, ) .
For functions of H(Ks, ) the projection decomposition (6) coincides with the
ANOVA decomposition; see [5]. The ANOVA decomposition of a function is dened
recursively by
fu (x) =
f (x)dxu
fv (x) for = u A,
[0,1]s|u|
vu
603
n1
s,u
(xi,j , xk,j ).
n2
ju
i,k=0
|Is (f ) Qn,s (f )|
=uA
(7)
=uA
the integration error Is (f ) Qn,s (f ) can be bounded in terms of D() (P ) (after using
the CauchySchwarz inequality in (7)):
|Is (f ) Qn,s (f )|
s
=1
where
(8)
||f() ||2 =
||fu ||2H(Ks, ) .
uA, |u|=
The quantity D() (P ) is called the order- discrepancy; see [29]. It measures the
uniformity of the projections of order- taken together and is the worst-case error of
the algorithm Qn,s (f ) over the unit ball of the RKHS H(K() ), i.e.,
D() (P ) = D(P ; K() ),
where K() (x, y) is the reproducing kernel
K() (x, y) :=
(9)
uA, |u|=
Ku (xu , yu )
604
XIAOQUN WANG
and (8) gives the square norm of functions in the RKHS H(K() ).
Instead of trying to nd point sets P such that D(Pu ; H(Ku )) is small for all
u A, one may try to nd P such that D() (P ) is small for = 1, 2, . . . , s or at least
for 3. It is shown that for relatively small n (in 1000s) and large s (say, s = 64),
the Sobol points have order-2 and higher-order discrepancies no smaller than those
of random points [29]. It is challenging to construct point sets with smaller order-2
and higher-order discrepancies. We are interested in whether good lattice points can
do better.
The order- discrepancy has two drawbacks. First, for an arbitrary point set P
and arbitrary weights s,u , computing D() (P ) is expensive: direct computation based
on its denition is equivalent to computing D(Pu ; H(Ku )) for each subset u A with
|u| = , the number of which is O(s ). Second, the order- discrepancy does not have
the shift-invariant property: for a point set P , its shifted version P + := {{xk +
}, k = 1, . . . , n 1}, where [0, 1)s , may have a dierent order- discrepancy.
For example, consider a point
set P := {k/n : k = 0, . . . , n 1} in [0, 1]. It is easy to
calculate that D(1) (P ) = 3/(3n). However, for = 1/(2n), the shifted point set is
It turns out that this is related to the theory of the shift-invariant kernel.
3. The construction of good lattice rules. We are interested in constructing
good lattice points such that their mean-squared order- discrepancies (as dened
above) are small, at least for 3. A suitable quality measure is crucial for this
purpose. The theory of the shift-invariant kernel is useful.
3.1. Shift-invariant kernel. For an arbitrary reproducing kernel K(x, y), its
associated shift-invariant kernel is dened as
sh
K (x, y) :=
K({x + }, {y + }) d.
[0,1]s
x, y [0, 1]s .
This implies that there exists a shift [0, 1)s such that
D(P + ; H(K)) D(P ; H(K sh )).
In other words, D(P ; H(K sh )) gives an upper bound on the value of D(P + ; H(K))
with a good choice of and measures the average performance of D(P + ; H(K))
605
with respect to the random shifts . Moreover, D(P ; H(K sh )) is shift-invariant: i.e.,
D(P + ; H(K sh )) = D(P ; H(K sh )) for any shift [0, 1)s . There is an added
benet to using the shift-invariant kernel: for a rank-1 lattice point set
kz
(11)
: k = 0, . . . , n 1 ,
P := P (z) :=
n
the computational cost of evaluating the worst-case error D(P (z); H(K sh )) is reduced.
Indeed, from (2) and from the shift-invariant property of K sh (x, y), we have (see [9])
(12)
D2 (P (z); H(K sh )) =
K sh (x, 0) dx +
[0,1]s
n1
1 sh
K (xk , 0).
n
k=0
Although D(P ; H(K)) typically takes an O(n2 ) operation to evaluate for a general
point set (see (2)), D(P (z); H(K sh )) takes only an O(n) operation to evaluate for a
lattice point set.
3.2. Quality measures of lattice rules. In dimension s 3 one normally
nds lattice rules by computer searches based on some criterion of goodness. There
are a number of such criteria available; see [9, 18, 20]. Here we will use a criterion
based on the worst-case error of the RKHS, whose kernel is the shift-invariant kernel
of the kernel given in (4).
For the kernel (4), its associated shift-invariant kernel can easily be found to be
sh
Ks,
(x, y) = 1 +
s,u
B2 ({xj yj }),
=uA
ju
1
where we have used 0 ({x + }, {y + })d = B2 ({x y}). For an arbitrary point
set P in (1), from (2) it follows that
sh
D2 (P ; H(Ks,
)) =
n1
1
n2
s,u
i,k=0 =uA
ju
For the rank-1 lattice point set P (z) in (11), based on (12) we have a simplication
sh
D2 (P (z); H(Ks,
)) =
n1
kzj
1
.
s,u
B2
n
n
ju
k=0 =uA
Note that if the weights s,u are of the product form (5), the computation of the
sh
quantity D2 (P ; H(Ks,
)) is simpler. For example, for the rank-1 lattice point set
P (z) in (11), we have
(13)
sh
(P (z); H(Ks,
))
n1 s
kzj
1
.
1 + s,j B2
= 1 +
n
n
j=1
k=0
606
XIAOQUN WANG
sh
We will use the worst-case error corresponding to the shift-invariant kernel Ks,
(x, y),
sh
i.e., D(P ; H(Ks, )), as a quality measure for lattice rules. It depends on the weights
s,u and is a mixture of the uniformity of various projections. Judged by such a
measure, the goodness of a lattice point set depends not only on the point set itself, but
sh
also on the weights. A lattice point set with small D(P ; H(Ks,
)) does not necessarily
have good low-order projections, say, order-2 projections; see section 4. Our purpose is
to choose appropriate weights s,u , such that good low-order projections are achieved,
or such that the projections of some specic orders are optimized.
Remark 1. Since the Bernoulli polynomial of degree 2 can be expressed as
1
B2 (x) =
2 2
(14)
h=
e2ihx
,
h2
x [0, 1],
where the prime on the sum indicates that the h = 0 term is omitted, the shiftsh
invariant kernel Ks,
(x, y) can be written as
sh
Ks,
(x, y) = 1 +
s,u
=uA
ju h=
e2ih(xj yj )
,
h2
where
s,u =
s,u
.
(2 2 )|u|
This is the reproducing kernel of a weighted Korobov space with the weights s,u and
= 2; see [6]. If the weights s,u are of the product form (5), then the weights s,u
are also of the product form
s,j
s,j with s,j =
s,u =
,
2
2
ju
sh
and the kernel Ks,
(x, y) can be written as
s
sh
Ks, (x, y) =
1 + s,j
j=1
h=
e2ih(xj yj )
h2
ju
sh
D2 (P ; H(K()
)) =
n1
1
n2
s,u
ju
For the rank-1 lattice point set P (z) in (11), from (12) we have a simplication
n1
kzj
1
2
sh
D (P (z); K() ) =
.
s,u
B2
n
n
ju
k=0 uA, |u|=
sh
Based on (10), it is obvious that D(P ; K()
) gives an upper bound on the order-
discrepancy of the shifted point set P + by a good choice of .
607
3.3. The elementary order- discrepancy. In the rest of this paper we are
interested in the order-dependent weights (see [6]), that is,
(16)
s, = 1, s,u := |u| ,
for some nonnegative numbers 1 , . . . , s (they may depend on the dimension s). The
order-dependent weights depend on u only through its order. A benet of using ordersh
)) as
dependent weights is that there exists a fast algorithm to compute D(P ; H(K()
shown below. The order-dependent weights are also of the product form if and only
if = for some constant .
For an arbitrary point set P in (1), its elementary order- discrepancy, denoted
by G() (P ), is dened as
G2() (P ) :=
n1
1
n2
sh
The value of G() (P ) is equal to D(P ; H(K()
)) in (15) with s,u = 1 for all u A
with |u| = . If P is a lattice point set (11), then
G2() (P (z))
n1
1
=
n
B2
kzj
n
.
If P is a set of independent samples from the uniform distribution over [0, 1]s , then
s
1
(17)
.
E[G2() (P )] =
6n
sh
Obviously, for order-dependent weights (16), D(P ; H(K()
)) given in (15) can be
expressed in terms of the elementary order- discrepancies:
sh
)) = G2() (P ),
D2 (P ; H(K()
sh
)) can then be expressed as the weighted sum
and the worst-case error D(P ; H(Ks,
of the elementary order- discrepancies:
(18)
sh
D2 (P ; H(Ks,
)) =
s
=1
sh
D2 (P ; K()
)=
s
G2() (P ).
=1
The elementary order- discrepancy G() (P ) measures the quality of all order-
projections taken together. Its advantage is that it does not depend on any weight
(it depends only on the point set itself). We will use it to compare dierent (lattice)
point sets. We will show that such a measure is more informative than the worst-case
error.
The formulas for the elementary order- discrepancies involve quantities of the
form
Cj for some C1 , C2 , . . . , Cs .
uA, |u|= ju
Such quantities can be eciently computed using a recursive formula as shown in [6].
This recursive method will be used in our algorithms below.
608
XIAOQUN WANG
sh
)) =
D2 (P (z(as )); H(Ks,
n1
=1 k=0
B2
uA, |u|= ju
kaj1
s
n
.
(19)
sh
(P (z1 , . . . , zs ); H(Ks,
))
n1
=1 k=0
uA, |u|= ju
B2
kzj
n
is minimized.
The CBC lattice rules have the advantage of being extensible in dimension s if
the weights are independent of s, and they achieve the (strong) tractability error
bound under appropriate conditions on the weights [6]. A faster CBC algorithm was
developed in [19]. The error bounds of Korobov rules in weighted spaces were studied
in [30] for product weights (the same analysis applies to nonproduct weights). Faster
Korobov algorithms are also desirable. We are mainly interested in the potential
possibility of these constructions in achieving good elementary order- discrepancies
by suitably choosing the weights 1 , 2 , . . . .
If the order-dependent weights are also of the product form, i.e., = for some
sh
)) is to use the formula (13).
, then a more ecient way to compute D2 (P (z); H(Ks,
In this case, the quality of low-order projections of the resulting lattice points can be
sensitive to the parameter , and thus this parameter should be chosen carefully to
reect the relative importance of ANOVA terms of dierent orders (for this reason,
product weights are not the focus of this paper).
4. Good lattice points in high dimensions: How well are their projections distributed?. We take a completely dierent strategy from [6] to investigate
the quality of good lattice points; namely, we focus on the projections of dierent
orders. This allows us to assess the quality of the projections of some selected orders
609
and to examine the possible advantages of lattice points over digital nets and random
points. Note that an arbitrary lattice point set (11) with prime n has the same order1 projections, namely, {k/n : k = 0, 1, . . . , n 1}, which are automatically perfectly
distributed. Thus we need only focus on order-2 and higher-order projections.
sh
)) is
The formula (18) indicates that the square worst-case error D2 (P ; H(Ks,
the weighted sum of the square elementary order- discrepancies. The weights allow
greater or smaller emphasis on the projections of dierent orders and aect the goodness of projections of various orders. The choice of the weights should reect the
characteristic of the problems at hand. We consider two choices of order-dependent
weights:
Choice (A): 1 = 2 = 1, and = 0 for 3.
Choice (B): 1 = 2 = 3 = 1, and = 0 for 4.
The corresponding lattice rule (found by Algorithm 1 or 2) is known as (Korobov or
CBC) lattice rule (A) or (B). The choice (A) puts all emphasis on order-2 projections
and thus may result in lattice rules with optimal order-2 projections, while the choice
(B) puts emphasis on both order-2 and order-3 projections. The relative size of 3
with respect to 2 is important. The choice in (B) is intended for use with nance
problems in section 5 and is based on the matching strategy in [28].
We compare the Korobov or CBC lattice rule (A) or (B) with the Sobol points
[25] by comparing their elementary order- discrepancies. The root mean-squared
elementary order- discrepancy (see (17)), is included as a benchmark.
We are also interested in the following problem: given a method of construction,
say, the Korobov or CBC construction, with the freedom of choosing order-dependent
weights, what is the minimum value of the elementary order- discrepancy which can
be achieved by this construction for a xed ? This value and the mean value for a
random point set can be used as benchmarks. They indicate how good (or bad) the
order- projections in the optimal (or average) case can be, implying the possibility
or impossibility of good lattice points.
It turns out that for a xed with 1 < s, the optimal elementary order-
discrepancy, which can be achieved by the Korobov or CBC method, can be easily
found: one just chooses the weights to be
(20)
= 1,
but
j = 0
for j = .
This choice of weights in the Korobov or CBC algorithm results in lattice points
achieving the minimum elementary order- discrepancy among all possible orderdependent weights using the Korobov or CBC algorithm, respectively. (An interesting problem is the following: what is the minimum value of the elementary order-
discrepancy without restricting a method of construction?)
Table 1 presents the comparisons of the elementary order- discrepancies for s =
64. The optimal values of the elementary order- discrepancies (obtained as indicated
above) for the Korobov or CBC construction are also included. The comparisons
clearly indicate the potential possibility and limitations of good lattice points. We
observe the following:
For = 1, lattice points (A) and (B) (and Sobol points) have elementary
order-1 discrepancies and convergence order much better than those of random points.
For = 2, lattice points (A) and (B) have elementary order-2 discrepancies
much better than those of the Sobol points and random points. The convergence order of the elementary order-2 discrepancies of lattice points (A)
610
XIAOQUN WANG
Table 1
Shown are the elementary order- discrepancies and the convergence orders in dimension s =
64. For the Sobol points, n = 28 , 210 , and 212 . The convergence order (i.e., the value r in an
expression of the form O(nr )) is estimated from linear regression on the empirical data. Mean
in the third column is the root mean square elementary order- discrepancy; see (17). The optimal
values of the elementary order- discrepancies are obtained (for the Korobov and CBC constructions)
using the weights (20) for each xed .
G()
G(1)
G(2)
G(3)
G(4)
n
251
1009
4001
r
251
1009
4001
r
251
1009
4001
r
251
1009
4001
r
Mean
2.06e-1
1.03e-1
5.16e-2
0.50
4.72e-1
2.36e-1
1.18e-1
0.50
8.77e-1
4.37e-1
2.20e-1
0.50
1.40e00
6.97e-1
3.50e-1
0.50
Sobol
2.13e-2
5.66e-3
1.34e-3
1.00
5.53e-1
2.95e-1
1.21e-1
0.50
8.13e-1
4.10e-1
2.13e-1
0.48
1.43e00
7.13e-1
3.59e-1
0.50
Lattice
Korobov
1.30e-2
3.24e-3
8.16e-4
1.00
3.09e-1
9.56e-2
2.48e-2
0.91
8.84e-1
4.19e-1
2.02e-1
0.53
1.40e00
6.99e-1
3.50e-1
0.50
(A)
CBC
1.30e-2
3.24e-3
8.16e-4
1.00
3.21e-1
9.31e-2
2.49e-2
0.92
8.82e-1
4.41e-1
2.29e-1
0.49
1.40e00
6.97e-1
3.50e-1
0.50
Lattice
Korobov
1.30e-2
3.24e-3
8.16e-4
1.00
3.11e-1
9.56e-2
3.39e-2
0.80
8.84e-1
4.19e-1
1.70e-1
0.59
1.40e00
6.99e-1
3.61e-1
0.49
(B)
CBC
1.30e-2
3.24e-3
8.16e-4
1.00
3.55e-1
1.27e-1
4.66e-2
0.73
8.39e-1
3.91e-1
1.67e-1
0.58
1.42e00
7.15e-1
3.58e-1
0.50
Optimal
Korobov
1.30e-2
3.24e-3
8.16e-4
1.00
3.09e-1
9.56e-2
2.48e-2
0.91
7.71e-1
3.87e-1
1.50e-1
0.59
1.40e00
6.92e-1
3.41e-1
0.51
values
CBC
1.30e-2
3.24e-3
8.16e-4
1.00
3.21e-1
9.31e-2
2.49e-2
0.92
6.77e-1
3.12e-1
1.32e-1
0.59
1.40e00
6.87e-1
3.38e-1
0.51
and (B) is much faster than that of the Sobol points (the Sobol points lost
the advantage over random points in order-2 projections on average). Lattice
points (A) give the best elementary order-2 discrepancies.
For = 3, the elementary order-3 discrepancies of lattice points (A) and (B)
are similar to those of the Sobol points and random points and are close to
the optimal values. Lattice points (B) have elementary order-3 discrepancies
better than those of lattice points (A).
For 4, lattice points (A) and (B) have order- projections no worse than
those of random points (this is a surprise). For 4, even the optimal
elementary order- discrepancy, which can be achieved by the Korobov or
CBC construction, is only slightly better than the mean values for random
points (in Table 1 we do not give the numerical results for > 4). It seems
that for 4, if n is small (say, in 1000s) and s is large (say, s > 50), one
should not expect to achieve much better order- projections than those of
random points. There exists a limit for given n and s. This indicates the
diculty in achieving good higher-order projections.
In all cases, the Korobov and CBC lattice points have similar behavior.
Remark 2. A lattice point set which achieves the optimal elementary order-
discrepancy with > 2 (see (20) for the construction of such a point set) may have very
bad elementary order-2 discrepancy. For example, the lattice points which achieve the
optimal elementary order-4 discrepancy have very bad elementary order-2 discrepancy.
Thus the Korobov or CBC lattice points which achieve the optimal elementary order-
discrepancy for 4 can be useful only when the integrand is dominated by order-
ANOVA terms (and the possible advantage is very limited). Our interest in them is
just due to the motivation to know what is the best possible order- discrepancy for
a given and n. On the other hand, the lattice points found with the weights (A)
and (B) have not only very good (order-1 and) order-2 projections but also higherorder projections no worse than those of random points. Note that an alternative
611
Table 2
Shown are the elementary order- discrepancies for the classical (Korobov or CBC) lattice rules
constructed based on the quality measure P2 and the mean values for random points in dimension
s = 64.
n
251
1009
4001
Mean
Korobov
CBC
Mean
Korobov
CBC
Mean
Korobov
CBC
G(1)
2.06e-1
1.30e-2
1.30e-2
1.03e-1
3.24e-3
3.24e-3
5.16e-2
8.16e-4
8.16e-4
G(2)
4.72e-1
8.22e-1
2.55e00
2.36e-1
3.66e-1
3.05e00
1.18e-1
3.48e-1
2.95e00
G(3)
8.77e-1
8.70e-1
2.70e00
4.37e-1
4.14e-1
2.90e00
2.20e-1
2.60e-1
2.83e00
G(4)
1.40e00
1.40e00
4.68e00
6.97e-1
6.98e-1
5.41e00
3.50e-1
3.61e-1
5.08e00
G(8)
3.24e00
3.24e00
6.88e00
1.62e00
1.62e00
7.29e00
8.12e-1
8.12e-1
6.29e00
G(16)
8.30e-1
8.30e-1
1.16e00
4.14e-1
4.14e-1
1.02e00
2.08e-1
2.08e-1
6.92e-1
G(32)
3.03e-5
3.03e-5
3.16e-5
1.51e-5
1.51e-5
1.85e-5
7.59e-6
7.59e-6
8.70e-6
G(64)
7.93e-27
7.93e-27
7.91e-27
3.96e-27
3.96e-27
4.02e-27
1.99e-27
1.99e-27
2.00e-27
choice of weights = with a small parameter may also produce lattice points
with good order-2 projections. The corresponding algorithm is more ecient but
less robust (since the parameter has strong inuence on the quality as mentioned
above).
The attempt to improve the elementary order- discrepancies for 4 does
not turn out to be rewarding for small n in high dimensions. To achieve better
higher-order projections, one has to enlarge n (or nd better construction beyond
Korobov or CBC). In practice, we recommend using lattice rules (A). Due to their
good projection property, it is reasonable to expect that they have good performance
for favorable functions (dominated by order-1 and order-2 terms), while doing no
harm for unfavorable ones (say, functions dominated by high-order terms). In this
sense they are robust, perhaps opening the way to wider use.
We turn to the classical lattice rules. By classical lattice rules we mean the
rules found by using the classical quality measure P in searching algorithms. For
simplicity of presentation, we assume = 2; then P2 has the following form (see [20]):
P2 =
hz0
h1 hs
2,
j = (2 2 )j ,
j = 1, . . . , s.
612
XIAOQUN WANG
have
sh
(P (z); H(Ks,
))
n1 s
kzj
1
2
1 + 2 B2
= 1 +
n
n
k=0 j=1
n1 s
1
exp(2ihkz
/n)
j
= 1 +
1+
n
h2
j=1
k=0
= 1 +
1
n
n1
h=
s
exp(2ihj kzj /n)
hj
(exp(2ih z/n))k
1 n1
= 1 +
s
2
n
hj
s
hZ
= 1 +
=
hz0
k=0
j=1
1
2
h1 hs
1
2
2 = P2 ,
h1 hs
hz0
613
5.1. Arithmetic Asian options. Consider an Asian call option based on the
arithmetic average of the underlying asset. The terminal payo is
s
1
gA (St1 , . . . , Sts ) = max 0,
St K ,
s j=1 j
where K is the strike price at the expiration date T and Stj are the prices of the
underlying asset at equally spaced times tj = jt, j = 1, . . . , s, t = T /s. We
assume that under the risk-neutral measure the asset follows geometric Brownian
motion (BlackScholes model):
(22)
where r is the risk-free interest rate, is the volatility, and Bt is the standard Brownian
motion. The value of the option at t = 0 is
C0 = EQ [erT gA (St1 , . . . , Sts )],
where EQ [] is the expectation under the risk-neutral measure Q. The simulation
method is based on the analytical solution to (22)
St = S0 exp((r 2 /2)t + Bt )
and on a generation of the Brownian motion
(Bt1 , . . . , Bts )T = A (z1 , . . . , zs )T ,
where zj N (0, 1), j = 1, . . . , s, are independent and identically distributed standard normal variables and A is an s s matrix A satisfying AAT = V with V =
(min(ti , tj ))si,j=1 . The standard construction (i.e., sequential sampling) takes A to be
the Cholesky decomposition of V .
5.1.1. The eciency improvement techniques. To improve QMC, we use
dimension reduction techniques, such as the Brownian bridge (BB) (see [3, 17]), and
the principal component analysis (PCA) (see [1]), and variance reduction techniques,
such as antithetic variates, control variates, and their combination. We take the payo
of the geometric average Asian call option (which is analytically tractable) as a control
s
1/s
variable: gG (St1 , . . . , Sts ) = max(0, j=1 Stj K). Note that
EQ [gA ()] = EQ [gA () b gG ()] + b EQ [gG ()],
where b is some multiplier b. The optimal b is obtained by minimizing the variance
of gA bgG . The same multiplier b will be used in QMC. Importance sampling and
weighted importance sampling in QMC can also be eectively used by choosing a
suitable importance density. This problem will be studied in the future.
5.1.2. The algorithms and the error estimation. We compare the eciency
of the following QMC algorithms with that of MC:
the Korobov and CBC lattice rules (A) and (B);
the classical Korobov or CBC lattice rules constructed based on P2 (see [22]).
the QMC algorithm based on the Sobol points.
614
XIAOQUN WANG
In order to get an estimation of the accuracy for each method, we compute the
sample variance and the standard deviations. For the Sobol points, we use a digitalscrambling method as used in [28]. For lattice point set P := {xk }, we use a random
shift method; see [21]. Dene
Qn,s (f ; ) :=
n1
1
f ({xk + })
n
k=0
and
1
Qn,s (f ; j ),
m j=1
m
(23)
Qn,s (f ) :=
where 1 , , m are independent and identically distributed random vectors, uniformly distributed over [0, 1]s . The sample variance of the estimate (23) is estimated
by
n,s (f )] =
Var[Q
1
[Qn,s (f ; j ) Qn,s (f )]2 .
m(m 1) j=1
m
The standard deviation is just the square root of the sample variance. The relative
eciency ratio or the variance reduction factor of the estimate (23) with respect to
the crude MC estimate (i.e., without variance reduction and dimension reduction) is
dened by
crudeMC (f ))/Var[Q
n,s (f )],
E [Qn,s ] := Var(Q
crudeMC (f )) is the sample variance of the crude MC estimate based on
where Var(Q
mn samples. The numerical results are presented in Tables 3 and 4.
5.1.3. The numerical results. Tables 3 and 4 present the comparisons for
pricing arithmetic Asian options. The experiments show the following:
The classical Korobov or CBC lattice rules constructed based on P2 behave
very badly. Their performance can be worse than that of MC. The reason is
that they have very bad elementary order-2 discrepancies as shown in section
4. Thus such rules are unsuitable for functions for which order-2 ANOVA
terms are important.
The eciency of lattice rules (A) and (B) is dramatically improved. They are
more ecient than the Sobol points when standard construction is used to
generate the Brownian motion, and are at the same level of accuracy with the
Sobol points if dimension reduction techniques are used (since the possible
bad order-2 and higher-order projections of Sobol points in latter dimensions
now have less inuence due to the reduced eective dimensions).
The relative eciency of QMC algorithms is aected by the strike price K:
the smaller the K is, the more ecient the QMC algorithms are. This is
consistent with the fact that a smaller K leads to functions of lower superposition dimension (small K leads to integrands dominated mainly by order-1
ANOVA terms); see [27].
Dimension reduction and variance reduction techniques provide further signicant improvement for QMC (dimension reduction techniques are useless in
MC). In most cases, lattice rules (A) and (B) combined with these techniques
are more ecient than the Sobol points combined with the same techniques.
615
Table 3
Shown are the standard deviations and relative eciency (in parentheses) with respect to crude
MC for the arithmetic Asian option with m = 100 replications for s = 64 (n = 4001 for MC and
good lattice points, n = 4096 for the Sobol points). The parameters are S0 = 100, = 0.2, T =
1.0, r = 0.1. In the second column are the path generation methods: STDstandard construction
(sequential sampling), BBBrownian bridge, PCAprincipal component analysis.
Strike
price
90
STD
BB
PCA
100
STD
BB
PCA
110
STD
BB
PCA
MC
1.68e-2
(1.00)
1.68e-2
(0.998)
1.68e-2
(0.997)
1.37e-2
(1.00)
1.37e-2
(0.999)
1.37e-2
(0.998)
9.07e-3
(1.00)
9.08e-3
(0.999)
9.08e-3
(0.998)
Sobol
4.31e-3
(15)
7.10e-4
(561)
5.82e-4
(835)
6.43e-3
(4.5)
7.88e-4
(302)
5.28e-4
(674)
6.48e-3
(2)
6.32e-4
(206)
4.45e-4
(416)
Classical
Korobov
CBC
1.21e-2
6.87e-2
(1.92)
(0.06)
2.11e-2
6.22e-3
(0.63)
(7.32)
7.16e-3
8.73e-4
(5.53)
(372)
1.91e-2
1.05e-1
(0.51)
(0.02)
2.81e-2
7.39e-3
(0.24)
(3.43)
7.09e-3
8.40e-4
(3.73)
(266)
1.85e-2
1.00e-1
(0.24)
(0.01)
2.87e-2
7.46e-3
(0.10)
(1.48)
8.52e-3
9.38e-4
(1.13)
(94)
Lattice
Korobov
1.92e-3
(77)
1.10e-3
(235)
5.96e-4
(796)
2.91e-3
(22)
1.28e-3
(115)
5.30e-4
(669)
3.01e-3
(9)
1.29e-3
(50)
4.73e-4
(368)
(A)
CBC
1.65e-3
(104)
7.62e-4
(487)
5.53e-4
(927)
2.46e-3
(31)
7.67e-4
(319)
5.06e-4
(734)
3.15e-3
(8)
8.42e-4
(116)
4.40e-4
(425)
Lattice
Korobov
2.14e-3
(62)
1.55e-3
(118)
5.67e-4
(881)
2.83e-3
(23)
2.17e-3
(40)
5.50e-4
(619)
3.16e-3
(8)
2.19e-3
(17)
4.94e-4
(337)
(B)
CBC
2.22e-3
(57)
7.51e-4
(502)
5.64e-4
(888)
3.33e-3
(17)
7.84e-4
(305)
5.03e-4
(741)
3.19e-3
(8)
7.30e-4
(154)
4.36e-4
(434)
Table 4
The same as Table 3, but with variance reduction techniques (Crudecrude estimate without
variance reduction, AVantithetic variates, CVcontrol variates, ACVcombination of AV and
CV). The strike price is xed to be K = 100. We give only the variance reduction factors (the
standard deviation of the crude MC estimate is 1.37e-2).
Path
STD
BB
PCA
Crude
AV
CV
ACV
Crude
AV
CV
ACV
Crude
AV
CV
ACV
MC
1.0
5.6
1145
2113
1.0
5.6
1151
2143
1.0
5.6
1156
2142
Sobol
4.5
4.8
1331
1973
302
756
12849
16792
674
35868
132610
655760
Classical
Korobov
CBC
0.5
0.017
0.5
0.017
110
5.3
113
6.6
0.2
3.4
0.2
3.4
61
177
69
229
3.7
266
4.2
463
29
6244
90
13359
Lattice
Korobov
22
26
4914
11716
115
192
13693
23535
669
8351
31161
38484
(A)
CBC
31
33
6015
13098
319
690
19282
36038
734
13851
48900
89657
Lattice
Korobov
23
31
4895
10421
40
252
12037
22590
619
8875
30316
68602
(B)
CBC
17
21
3910
6361
305
697
20584
38287
741
21090
69510
89226
5.2. American options. American options are derivative securities for which
the holder of the security can choose the time of exercise. The American options
additional complicationnding the optimal time for exercisemakes pricing it one
of the most challenging problems in nance. Recently, several simulation methods
are proposed for the valuation of American options (see, among others, [2, 15]). Our
experiments are based on the least-square Monte Carlo (LSM) method [15].
Consider an American option which can be exercised only at discrete times tj =
jt, j = 1, . . . , s, t = T /s (such options are sometimes called Bermudan options).
The payo function is assumed to be g(t, St ), where St is the asset price. The riskneutral dynamics for the asset price are the same as in (22). There are several major
steps in the LSM method (see [15] for a more detailed description):
Generate N realization paths {Sti , t = 0, t1 , . . . , ts ; i = 1, . . . , N }.
616
XIAOQUN WANG
Determine for each path i the optimal exercise time i . This is done by
comparing (i) the payo from immediate exercise and (ii) the expected payo
from keeping the option alive (i.e., the continuation value). The continuation
value is estimated by a least-square regression using the cross-sectional information provided by simulation. This procedure is repeated recursively going
back in time from T .
Estimate the options price by
p =
N
1 ri
e
g(i , Si ),
N i=1
617
Table 5
Shown are the standard deviations and relative eciency (in parentheses) with respect to crude
MC for American put option with m = 100 replications (n = 4001 for MC and good lattice point,
and n = 4096 for the Sobol points). The parameters are K = 100, = 0.2, T = 1.0, r = 0.1. The
exercise periods are s = 64. In the second column are the path generation methods (STDstandard
construction).
Initial
stock price
90
STD
BB
PCA
100
STD
BB
PCA
110
STD
BB
PCA
MC
8.78e-3
(1.00)
8.26e-3
(1.1)
8.56e-3
(1.1)
9.09e-3
(1.00)
9.89e-3
(0.8)
9.81e-3
(0.9)
6.54e-3
(1.00)
7.51e-3
(0.8)
7.07e-3
(0.9)
Sobol
6.69e-3
(1.7)
4.02e-3
(4.8)
4.16e-3
(4.5)
6.14e-3
(2.2)
5.58e-3
(2.7)
2.97e-3
(9.40)
4.59e-3
(2.0)
3.10e-3
(4.4)
1.97e-3
(11.1)
Lattice
Korobov
4.55e-3
(3.7)
4.83e-3
(3.3)
3.66e-3
(5.8)
4.47e-3
(4.1)
3.19e-3
(8.1)
3.08e-3
(8.7)
3.93e-3
(2.8)
2.46e-3
(7.1)
2.03e-4
(10.4)
(A)
CBC
4.58e-3
(3.7)
4.42e-2
(3.9)
3.73e-3
(5.5)
4.79e-3
(3.6)
3.51e-3
(6.7)
3.10e-3
(8.6)
3.38e-3
(3.7)
2.33e-3
(7.9)
2.29e-3
(8.2)
Lattice
Korobov
4.45e-3
(3.9)
4.63e-3
(3.6)
4.41e-3
(4.0)
5.41e-2
(2.8)
7.05e-3
(1.7)
2.83e-3
(10.3)
3.31e-3
(3.9)
6.09e-3
(1.2)
2.07e-3
(10.0)
(B)
CBC
4.29e-3
(4.2)
3.16e-3
(7.7)
3.50e-3
(6.3)
5.40e-3
(2.8)
2.96e-3
(9.5)
2.89e-3
(9.9)
4.24e-3
(2.4)
2.16e-3
(9.1)
1.92e-2
(11.6)
Table 6
Similar to Table 5, but for American-Bermudan-Asian call option. The parameters are K =
100, = 0.2, T = 1.0, r = 0.1. The time is discretized into 64 steps. The standard deviations are
given only for crude MC estimates (in parentheses).
(A0 , S0 )
(90,90)
(90,110)
(100,90)
(100,110)
(110,90)
(110,110)
STD
BB
PCA
STD
BB
PCA
STD
BB
PCA
STD
BB
PCA
STD
BB
PCA
STD
BB
PCA
MC
(5.31e-3) 1.0
0.8
0.7
(1.43e-2) 1.0
0.8
0.8
(6.23e-3) 1.0
0.8
0.7
(1.43e-2) 1.0
0.8
0.8
(7.01e-3) 1.0
0.8
0.9
(1.34e-2) 1.0
0.9
0.9
Sobol
1.6
178.1
275.0
9.7
238.1
382.4
1.8
104.2
221.6
11.6
153.9
181.8
2.1
12.9
50.6
9.0
67.3
146.4
Lattice
Korobov
7.7
33.4
207.4
52.9
158.7
375.9
8.7
42.4
123.4
41.6
108.7
117.7
9.9
22.2
36.4
25.5
43.6
57.0
(A)
CBC
6.3
69.0
318.7
56.3
96.8
356.6
8.7
81.8
177.4
39.1
96.4
213.3
9.8
28.9
47.9
23.7
66.2
80.5
Lattice
Korobov
5.7
6.2
220.6
41.7
103.6
360.4
7.6
12.6
157.7
31.9
104.4
159.9
7.4
21.9
25.2
25.7
83.0
105.9
(B)
CBC
6.1
89.7
288.2
32.5
271.4
334.3
7.5
117.6
249.1
31.2
178.3
199.4
7.3
49.4
40.9
17.2
129.0
93.0
are dierent from the ones constructed in [28], which are problem-dependent (they
have similar eciency). Also, the diculty of choosing suitable weights is partially
avoided (since we could use the same weights regardless of the problems at hand, the
dimension, and the methods of path generation). We thus recommend using lattice
points (A) in practice and present in the appendix (see Tables 7 and 8) some generators
(Korobov and CBC) for n = 1009 and n = 4001 up to dimensions s = 128 (faster
searching for generators for larger n and s is an important future work). Note that
once the generators are found (or given), lattice points are much easier to generate
than random points and other quasi-random numbers.
618
XIAOQUN WANG
619
as
1
390
382
382
309
160
156
156
156
156
156
156
156
304
156
304
156
304
304
249
249
249
249
249
249
466
177
177
177
57
177
177
57
177
65
65
65
326
380
380
77
77
77
zs
1
390
295
221
110
187
350
138
316
213
323
281
131
54
225
468
420
143
264
291
89
494
107
320
353
435
180
472
397
141
57
16
106
313
497
400
244
329
385
172
490
28
67
s
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
as
380
77
380
380
380
380
77
77
380
77
380
380
425
380
380
342
342
342
342
342
342
342
342
342
342
262
181
181
181
181
262
262
181
181
181
262
342
342
342
342
95
95
308
zs
388
37
12
305
152
26
240
492
31
193
206
242
273
500
34
284
177
421
257
275
126
436
198
266
154
10
298
113
135
487
317
371
192
293
181
101
235
150
280
337
151
72
315
s
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
as
95
95
95
95
95
95
95
430
237
430
430
430
237
237
430
237
237
430
430
430
430
237
430
430
430
430
430
237
430
233
498
498
498
233
233
498
233
233
233
233
233
233
zs
499
465
165
81
479
23
402
159
95
440
389
39
297
355
189
322
119
219
406
83
223
263
183
148
186
147
85
7
271
248
44
116
188
381
432
241
409
38
168
199
123
338
620
XIAOQUN WANG
Table 8
The same as Table 5, but for n = 4001.
s
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
as
1
1478
1527
823
1134
1134
1563
547
547
1419
547
547
933
933
1175
933
933
1175
1175
1175
1175
933
1175
774
336
336
774
336
336
774
933
933
933
1175
933
933
933
933
933
933
1375
1452
1375
zs
1
1478
1446
1031
555
1180
390
1902
689
1471
974
1754
1725
957
1746
1186
372
1477
517
1093
1451
775
981
1543
855
476
1572
534
453
1006
1747
1492
1206
809
639
64
724
1960
214
936
688
895
1541
s
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
as
1452
1375
1375
1397
358
358
358
1397
358
358
1397
358
1247
1247
1585
1585
1639
1306
1639
1639
1639
1639
1306
1306
1306
367
556
556
367
556
367
367
556
556
1495
1495
372
372
372
372
1495
272
272
zs
1254
122
808
1481
1870
703
1420
874
1287
51
1106
1931
450
734
494
666
1759
760
745
1875
1558
898
458
1364
146
1320
997
1367
376
1124
1123
189
1744
166
665
1658
863
1066
299
1462
1205
1453
1645
s
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
as
1495
372
372
372
372
372
372
372
372
372
372
1495
1495
372
372
1495
372
372
372
905
905
905
84
905
84
84
84
905
84
84
372
372
372
372
1495
1495
372
372
387
387
387
387
zs
1418
112
1850
348
108
159
645
1362
1854
1548
1442
1441
1228
1901
1919
1516
240
595
1324
536
970
1764
1715
621
1078
208
335
1019
1239
313
184
1270
66
303
918
1233
1778
1500
1997
1131
1221
1985
Acknowledgment. The author would like to thank the referees for their valuable
comments.
REFERENCES
[1] P. Acworth, M. Broadie, and P. Glasserman, A comparison of some Monte Carlo and
quasi-Monte Carlo techniques for option pricing, in Monte Carlo and Quasi-Monte Carlo
Methods 1996, H. Niederreiter, P. Hellekalek, G. Larcher, and P. Zinterhof, eds., SpringerVerlag, New York, 1998, pp. 118.
[2] M. Broadie and P. Glasserman, Pricing American-style security using simulation, J.
Econom. Dynam. Control, 21 (1997), pp. 13231352.
621
Reproduced with permission of the copyright owner. Further reproduction prohibited without permission.