American Journal of Computational Mathematics, 2011, 1, 294-302
doi:10.4236/ajcm.2011.14036 Published Online December 2011 (http://www.SciRP.org/journal/ajcm)
Copyright © 2011 SciRes. AJCM
First Order Convergence Analysis for Sparse Grid Method
in Stochastic Two-Stage Linear Optimization Problem
Shengyuan Chen
Department of Mathematics and Statistics, York University, Toronto, Canada
E-mail: chensy@mathsta t.yorku.ca
Received August 27, 2011; revised September 20, 2011; accepted Septem b e r 30, 2011
Abstract
Stochastic two-stage linear optimization is an important and widely used optimization model. Efficiency of
numerical integration of the second stage value function is critical. However, the second stage value function
is piecewise linear convex, which imposes challenges for applying the modern efficient spare grid method. In
this paper, we prove the first order convergence rate of the sparse grid method for this important stochastic
optimization model, utilizing convexity analysis and measure theory. The result is two-folded: it establishes
a theoretical foundation for applying the sparse grid method in stochastic programming, and extends the
convergence theory of sparse grid integration method to piecewise linear and convex functions.
Keywords: Convergence Analysis, Stochastic Optimization, Scenario Generation, Convex Analysis, Measure
Theory
1. Introduction
Stochastic two-stage linear optimization, also called sto-
chastic two-stage linear programming, models a sequen-
tial decision structure, where the first stage decisions are
made now before the random variable manifests itself;
and the second stage decisions are made adaptively to the
realized random variable and the first stage decisions.
The adaptive decision model has been applied many im-
portant application areas. For example, in the introduc-
tory farmer’s problem [1], a farmer needs to divide the
land for different vegetables in spring. The farmer’s ob-
jective is to maximize profit in the harvest season. The
profit is related to the market price at that time and the
weather dependent yield. Neither the price nor the
weather is known at the present time, hence the farmer’s
decision in spring has to take into account multiple sce-
narios. It is not a simple forecasting problem though,
since the farmer’s second stage decision in fall, which
adapts to different scenarios, also jointly determines the
profit. The second stage decision problem is also called
“recourse” problem. [2] collects more recent applications
in engineering, manufacture, finance, transportation, tele-
communication et al.
A stochastic two-stage linear problem with recourse
has the following general representation:


,d,and
min
,=min
..=,0,
T
xX
T
cxx P
xsy
st QyTxy
 


(1)
where
is a random vector properly defined on
P,,,
X
is a polytope feasible region for the first
stage, mn
Q
, s and are a vector and a matrix of
proper sizes, T

:, ,X
 is a real valued
function.
The high dimensional in tegration in (1) is difficult and
is usually approximated by using a set of scenarios and
weights
,,1,,
kk
wk K
K
as:


1
min, ,and
,min
.., 0.
Tk
k
xX k
kT
k
cxw x
xdy
st QyTxy


 
(2)
Under this scenario approximation, the optimal objec-
tive value *
K
z
of (2) provides an approximation of the
optimal objective value of (1). An optimal solution
*
z
*
K
x
of (2) prov id es an appr ox imation o f an op t imal so lu -
tion *
x
of (1).
Monte Carlo (MC) method has been widely used in
this approximation, where are random
,=1, ,
kk
K
S. Y. CHEN295
sampling points and =1
k
wK. The convergence theory
of Monte Carlo method has been extensively studied
[3-6]. The core result is the epi-convergent theorem: un-
der mild assumptions, *
K
z
converges to w.p.1 as
; and any clustering po int of the =1 , which
is the sequence of optimal solutions to (2), is in the op-
timal solution set of the original problem. Quasi Monte-
Carlo (QMC) method has also been recently studied [7],
and similar convergence result has been achieved.
*
z
*
{}
KK
x
K
The sparse grid (SP) method is an established high
dimensional quadrature rule, which was originally pro-
posed by Smolyak [8], and has been studied by many
authors in the context of numerical integration [9] (and
references therein). Its application in the stochastic two-
stage linear optimization is only shown in a recent nu-
merical study in [10]. Though [10] shows the superior
numerical performance of sparse grid method, compared
with both MC and QMC, the convergence analysis is
based on an assumption that the recourse function is in a
Sobolev space, which only holds for a very narrow sub-
set of the two-stage linear problems, i.e., separable prob-
lems. The contribution of this paper are 1) establishing
the epi-convergence of the sparse grid method for this
important decision model; 2) prove the first order con-
vergence rate of the method.
We first introduce the spare grid approximation error
for integrand functions in Sobolev spaces.
Let
D denote the partial derivative operator

 
=
jj
f
Df xx
x
Let

1
=,,,
dj
 
 be a multi-index, and
define
=1
=1
==
dj
jd,
j
jj
j
DD x
where 1d

. The Sobolev space with smooth-
ness parameter is defined as
1r

2
:0,1for all ,
d
r
d
f
Df r
 
where r
means component-wisely ,
jr
1, ,j d. Sobolev spaces could also be defin ed using
p
norms, see Evans [11]. The derivatives in the defini-
tion of Sobolev space are weak derivatives. Formally,
Df
is the
-derivative of f if for all ,

00,1 d
vC
i.e., infinitely differentiable function on ,

0,1 dDf
satisfies the following equation:


 


0,1
[0,1]
d
1d
d
d
Dfxvxx
.
f
xDvxx

For example,
=
f
xx defined in
0,1 has the
first order weak derivative function
Df xsign; but
the function is nondifferentiable at 0 in the usual strong
sense. It has been shown that weak derivative is essen-
tially unique, and coincides with the classical strong de-
rivative when it exists. Various properties of strong de-
rivatives carry over to weak derivative as well, for ex-
ample,

DfDDfD
 

Df
for all multi-
index ,, r
 
. For more calculus rules regard-
ing weak derivative, including the extended Leibniz
theorem, see Evans ([11], Section 5.2.3).
The norm of the defined Sobolev space is
22
,
r
dr
fDf
where 2
is the 2-norm of a function, L r
component- wisely, 2
is the Euclidian 2-norm of a
finite vector:

1/2
2
[0,1]
2d
d
ffx
x
1/2
2
21.
k
j
j
hh



For , the sparse grid method achieves the fol-
lowing convergence rate [12]:
r
d
f





0,1 =1
11
,
d
log ,
Kkk
dk
dr
r
rd rd
fwf
Kf
K


(3)
where
K
is the number of function evaluations, ,rd
is a constant independent of
f
, increasing with both d
and , see Brass [13]. Note that implies
rr
d
f
,
i
d
f
ir
. Since the norm r
d
f and ,dr
are non-
decreasing in r, and the term

11
ln dr
r
KK
is
non-increasing in for large , it is none trivial to
tell which space will yield the tightest bound. The prob-
lem is called fat
rK
F
problem in Wonzniakowski [14]. In
this paper, as we shall see, only is relevant for our
discussion. =1r
The convergence result in (3) only holds for the two-
stage stochastic linear programming (1) in the trivial case,
i.e., when the integrand function is
separable. For example,

,:x

11
1212 1122
,
,,, =min
yy
x
xyyy



y
Copyright © 2011 SciRes. AJCM
S. Y. CHEN
296
1111
=
y
yx


222
=2
y
yx


1122
,,, 0yyyy

is equivalent to

12121 122
,,, ,
x
xx

x
where 1
11 1
x
, 1
22 1
x
 and the conver-
gence result in (3) can be applied directly.
However, in general,

,
x
is non-separable piece-
wise function, see Birge and Louveaux [1], and does not
belong to for any . For example,
r
d
r

1212 ,
,,, =
min
yy
x
xy


 y
2
121
=
y
yx


x
,0yy

is equivalent to

121212 1 2
,,,=,
x
xx

x
and does not have the (weak) derivative


1,1 12
,,,
D
xx
discon and non-differentiable
even in the we
 

1,11,0 0,1
=DfDDf if
nd in (3) can not be
applied to two-stage linear problem directly. The major
contribution of this paper is to prove the convergence of
(2) to (1) in the rate specified in (3) with =1r, i.e., the
first order convere rate, even though

1
,d
x
.
On the other hand, extends the co nvergence
theory of sparse grid method to convex multivariate
piecewise linear functions since for such a function
::
i
since istinuous
ak sense, while
Hence the erro
genc
this analys
(0,1)
Df
exists.

1,1
Df r bou
is
f
dx for i
x
B polyh
and

1,,
m
B partitions presented a

=mfx
; each B
, could be
i isedron
res1
B
in
,=
1,, .
y
i
y
ydxi m
The paper is organized as the followings. In Section 2,
we introduce a logarithmic mollifier function and prove
its various properties. The mollifier function is quite fa-
miliar to the optimization community as it is the barrier
function used in the Interior Point Method for linear pro-
gramming. In Section 3, we use the limiting properties of
the mollifier function to prove the uniform convergence
and the first order convergence rate for the objective
function. We also show the converging behaviour of the
optimal solutions *
K
x
in a subsequence. Finally, Section
4 presents our concions.
In the coming sections, we a
lus ssume . For a
m

=0,1
d
ith a inveore general continuous distribution wrse cu-
mulative function

1:0,1
d
F, one can apply trans-
formation

0,1
d= d
fF f


1dF.



The transforma re complexity in the
an
2. Mollifier Function
We make the followitions of the problem
:
tion brings in mo
ng mild assump
alysis without changing our conclusion, hence we as-
sume

0,1 d
 in the following sections and extend
the anugh inverse transformation and trunca-
tion in the Appendix.
alysis thro
(1):
A1
X
is compty relative interior; pact with nonem
,,<xX x

,

, or relative;
A2: completeness, and

,
x
has nonempty rela-
tive interior;
A3:
=rank Qm
;
A4:
is a contius random vnuoector with an in-
ve cuysis using the
In
rtiblemulative distribution function.
Assumption A1 is necessary for our anal
terior Point Method theory. Assumption A2 is for
convenience since otherwise we need to discuss the case
,=x

, which will drag our analysis to a different
ption A3 is implicitly assumed in many
analysis of linear programming, since the rows of Q
could be preprocessed such that the reduced Q has fu
row rank. Assumption A4 facilitates the conversion from
a unit cube
focus. Assum
ll
0,1 d to
through the inverse c.d.f. trans-
formation.
We define a mollifier function ,:

:
min T

,
.. =
x
,
s
y

By
Tx
st Qy
=1
=log
n
i
iy
. In the
(4)
where following, we call

By
,
x
ion, and

,x
the recourse funct
the molli-
fier function. We let

1
11
,,
T
n
gy Byyy

 



222
1
11
diag,,.
n
Hy Byyy


(5)
is in fact a barrier function widely used in the
()By
Interior Point Method for linear programming, and its
proper ties ar e well studied. As 0
, the converge n c e
of
,x
and
**
,,
,
x
x
yu

iin the following
theo
Theore
s stated
rem. m 2.1. For any 1
let

,0,xX
,
**
,,
,
x
x
yu

e mollifierbe the optimal primal andof th dual solutions
1We thank John Birge for pointing out this elegant argument.
Copyright © 2011 SciRes. AJCM
S. Y. CHEN297
function, then

,
0=,
lim x
x


** **
,,
0,,
lim
x
xxx
y
uyu

,
where

**
,
x
x
yu
unctio is an optimal primal and dual pair of the
recourse fn

,
x
.
Proof. Due to thr functioe barrien , the objective
fu ()B
nction of

,x
is strictly convTogether with
the relative ceness assumption, it is clear that

,x
ex.
omplet
has an unique optimal solution. Since the pro-
non-empty relative interior by assumption, the
central path


**
yu
blem has
,,
,
xx

exists for each
x
, and con-
verges to ther of the optimal set, see e analytical cent
property of
Roos et al. [15] Theorem I.30 and its Definition I.20 for
analytic center.
The converg ing
**
,,
,
x
x
yu

tant topic in
(central path)
as
ap-
pe
where . Clearly
0
has been an impor Interior Point
Med has been extensive studied by many authors.
For interested readers, in addition to the reference given
in the proof, we refer the extensive research in Megiddo
[16], an early work Fiacco [17], and a survey of degen-
eracy and IPM by Güler et al. [18]. For readers interested
in the interior point method in general, we refer Nesterov
and Nemirovskii [19], Renegar [20] and Wright [21].
The KKT condition of the optimization problem
thod, an
ared in (4) is

,,, 0,
T
x
sgQu
Fyu Qy Tx






(1)
,:nm m
x
F

,,,
x
F

is infinitely dore, ifferentiable. Furtherm

*T
,
,
,0
x
x
yu
H
yQ
FQ




,
0,
x
F
I





(6)
where
I
is an identity matrix,

,
,
x
yu
F
is invertible
since ()
H
is positive definite. the implicit
functioeorem, **
,,
,Hence by
n th
x
x
yu
are infinitely differentiable
functions of
. So
**
,,,
=T
x
xx
sy By

 
is
infinitely differtiable.
Proposition 2. 1. For a
en ny 1,

,0,xX

,():
x
.C

In the following, we directly derive te (strong) partial
de h
rivatives of

,x
D
for all 1
. Note that there
are 21
d number of
s satisfy1ing
. We also
prov partial derivatives are finite for all e these
0,1 ,
x
X. Finally, we show that their limits are fi nite when
0
<. Hereinafter, a vector <v or a matrix
M
Prop means the inequality holds cnent-wisely.
osition 2.2. For all

,0,1xX
,
ompo

*
,,xx
u

=<.

=<
xx
uFurthe

0,
lim

*
rmore,

grangian functi o. ization Proof. The Laptimon of the
problem in (4) is
,,Lyu
.
T
syByu uTx
 

TT
uQy 
,x
Since
T
**
,,, ,
,,=
xxxx
Lyu
 


,



**
,,
,,,,
**
*,
***
,,,
*,
*,
*,
=,
=
=
=,
xx
xx x
,,
**
,
,,
,
xx
xx
xx
x
T
u
Lu
Qu

xx
x
x
x
x
Ly
LLy
yu
y
usgy
u
Qy Tx
u












e last equality follows the KKT condition.

where th
Clearly
*,<
x
u
. As 0
, **
,
x
x
uu
by Theo-
rem 2.1.
We let

**
,,
diag,diag .
xxx
yy


x
at the Recall th*
x
y
1, refers to the limiting poin t defined
in the Theorem 2.not an arbitrary optimal solution of
,
x
.
Lemma 2.1. Fo y r an
,0,xX

1
,

*


1
2
,,
=,
T
xx
uQQ




*2
2
,,,
=.
TT
xxx
yQQQ


1
Proof.
,,
,, =
xx
Fyu

** 0defines implicit functions
**
,,
,
x
x

,
yu . With

,,
,
x
x
yu
F
F

given in (6),

1
,
,
x
yu
F
xplicitly cocan be ed as mpute


 
11
11111
1
11 1
,
TT TT
TT
I Q QHQQHHQQHQ
QH QQHQH Q








1
H
where is a shorthand notation of

*,
x
Hy
H
. Hence
by the ilicit function theorem

*
mp

,1
(,) ,,
*,
=,
x
y
ux x
x
y
F
F
u






and the conclusion follows straightforward computation
Copyright © 2011 SciRes. AJCM
S. Y. CHEN
298
tion 2.3. For all 1
. Furthermo
essentia a 2.1,
(7)
Hence
and (5).
Proposi

,0,xX
,


1
22T

re,
,,xx
QQ

2

0,0
lim x


Proof. By Propositiolly.
n 2.2 and Lemm


,,xx
QQ



1
22T

2,<,, .
x
x
X
 
 By Theorem
2.
If the optimal set of
1,


1
22
,
0=0 .
lim T
xx
QQ


,
x
ce the a is non-degenerate,
2T
xQ is non-singular , henbove limit is zero. If
al set is degenerate, then the limit 0
Q
the optim
is not
defined. However, in the following, we shat the
degenerated case has zero Lebesgue measure, hence the
limit is zero with probability one. Let’s first consider a
special case of degeneration. Let the first m columns
of Q be an optimal basis B and
12 1
,,,0,
mm n
yyyyy
 be
e the set


*1
= |=,=0,,,>0EyBTxyyy

ow th
degene
0
optimal b
asic feasible solu rated
. Since
set 0 has zero
a
tion. Defin
12 0
Bm
m

12
=|=0,,,>
m
Yyyy y
easure in m
, and B is both
 
for an arbitrarily d
n.
Lebes-
gue minjective and sur-
jective, hence
 

0mE mBY. Clearly the same
argument holdsegenerated optimal
basis B with an arbitrarily chosen degenerated basic
columFurthermore, there are only finitely many ways
to choose basis and degenerated columns. Hence the total
measure of the set
which leads to a degenerated

,
x

is zero. Hence we conclude that the limit is
ntially.
We continue tca
zero esseo lculate higher order partial deriva-
tives. Note that

2,x

is a mm matrix, and
 
3

,2,
=.
xx
ijk kij

  


 

In general, for any and any set of , kd1,,
k
ii

2,
3
x
ii
k


 s a m

im
matrix, and
 
,2,
12 3,
12
=.
kxx
ii iii
kk
ii
 
 



 
Proposition 2.4. For any
,0,xX
,
1
and any
set of indices 1,,
k
ii,

,
1
<,
k

,
01
=0
lim k
x
ii
k



essentially.
Proof. We first prove the conclusion fr, and
extend by induct i on. o =2k

1
22
,,
()= T
xx
kk
QQ

 


 


11
22
2
,,,
*
1, ,
*
2, ,
1
2
,
*,,
1
2
,
=
00
00
=2
00
0
,
TT
T
xxx
k
x
k
x
T
xk
nx
k
TT
x
QQ Q QQQ
y
y
QQ Q
y
QQ Q


























(8)
where


*1
,, *22
,,,
==
jx TT
xxx
jk
j
k
k
yyQQQ







is shown in lemma 2.1. Clearly (8) is finite for >0
.
Now taking limit 0
, by Theorem 2.1 and
tion 2.3: Proposi-

2,
0= 0essentially.
lim x
k
Furthermore, we prove by induction t hat

2,
x
ii
k



3
ii
k
x



is in the form

,,
2,,
x
x
SQ

 ,
ere
S
is an algebraic expression invowhlving multi-
plicatiummation
claim
on and s
is true for
of

1
2
,,
,T
xx
QQ

 , Q. The

2,x
k
it is
in (8). Suppose
true for

2,
31
x
ii
k



, then by the chain rule,
the term
2
,T
x
QQ
1
panded into multiplica-
tion of th
,x
Q
, Q, hence the
will be ex
e same terms
form
1
2
,
,T
x
Q

,,
2,,
x
x
SQ

 is established. It is clear that
020 essentially
lim
by the Theo-
,,
, ,
xx
SQ

rem 2.1 and Propositionivigher order 2.3. We also dere h
partial derivatives explicitly in the Appendix A.
Theorem 2.2. For any 1,

,0,xX

Copyright © 2011 SciRes. AJCM
S. Y. CHEN299

1
,
21
2
2
()< ;
xn

**
1
,1
,,
22
0() ,< .
lim xx
mx
nuu



Furthermore,

1
22
2
**
1, ,
22
,.
sup xmx
xX uu



Proof. follows the definition of the norm 1
n
e finite-
,
Propositions 2.2, 2.3, 2.4, and Theorem 2.1. Th
ness of *,
j
x
u, =1j, ,
mption of the
. Furthermo
m, follows the relative com-
pleteness two-stage linear problem and
duality thre,
assu
eory
X
is compt, hence
o (1
gence e.
ac
<.
3. First Order Convergence Rate
We first discuss the convergence of the objective func-
tion specified in the approximatin model (2) to the ob-
jective function of the true model), and the conver-
rat
Theorem 3.1. For all
x
X,





21
1,
0,1 =1
,, .
kk
dd
k
xd wxK
 

Hence,
K
log d
KK
.
Proof. For notation convenience, define operators
and

,
K
kk
w x
 




0,1
=1 ,
d
k
xd uniformly
E
K
A
as



0,1 d,
d
Ef f


=Kkk
K
Af wf
=1k
.
Th for any

,0,1,xX K
 ,
en














,
,,
,
,,
,
,
K
x
xKx
KxK
ExAx
Ex E
EA
AAx





 

 

Taking limiting on both sides, then the first
and third term on tnd side go to zero by Theo-
rem 2.1, and the sec is bounded by the classical
convergence rategrid method, see (3), Proposi-
tion 2.3 and Theorem
Let the objective function of the true problem (1) and
the approximated problem (2) be
0
he right ha
ond term
of sparse
2.2.




0,1
=1
=,d,
=,,
d
K
Tk
k
Kk
zx cxP
zx cxwx
Tx


an
d let the optimal objective value and optimal solution
set of the true model and approximated model be
*** *
,,,
K
K
3.3 state the resu
zXz X
respectively. Theorem 3.2 and Theorem
lts for the optimal objective value and
the optimal sets separately.
Theorem 3.2. The optimal objective value converges,
i.e., and the rate of convergence is
**
limKK
zz

,


21
** 1,
log .
d
Kd
K
zz
K

Proof. For the minimization problem we note that
*
zx zx for any **
,
x
XxX
, and
KK K
x zx
for any *,
KK
z
x
XxX
. Let

2
1,
log
=
d
Kd
K1
K

, then

 

**
=

K
KKKKK
K
z
KK
K
zxz xzxzxzxzx
xz x




 
 
***
*
,
*
*
=
K
KKKK
K
zx zxzx zxx
zx x

 


K
where the inequalities follow Theorem 3.1.
Theorem 3.3. For any
K
z
zx z
***
,
KK
x
Xx X
,
K
x
1) is feasible;
2)




21
*1,
log
2
d
Kd
K
zx zx
K

;
*of a subsequence 3) For any clust eri ng p oi nt
x
*
,,
K
KK
ttt
x
txX

, **
x
X
; furthermore,
**
=if
X
x
is a singleton, then *
=*
x
x
.
Proof.
K
x
satisfies the first stage constraints and by
the relativmpleteness assumption,e co
K
x
is feasible.
Toow t shhat
K
x
is also very close to amal solu-
tiom
3.1.
ny opti
n, we apply the similar technique used in Theore

**
K KK KK
zxz xz xzx



2,
2 ,
KK


since
=
K
zx zx
K
KKK K
zxz x
 

by Theorem 3.1, and
*
K
KK
zx zx

by the steps in the proof of
Theo *
x
is a clustering point,
rem 3.3. Since

** *
0
limtKt
zx zxzxzx



by the ine-
quality above. As a special case, if
*
=
X
x, then
**
=
x
x
.
tThe.3 is a classical result hird result of Theorem 3
Copyright © 2011 SciRes. AJCM
S. Y. CHEN
Copyright © 2011 SciRes. AJCM
300
based the uniform conve
bsequal sets are
eto, and onxpect op-
timality of clustering points. The result
h [23,24
onverg cnvergenc
K
he sparse grid method for the stoch-
mming not only converges but
order rate. Our constructive proof
Mathematics, Philadelphia, 2005.
R. J.-B Wets, “Epi-Convergency of Con-
Programs,” Stochastic and Stochastic Re-
ports, Vol. 34, 1991, pp. 83-92.
chastic Pro-
rgence, see Römisch [22]. The
result is stated in suence since the optim
not necessarily singlnse can only e
can also be
proved by epi-convergence, see Attouc]. Epi-
cene is implied by uniform coe, see
all [25].
4. Conclusions
The modern sparse grid method is very efficient in nu-
merical integration for integrant functions the Sobolev
space r
d
. However, the integrand function in two-
stage linear programming does not belong to r. We
prove that d
astic twot
stage linear progra
converges in the first also
uses a logarithmic mollifier function from interior point
method.
5. References
[1] J. R. Birge and F. Louveaux, “Introduction to Stochastic
Programming,” Springer, New York, 1997.
[2] S. W. Wallace and W. T. Ziemba, Eds., “Applications of
Stochastic Programming,” Society for Industrial and Ap-
plied
[3] A. J. King and
vex Stochastic
[4] A. J. King and R. T. Rockafellar, “Asymptotic Theory for
Solutions in Statistical Estimation and Sto
gramming,” Mathematics for Operations Research, Vol.
18, No. 1, 1993, pp. 148-162. doi:10.1287/moor.18.1.148
[5] A. Shapiro, “Asymptotic Analysis of Stochastic Programs,”
Annals of Operations Resesrch, Vol. 30, No. 1, 1991, pp.
169-186. doi:10.1007/BF02204815
[6] J. Dupacova and R. Wets, “Asymptotic Behavior of Sta-
tistical Estimators and of Optimal Solutions of Stochastic
Optimization Problems,” Annals of Statistics, Vol. 16, No.
4, 1988, pp. 1517-1549. doi:10.1214/aos/1176351052
[7] T. Pennanen and M. Koivu, “Epi-Convergent Discretiza-
tion of Stochastic Programs via Integration Quadratures,”
Numerische Mathematik, Vol. 100, No. 1, 2005, pp. 141-
163. doi:10.1007/s00211-004-0571-4
[8] S. A. Smolyak, “Interpolation and Quadrature Formula
for the Class a
s
W and a
s
E,” Doklady Akademii Nauk
SSSR, Vol. 131, 1960, pp. 1028-1031. (in Russian, Eng-
lish Translation: Soviet Mathematica Doklady, Vol. 4, 1963,
9129717644
pp. 240-243).
[9] T. Gerstner and M. Griebel, “Numerical Integration Us-
ing Sparse Grid,” Numerical Algorithms, Vol. 18, No. 3-4,
1998, pp. 209-232. doi:10.1023/A:101
Cost
[10] M. Chen and S. Mehrotra, “Epiconvergent Scenario Gen-
eration Method for Stochastic Problems via Sparse Grid,”
Stochastic Programming E-Print Series, Vol. 2008, No. 7,
2008.
[11] L. C. Evans, “Partial Differential Equations,” American
Mathematical Society, Vol. 37, No. 3, 1998, pp. 363-367.
[12] G. W. Wasilkowsi and H. Wozniakowski, “Explicit
Bounds of Algorithms for Multivariate Tensor Product
Problems,” Journal of Complexity, Vol. 11, No. 1, 1995,
pp. 1-56. doi:10.1006/jcom.1995.1001
[13] H. Brass and G. H¨ammerlin, Eds., “Bounds for Peano
kernels,” Vol. 112, Birkhäuser, Basel, 1993, pp. 39-55.
[14] H. Wozniakowski, “Information-Based Complexity,” An-
nual Review of Computer Science, Vol. 1, No. 1, 1986, pp.
319-380. doi:10.1146/annurev.cs.01.060186.001535
[15] C. Roos, T. Terlaky and J.-P. Vial, “Interior Point Methods
for Linear Optimization,” Springer, New York, 1997.
ew
De-
[16] N. Megiddo, “Progress in Mathematical Programming,
Chapter Pathways to the Optimal Set in Linear Program-
ming,” Springer-Verlag, New York, 1989, p. 132.
[17] A. V. Fiacco, “Int roduction to Se nsitivity and Stab ility Ana-
lysis in Nonlinear Programming,” Academic Press, N
York, 1983.
[18] O. Güler, D. den Hertog, C. Roos and T. Terlaky, “
generacy in Interior Point Methods for Linear Program-
ming: A Survey,” Annals of Operations Research, Vol.
46-47, No. 1, 1993, pp. 107-138.
doi:10.1007/BF02096259
[19] Y. Nesterov and A. Nemirovskii, “Interior Point Polyno-
mial Algorithms in Convex Programming,” Society for
2001.
Interior-Point Methods,” Soci-
800
Industrial and Applied Mathematics, Philadelphia, 1994.
[20] J. Renegar, “A Mathematical View of Interior-Point Me-
thods in Convex Optimization,” Society for Industrial and
Applied Mathematics, Philadelphia,
[21] S. J. Wright, “Primal-Dual
ety for Industrial and Applied Mathematics, Philadelphia,
1997.
[22] W. Römisch, “An Approximation Method in Stochastic
Optimal Control,” In: Optimization Techniques, Part 1,
Lecture Notes in Control and Information Scienc es, Sprin-
ger-Verlag, New York, 1980, pp. 169-178.
[23] H. Attouch, “Variational Convergence for Functions and
Operators,” Pitman (Advanced Publishing Programs),
1984.
[24] H. Attouch and R. J.-B. Wets, “Quantitative Stability of
Variational Systems: I. The Epigraphical Distance,” Tran-
sactions of the American Mathematical Society, Vol. 328,
No. 2, 1991, pp. 695-729. doi:10.2307/2001
Vol. 11, No. 1, 1998, pp. 9-18.
[25] P. Kall, “Approximation to Optimization Problems: An
Elementary Review,” Mathematics of Operations Re-
search,
doi:10.1287/moor.11.1.9
[26] T.-W. Ma, “Higher Chain Formula Proved by Combina-
torics,” The Electronic Journal of Combinatorics, Vol. 16,
No. 21, 2009.
S. Y. CHEN301
tion and
or a random vector on with invertible cumulative
ion
om
Appendix A. Inverse Transforma
Truncation
F
distribution function

1:0,1
d
F, the integrat
domain of a mollifier function can be transformed fr
to

0,1 d:
 

1
,,
d= d,
xx
FF

[0,1]d
 

then we apply the sparse grid method to generate scenar-
s and weights iofor on the cube. We need to
check the properties of the integrand
. Its differentiability on ly depends on

0,1 d
function

1
,xF
1()F
since ,()
xC
. Most commonly used invertible cu-
mulative distribution functions, for eple, inverse of
normal distribution function

1
, is also in C
xam
. The
finiteness of the partial derivatives of
1
,xF
only d
also
epends on
1
F since partial derivatives of

,x
are finite for any multi-index 1
compo-
nent-wisely.
The higher order partial derivative of aposite
function can be calculated exp licitly. For
:fg
hxXy Yz , we can ap-
ply the Faà di Bruno’s formula:
com










n
d==,
d
nnPB
nPBP
f
gxf gxfgxgx
x
where n
is the set of all partit
ions of the set n
J
of
intege . A partition of rs 1,,nn
J
s of is a family oir-
wise dinempty subsetf pa
sjoint non
J
whose un ision
n
J
.
A
means the cardinality of the set
A
. For a vec-
p
rm
tor comsite function:
:,
fg
hy Yz

 
We apply Tsoy-Wo Ma’s higher chain foula [26]:
o
x X

,, =1
11
=! !!
mk
p
msk
mp
k
spm kkk
z

z y
mpxy x





w here is the set of all decompositions of multi-index
with multiplicities Calculat ion involving a multi-
x m.
inde

1,,
 
 follows rules:
=1 =1
=,!=!,
= .
jj
jj
j
z
=1 =1
=,
j
j
jj
j
x
xz
x
x



A multi-index

decomposes into
s
parts
1,,
s
pp in
with multiplicities 1,,
mm in
respectively if the decomposition equation
1122
=
s
s
mp mpmp

holds and all parts are different. The total multiplicity is
defined as
12
=.
s
mmmm

The list
,,
s
pm ed a is call
-decomposition of
.
nsure all parts are different we may impose
12
0
To e
pp p , where


means =,,=
111
j
jj
 
, but <
j
j
, for a
j
.
For the problem under discussion, let 1
compo-
nent-wisely, i.e., =1r, note that are 21
d number of
such
(s). Folg the higher c formula,
we get lowinhain rule



1
,
11
=.
x
k
F



,!
m
p
msk
x
mp
k


ce
,, =1 !!
spmkkk
mp


Furthermore, sin
2
0,
mx

by Proposi-
tion 2.3, the computation of
=0
li

1
F
0,
lim x
can be simplified significantly. In this case, only the de-
compositions
1,, i
e
,
-zeros in the
i
e is the th unit basis of cor-
respond to nonula. Otherwise, for
i
above form
d
,
, 1
=
s
m
mm
, and

0,=0
lim
m
x
m

.
2
s
Hence


,
00=1
imdxi
iiw
1
,=
lim l
xi
F
*
,
=1
=.
di
ii
x
i
uw





Hence,

1*
,,
1
0
=<
d
xi
=
1
212
,
lim i
ix
di
Fu
 


if and only if
<
i
almost surely
1,=1, ,id
 . For some distributions, the condi-
tion might not hold. For example, the inverse of a cumu-
lative function of the normal distribution does not have
this property nearby 0 or 1. To remove the singularities,
truncation of the cube [0, 1] could be applied:
where


d
0,1 ,1
()d d,
dhxxhxx


0< <1
han
we need t
is a small positive number. To com-
d siderd sparse grid
method,o change the variable to, where
pute the right using the standa y
Copyright © 2011 SciRes. AJCM
S. Y. CHEN
302

=12:0xy
 
,1,1
dd

1

 . Hence

 



,1 0,1
d=1212 d.
d
dd
hx xhy y
 


Hence for a two-stage linear problem with an invert-
ible but unbounded cumulative distribution function
F
, we shall first generate the standa
weights


,


1
=12,=12,
d
kkkk
F
ww
 
 
finally use the and
,
kk
w
in the approximation
model (2)
The erproximation model is exactly the
.
ror of this ap
sum of truncation error t
e, and sparse grid approxima-
tion error
s
e. Erro
rd grid points and
=1
K
kk
w
using the sp
karse grid method,
then scale and transform them to the original random
variable
by
r goes down with
t
e
and error
s
e goes do wng wn ith increasi
K
at der rate
sparse gethod. the first or
of rid m
Copyright © 2011 SciRes. AJCM