This content was uploaded by our users and we assume good faith they have the permission to share this book. If you own the copyright to this book and it is wrongfully on our website, we offer a simple DMCA procedure to remove your content from our site. Start by pressing the button below!
Oi,i
(2.68)
= l^.
The deterministic equivalent of problem (2.67)—(2.68) is (2.69)
max MX), Flx(gi{X))
= m,
ffi{X) < 0,
(2.70)
i=T~n^
i = T>.
(2.71)
where we are required to define (deterministic) vectors X and g(X) = {gi(X), ...,gm(X)}. When we have a continuous distribution function, the deter ministic equivalent of problem (2.67)—(2.68) becomes (272)
max MX)
(2.73)
F-}(ai) <0,i-T^.
We shall consider a chance-constrained stochastic programming problem in nota tion 2): max MX) (2-74) P{f(X)
< 0} > a.
(2.75)
In accordance with the following assertion, the deterministic equivalent of problem (2.74)-(2.75) is max MX), (2.76) Fx(g(X))
= a,
(2.77)
g(X) < 0,
(2.78)
where we are required to calculate deterministic vectors, X and g(X). T h e o r e m 2.1. [35] If the joint distribution function F(X) of components of a, random vector f(X) = {fi(X),..,, fm(X)}, is continuous for each X, then problem (2.76)—(2.78) is a deterministic equivalent of the stochastic problem (2.74)—(2.75). Proof. If we prove that both problems have the same domain of definition, we prove the theorem, since the objective functions of both problems coincide. Let g(X) satisfy the conditions of problem (2.76)—(2.78). Then P{f(X)
= a,P{f(X
<0}>a,
i.e. X satisfies the conditions of the stochastic problem (2.74)-(2.75). Now let X be a feasible solution to problem (3.69)-(3.70). We have P{f(X)
< 0} > a.
2.4. EXAMPLES
OF STOCHASTIC
PROGRAMMING
PROBLEMS
23
For each X, we divide the numbers of the vector f(X) components into three sets: ! x'i Ix^fpi where i G I$] if f,{X) is a deterministic variable; i 6 1$ if f)(X) is a random variable with a negative upper bound, /,(A') < J,{X) < 0; finally, i E / f if fi(X) is a random variable for which P{fi(X) > 0} > 0. The plan of problem (2.74)-(2.75) defines the deterministic vector
g(X) =
fi(X)
,ieq(1)
f,W
,i€l(2) x ,ielx
(3)
We have P{f(X)<0}
=
P{f(X)a.
In accordance with the condition, the distribution F-^(g(X) is continuous and hence there exists g(X) such that
g{X) < g(X) < o,
P {f(X) < 9{x)} = a.
The vectors A' and g(X) then satisfy the conditions of problem (2.76)-(2.78). This completes the proof of the theorem. To solve applied problems, one may efficiently use computational algorithms and programs where the deterministic equivalents of stochastic problems are the linear and convex programming problems. In this respect one may address oneself to [36] which distinguishes those cases where convexity is ensured for the objective function and the domain of definition of a deterministic problem that is equivalent to a stochastic model. To solve convex stochastic programming problems (to compute posterior decision rules), one may use any numerical method of convex programming in the Hilbert space. Problems of mathematical programming in function spaces can be effectively solved by the method of feasible directions which is summarized and justified in [40] for finite-dimensional convex problems. In addition, the methods studied in [27], [28] can also be used to solve stochastic programming problems.
2.4
Examples of Chance-Constrained Stochastic Programming Problems
In practical problems of planning, designing and management, one has to take deci sions when the information on initial data is insufficient. In general, stochastic models for decision-making in complicated situations prove to be more adequate to actual phenomena and processes than deterministic statements of problems. In the setting of a stochastic problem, the objective functional and its domain of definition are often
CHAPTER
'2-1
2.
CHANCE-CONSTRAINED
SPP
taken to be statistical characteristics of t h e event being modeled, such as m a t h e m a t ical expectation, variance, and probability of falling within a r a n d o m region. Such probability can be conditional or unconditional. Below are given examples of chance-constrained stochastic programming problems in various applications. Following [9], we shall consider the coal output planning problem where initial d a t a can be only partially predicted. Suppose we have several coal mines. We are required to work out for the i-th mine i = ( 1 , 2 , . . . , m ) several development projects j = ( 1 , 2 , . . , , » ) a n d define for each mine development projects t h a t are most practical from t h e point of view of t h e coal mining association as a whole. Let x,3 equal 1 if t h e j - t h project is to be implemented for t h e i-th mine, and 0 otherwise. Note t h a t only one development project can be implemented for each mine. Hence ^ l y
=
1,» =
1,2,...,TO.
3= 1
We introduce t h e following notation: cy(w) — a n n u a l mining costs for the i-th mine according to t h e j - t h project when w random parameters of conditions determined by n a t u r a l factors are realized; dij(uj) — annual output for t h e i-th mine according to t h e j - t h project when u random natural factors are realized; stJ(u>) — coal ash content at t h e i-th mine operating according to t h e j - t h project when UJ random conditions of t h e problem are realized; fcy(w) — annual wages fund for t h e i-th mine operating according to t h e j - t h project when w random natural factors are realized; d — t h e annual coal o u t p u t for t h e whole association as required by a higher level plan (or as determined by demand); 3 — admissible coal ash content for t h e mining association as a whole; k — t h e total annual wages fund for t h e association; ad,ot3,ak -probabilities t h a t t h e restrictions placed by a higher organization on satisfaction of demand, coal ash content and wages fund are observed. Using t h e above notation and terminology, t h e coal o u t p u t planning model for the mining association can be represented as [ m n,
\
£|£Ece(w)*tf|->min> P
p
(
m
n,
i
J E E 4 > H z , ; , > 4 >ad, E E ^ » u < s
(2-79)
>«.,
(2.80) (2.81)
2.4. EXAMPLES
OF STOCHASTIC [|
PROGRAMMING
m 77i n, n,
PROBLEMS
1
PJ££M")*o <*}>«*. (.. = 1 . 7 = 1 J P lT.ll
25
kij{^)xhak,
(2.82)
n, 71,
.,m, = l1,2,. £ z < , = l>t = 1 2,...,m,
(2.83)
1=1
{ 0 , l } , j = 1,l ,...,n,; .., n t ; i = 1,1,. = {0, ...,m. xs,j ., m. v = The solution of this problem is a random vector-plan for coal mining in the subsequent period. In this case, the future development planning of the mining association incorporates actual realizations of natural factors. Suppose the parameters dt},st),kij are random variables that are mutually independent and normally distributed. Let c,j, dij, 5,-j, ktJ, a2(ctJ), a2(dtj), <72(s,-j), a2(ktj), be respectively the mean value and the variance of random parameters of the problem. The deterministic equivalent of problem (2.79)—(2.83) becomes m m
n, n,
min^^c,;Xij
(2.84)
;= i ., = i i 1=1
Ei>w4>rf.
771 7n 771 1 ,,
m
-<*A i = l j; == 1l .=1 77i 771
-°-\
,. ==i i jJ == ii
EE^
(2.85)
\|. = l j = l .=13=1
7n 1 ,,
771 7Tii 771 1,
rii
m 7i[ n, m
(2.86)
^ ,. = ii _, ,== ii m
n,
+ * " ' ( i - 4 £XM*«)4 >fc,
(2.87) (2-87)
JTTXx'J^ = ,i = = lM = 1,2, 1,2, ...,m, ...,m,
(2.88)
\ | .. ==i i j}==ii
>=i j = i
ii = = ii Xii X%J
= {0,1}, j = l,2,...,n,;j 1,.., m. 1,2,...,«,-;» = l,...,m.
(2.89)
The above statement of the coal output planning problem with random data is the simplest one, but it can be complicated, e.g., by assuming that random parameters in the conditions of the problem are correlated. This problem can be viewed as dynamic when the development planning and actual realization of natural factors are accomplished in periods (in stages). Another example is provided by the agricultural production planning model which is actually a chance-constrained stochastic LP problem [7]. We introduce the following notation: m — number of farm crops (1 < i < m); n, — number of crop grades; p — number of types of soil on which agricultural production is to be carried out (1
CHAPTER 2. CHANCE-CONSTRAINED
26
SPP
Sk — available area of soil of type k; R, — planned output of the i-th crop; 1 — a.-admissible risk of failure to fulfil the 2-h crop production plan; Vijk — yielding capacity of the z-th crop of grade j on the soil of type k; ctJk — cultivation costs per unit area of the fc-th soil for sowing the i-th crop of grade j ; l—j3 — admissible risk under which the actual costs may exceed the planned costs; x ijk — the fc-th soil area for the i-th crop of grade j . Using the above notation, the agricultural production planning problem becomes y ->■ min, EEE
c
u P V - < y \=I3,
1=1 ; = 1
p
fc=l
(2.91)
J
\ E E ^jkx1]k >Ri\> m
(2.90)
a,; (1 < * < m),
(2.92)
n,
£ £ > , ■ , * = &, ( ! < * < ? ) .
(2.93)
Xijk > 0; i = l,m; j = l,n,-; k = I, p.
(2.94)
We have a knapsack problem which can be formulated as V-model. This problem can be solved by employing the confidence region method. The original problem is of the form:
{
jI
[ nn
m'mfs.t.P
< Y1CJXJ b =l c
7i n
< / f > ai^2aJxi J
x
a
3=1 a
x
= M < £j < bj,j = l , n b
x
b
E i i ^ / 1 > ' E i 3 = >° < i < J'j = M 71
(ft) (Po)
It is assumed that a3 > 0,i ; > 0 are constants with j = l,n; E a}bj > b; 1/2 < a < l;c = (ci,...,c„) is a random vector of the multinormal distribution N(iu.,V), where fj, = ( / i i , . . . , ^ ) are unknown and V is a diagonal covariance matrix V = diag(a PQ is equivalent to P0 min - ^ ^1/2 ^ s.f. .t. nin E //UjSj W + M EE o ffjij) j=i 71
j=i
(Po)
a + E ajXj = b,0< XJ < bhj = l,n, j=i
where /s.'a = $"'(0-),$ is a normal distribution function of /V(0,1). To construct a minimax model, we construct a confidence region S for distribution of parameters (jj.1,cr'j) based on a random sampling from the relevant multinormal distribution of N(fM,V). Suppose we have:
2.4. EXAMPLES
OF STOCHASTIC
PROGRAMMING
PROBLEMS
27
m = (mi,...,m n ) is an average sampling vector; S = (S2,..., Si) is a sampling variation vector; N is a sampling dimension; (3 is a significance level. We shall construct confidence regions for (i}. Since JNZu
£ (/"j _ Wj?IS]
' s governed by the probability law of F-distribution
with (T», N — n) degrees of freedom, the confidence region for n3 is given by:
pp> - v,m * w(N^)Fs{n'N -n)-
(2 95)
-
where Fp(n, N — n) is a quantile of F-distribution with (n, N — n) degrees of freedom for the significance level of 0. We now construct confidence regions for a2 For simplicity, we assume that the variables c, are distributed independently. Since (N — l)S2/a2 is distributed in accordance with chi-square distribution with (N — 1) degree of freedom, the confidence region for <x| is given by the constraints: (N - l)S*/xl(N
- 1) < o) <(N-
where v = 1/2 — /3 1//n and \l(N with N — 1 degree of freedom. Now S is given by
- l),j = M ,
l)S*/xU(N
(2.96)
— 1) is the i/-th quantile of chi-square distribution
S = {(^,,...,^ n ,<7 2 ,..., ^ s a t i s f y i n g (2.95) and (2.96)j = T^n}
(2.97)
Using S, we formulate the minimax model n
min max5"^fijxj i=i
/ I n
+ Ka j
\ 1/2 \
J^
v=i
n
s.t.(fj,,A)^S,^2ajX3—b,Q<.x3
(P)
s=i
/
where A = {a2, ....o-*). We consider an auxiliary problem together with P: maxE^;+^(E^^2)1/2 s.t.{n,S) C S The nonnegativity of i ; implies that the Pl-optimal variables a2 = (N-
l ) 5 2 / x ? _ , ( ^ " 1).J = T ^
We substitute (2.98) into (PI) replacing the variables fj = (fj,3,...,m3)/s3. (PI) goes over into (P2): max ^
5JXJ6
+^2mjXj+
(2-98) Then
CHAPTER 2. CHANCE-CONSTRAINED
28
SPP
1/2
+ka {(N - l)/xlJN
- 1)}' / 2 (Tstf)
a
. t . £ £ < <&. > 0 ,
(P2)
where <% = $E$F0(n,N - n). Let us introduce a subproblem p2(A): min
x
jii
> A:
fi > o.
3
Let £(A) = (£i(A),...,£ n (A)) be an optimal solution to P2(A). If £(A*) satisfies i) and ii),then f(A*) is also an optimal solution to P2; i)E£j(A*) = d^, ii) E ^ ^ ( A * ) = A* Since P2(A) is a convex problem, its solution satisfies the Kuhn-Tucker conditions i/j = (j -r)s}xhj
= \,n
E sjx^3 > X, & > 0, v} > Oj = T^rj > 0 i Vjti = 0j = TjlT](\ - £SjXjfj) = 0,
(2.99)
where r],v} are the Lagrange multipliers. Hence it follows that the optimal solution of P2 is given by : a/ 2 10
e; = 4 > w (E*M) \i = ^ 12
( - °)
and the optimal value of (P2)= £ m ; X j + ^(E s]x)Y , where
s.t.^ctjXj
— b,
0 < x: < b3j = 1, n.
-
(2.101)
The finite solution algorithm for problems of this class has been developed and yields a solution in 0(n2 log n) steps. We have chance-constrained stochastic problems min/(x) s.t.P{ft(x,f(w)) < 0,» = l,...,m} > a
^■iU^
min/(x) s.t.P{g,(x,{t(uj)) < 0 } ><*,•,» = l,...,m,
(2.103) (2.104)
where £ is a random vector; £, is a random variable defined in the probability space (ft, F, P); a, a, are confidence levels. Denote by s(f, a) the set of feasible solutions of problem (2.102). Let us consider a sequence {x\3',a^},j = l,...,oo. We study the behavior of s(£"',a' 5 ') where £"' converges to £, while a' J ' converges (in a sense) to a. Such studies were made by Salinetti for linear constraint functions and a special form of convergence.
2.4. EXAMPLES
OF STOCHASTIC
PROGRAMMING
We shall find conditions under which lim s(^J\
PROBLEMS
29
a ( j ') = sU,ct). Here the limit is
j—► oo
taken in the sense of Kuratowski, i. e. lim ,s(£' ; '',a' J ') = s i-H-oo
lim sups(£ U J a l J ) ) C s(f,a) C lim inf s(CJ>a where n s u p s ( { ( j ) a w ) = {x : x}k -> x;xJk g s f ^ W ' * ' ) lim infs(f y ) a ( i ) ) = {x : x3 -^ x;x3 e
s(({j)a{f))
except perhaps a finite number of points Xj}. Let Fg(x^(t) be a distribution function ofg(x,£); then P{g{x,£) < 0} = F s(x{ )(0); P{9(*,(U))<0} = Fg{l,ilj))(Q). Denote G(x) = Fg{x<(){0);G^{x) = Fg{lJiij]){0). Then (2.102) is equivalent to G(x) > a, and s((,a) = {x\G(x) > a} is equivalent to s(£ (j) , a(j>) = {x\Gu)(x) >
o«}. The following results were obtained for the model studied. Theorem 2.2. Let g(x,u) be continuous and suppose that for any £ the vec tor g(x,£) has a continuous distribution function. Then lim G^\x) = G(x) when j—foo
fW4f. Theorem 2.3. Suppose that: i) the conditions of Theorem 2.2 are satisfied; 2) there exists a sequence {x^} : G^'(XJ) > a} and x3 —¥ x; 3) ^ 4 £, a' j ) -)■ a. Then a; satisfies G(x) > a. From Theorem 2.3 we immediately have the inclusion lim sup s({«W J ') C s(£, a). j—H-oo
For the other inclusion, it suffices to impose an auxiliary weak condition on G(x). T h e o r e m 2.4. For any x : G(x) = a let there be sequences {xk} and {ejt}; G(xk) = a + sk, tk > 0, £k -> 0, xk -> i . Under the conditions of Theorem 2.1, for any x we then have from s(£,a) that there exists {x3} : G^3\x1) >ctj and XJ—>x or, which is equivalent, lim inf s({' J 'a' J ') 3 s((,a).
The constructive proof (i.e., the method of constructing x} in terms of xk) is also provided. Combining Theorems 2.3 and 2.4, we obtain the continuity theorem for the set of feasible solutions to problem (1). Theorem 2.5. Let £ (j) be such that ( (j) -4 f, a w -4 a. l j g(x, 0 is continuously distributed Vx; (x, u) is continuous; 2) Vx : G(x) = a, 3{xi} : G(xk) = a + £fc, tk > 0, £t -> 0; £* -J- x. Then lim s u p s ( £ ' J ' a " >) C s(£,a) C lim inf s(£«>ak')). j'-H-oo
J-Voo
CHAPTER
30
2.
CHANCE-CONSTRAINED
SPP
T h e implication of conditions 2) is t h a t for the boundary points of t h e set s(x, a) x0 there is no neighborhood of U(x0) belonging to the same space as s(£, a ) such t h a t G(x) = a for x G U{x0) l~l s(f, a ) . A simple criterion is given for verifying conditions (2). T h e o r e m 2.6. Suppose that: 1) chance constraints are of the form h(x) = P{gi(x,£) < 0,...} > a ; 2) gi, ...,gm are concave; ( has a log-concave density distribution function; 3) s(f, a ) has at least one point h(x) > a. Then any boundary point x0 of the set s(f, a ) i a s a sequence {xi:} : h(ik) > a and xk —>■ XQ. We now revert to problem (2) with individual constraints: (2)min/|Gi(x)>al. T h e o r e m 2.7. 1) Let g, be continuous in any x and let g%(x,^t) be continuously distributed; then limsup S ( ^ ( J ' Q ; ' J ) C s ( f , a ) as soon as £' J ' —y £ , a ' J ' —> a, where f = ( 6 , . . . , ^ ) ^ W = ( d i ) , - , ^ ) , " = ( a l l . . . , a m ) , a W = (aF)I...,aW); 2) ifG,{x)i = l , m : V.r 6 s(f,a)3{£)fc} { i | G , ( s ) = a , } and conditions 2) of Theorem
xk -> i-;G,(i, ; ) > a,Vi G / ( x ) = 6 hoid, then lim inf s ( £ ' J \ a ' J ' ) C
s ( £ , a ) . What is given below may be taken as a test for condition 2). T h e o r e m 2.8. For problem (2), iet ah G,(x) be log-concave functions and suppose there exists at least one point ~XQ GZ(XQ) > a;V?. Then for any boundary point x of the set s ( f , a ) there exists {x^} '■ G,(ibk) > oit\/i,Xk —> x (the "boundary point" means here I(x) ^ 0 ) . The main result is the stability theorem. T h e o r e m 2 . 9 . Let f, { { ' J ' } , a, { a ( j ' } be given. Let Z, { Z ( j ) } , M, {M<J'>} be optimal values and sets of optimal solutions to problem (1) or (2) with the problem parameters specified. If the conditions of theorems 2, 3, 6 hold, then lim sup i l / ' J ' C J—^oo
M. Furthermore,
: x G M,xu~> G M ( i i ; i ( i l -> z, '-hen lim Z ( j ) = Z.
if 3 s ; {x{l)}
We shall now consider applications of the cutting-off m e t h o d to chance-constrained convex programming problems. Suppose we have the problem: min{c'a; : x € R},
(2.105)
where R = {x\P(g(x)
> f(b)x
G Rxb G Rb{x
< b} > p}
(2.106)
are convex polyhedrons, and 6 is a random vector of dimension m, A = (m x n). For simplicity, we assume f[b) = /i(&i),..., fm(bm). If there are no constraints as in (2.106), then problem (2.105)-(2.106) reduces to a certain deterministic convex programming problem min{c'x : x G R} (2.107) R=
{x: P[Ax
pxe
Rxb G R.b}
(2.108)
2.4. EXAMPLES
OF STOCHASTIC
PROGRAMMING
31
PROBLEMS
(2.107)-(2.108) reduces to (2.109)-(2.110): minjc'j- : x £ R} R= {x,b: Ax>b,b£
(2.109)
Rb,x£
Rx};
m
II[l-ft,(i.)]>P,
(2-110)
i=l
where ^,(6,) is a distribution function of bt. There are optimization methods reducing the convex programming problem to a sequence of linear problems. Their general scheme is as follows. Construct Rk: Rk
D
Rk+i
_
D
D Rxk
_
m g m m
^
x
\ .
x
g
Rk}
Rk is a parallelogram. It is stated that xk —> s,where x is a. solution to a convex problem. In order to solve problem (2.105)—(2.106), it is proposed to use the Kayley cuttingoff scheme. Step 1: Imbed R in R° R° = {x\P(Ax < b) > pb £ Rhx e Rx}. Let x°,6° solve for R° Let us define the residue maximum max 1 < j < mi{gJ{x) — fj(b3)} = 9jo(x) ~ /JO(^JO) anc ^ expand g]0{x) and fJa(bjo) into a Taylor series restricting our consideration to linear terms. The following approximation of x 1 , b1 is obtained from problem (2.111)-(2.112): {m'mc'x\x £ R1} (2.111) R1 = {x\P[gJz°)
+ V 5 j o (x 0 ) T (x - x°) < fjo(b°0) + f'Jb°J(bM
- bl)Ax < (2.112).
p}
By (2.109)-(2.110), problem (2.111 )-(2.112) is equivalent to the following problem: Let x\bl be a solution to (2.111 )-(2.112). Find max {gAx) - /,(&)} = l<j<mi
= 9h ( x ) ~ /JI (^3i )> ■E2' k2 We seek as a solution to the problem: min-fc'x : x £ R } R2 = {x: P[gJ0(x°) + VgM(x°)T(x
- x°) < fJb°M) + f'K(b°n)(b3a - &°> j l (x 1 ) +
+ Vgn(xl)T(x - x1) < ffliVfl) + &(4k)(6,-i - ^ M * < M € f l J 6 ft] > PJ Which is equivalent to x, b : min{c'x|x £ R2} R2 = {x,b: Ax
Rb,x £ Rx},
fc.(*°) + Vft 0 (x°) (x - x 0 ) < fh{b%) + f'Jbl)(b30
- b%))
CHAPTER 2. CHANCE-CONSTRAINED SPP
32
The fc-th step is written is a similar way. Convergence is proved in much the same way as t h a t of the Kayley scheme. It is pointed out t h a t alternative methods of imbedding can also be used for solution of chance constrained stochastic problems. We shall now formulate a regional development problem with due regard for river pollution. To this end, we shall consider the collection of production facilities that are located in territory of a river basin (a region) and allow reconstruction. For the purposes of building new projects, suitable areas are chosen within the region. T h e regional development model is formulated as a chance constrained stochastic programming problem. We shall need the following constraints: 1. for execution of the production program specified for t h e regional economic system (within a river basin), we have
P< I E E QU")*lk > £.H I > «.'; i = U,
«) > o.5; u e fi;
Q
J2x>k = l;i = l,I;k=
1, A";
3=1
x]k = {0,1}; i = Ij;k
= 177T; q = T^Q,
where Bi(u>) — the regional production program for basic products i\ a'k(w)-output of basic products i by the technological m e t h o d q at a point k (the river is divided into K sections and the section number is assumed to coincide with t h e production facility seat number); x ik — t n e integer variable which shows whether or not t h e production program incorporates the technological method q (whose capacity is aqik) for producing basic products i at the point k; a] — t h e probability of satisfying conditions of the first group; w,w € fi — a random realization (state of nature); i 6 / — the index of basic products produced in the region (I - t h e complete range of basic products); q, q 6 Q — the index of the technological method (Q is the set of all technological methods for producing basic products); k,k e A — t h e river section number (the number of the seat of new production facilities); 2) for execution of a balance on production of sideline p r o d u c t s , we have
p
f i t £XX*M^ + E E E ^ H A - »', = oj > a) (t=ig=i!=i
/t=ip=i;=i
a) >0.5;j =I77;w G 0, p
Es& = l;' = V ^ = M?;
J
2.4. EXAMPLES OF STOCHASTIC PROGRAMMING PROBLEMS
33
2& = {0,l};; = l,/;fc = l,A';p = l,P, where tfjiki1^) — production of sideline products j by realizing the technological method q for producing basic products i at the point k; c^s/t(u;) — production of sideline products j when wastewater from production of products i is treated at waste treatment facilities, with the technological method p realized at the point k; Wj — total production of sideline products j for the region as a, whole; yfj. — the integer variable which shows whether or not the waste treatment facility (with the technological version p) is used for treatment of wastewater from production of products i at the point k; a* —the probability of satisfying the constraints of the second group; j,j £ / — the index of sideline products (/ is the range of sideline products); 3) for use of tight resources (investments,labour resources,fuel and power resources, material resources), we have
p't(E£
= I ' E t ^ ) ^ < O „ H } > <*l °2 > o-s
U=i
;=i
>
OJ e
n,
v—
I, R,
where dqkui{u>) — consumption of tight resources v at the point k where the technological method q is used for production of products i; Dnu(u>) — the total supply of tight resources v in the region; v,v G R — the index of tight resources (R is the set of resources); al — the probability of satisfying the constraints of the third group; 4) for the fresh river water intake, we have
PtlJZT.
vli»>)*U < Vk(u) j > a*kl 4 > 0.5; k = I,K\UJ e n,
where v UJ lii ) — t n e m t a k e of fresh river water at the point k where the technological method q is realized for production of products i; Vk{u>) — river discharge at the point fc; OL\ —the probability of satisfying the conditions of the fourth group; 5) for a balance on "production" and treatment of wastewater, we have Pi j E fXWzl
= FiM
+ * a ( « ) J > a5tk,
a% > 0.5,
CHAPTER 2. CHANCE-CONSTRAINED
34
SPP
k = l,A';i = 1J;UJ G ft,
^ Ifl*= £ <&(")!/& } > <*%, <4 > 0.5, fc = l , / f ; J = T77;w G n, where ffk(io) — daily discharge of wastes from production of products i at the point k by the technological method <j; F,-fc(tj), $t*r(w) — respective volumes of treated and nontreated water from produc tion of products i at the point k; 9ik{u) — capacity of the waste treatment facility which is used for treatment of wastewater from production of products : at the point k by the technological method p; a\k, afk — the probabilities of satisfying the conditions of the fifth and sixth group, respectively; 6) for a balance on wastes discharge, we have P«L7{FlkHa!kH
+ * « H A * = cjk(u>)} > afk,
a-*! > 0-5;fc= hk;« = 177; T = I7TT; w e a, where ajjfc(«),j0jj.— respective concentrations of pollutants TT in treated and nontreated wastewater from production of products i at the point k; cffc(w) — disposal of pollutants n from production of products i at the point k; ajk— the probabilities of satisfying the conditions of the seventh group; 7) for a river water balance at the point k, we have
p£{vk{u) - vi(w) - F W (W) - * w (w) - F ; H - *;(«) = o} > <4, jfe = T7I;w efl;a8k>
0.5,
3
P, {H(w) - 14-r(w) - V k M + V i / H + VT(u) = 0} > a t /o = 2,..., if; o:| > 0.5; a; g fi, P
10
{ 1 4 H - V4»(w) + Vk,H
+ Vk*{u) = 0} > a 10 ,
fe= l ; a ' 0 > 0.5;u> £ 0, where 14(w) — discharge of gray water in the river at the point k subsequent to wastes disposal; Vk(u>) — river water discharge at the point k prior to wastes disposal; Vkn(u>)— natural water input at the point k; Vk*(u>) — total industrial water consumption at the point k; Ftr(w),$fc/(w) — disposal of treated and nontreated wastewater from communal activities at the point k;
2.4. EXAMPLES
OF STOCHASTIC
PROGRAMMING
PROBLEMS
35
■f1fc(tJ)>^,it(w)—total discharge of treated and notreated industrial sewage water at the point k; Vki{w) — domestic consumption at the point, k; Vkn{^>), Fkr(u>), $fc/(w), Vki(u) —exogenously given parameters; a /ti ak> al°-the probabilities of satisfying the conditions of suitable groups; 8) condition on nonnegativity of continuous variables Wj > 0; Ftk{u) > 0; Phiih{u)
> 0;/?£ > 0;c&(w) > 0;
$ * M > 0; Vk(u>) > 0; Vk{w) > 0; V£(w) > 0; w 6 fi, J = T77; i = 177; k = 171; 7T = TjT; The objective function of the model is taken to be the threshold minimization G of total production expenses and economic damage by water pollution with due regard for wastes treatment and effect of realization of sideline products from waste re-utilization determined with probability a0. We have min G
PoJlEEE 7iM*St + E E E -&H& - E s («)^+ [
fc=l1=19=1
i-=l i = l p = l
, =1
+ E E E * » 0 * ] < G) = a0, a0 > o, 5, fc=] 1 = 1 71=1
J
where 7,'jt(w) — expenses on production of products i by the technological method p at the enterprise located at the point k; liki1^) — expenses on treatment of wastewater from production of products i by the technological method p at the enterprise located at the point k; /^-selling unit price for products j ; ^{u)j3fk— the function of economic losses due to disposal of pollutants 7r by the enterprise which is located at the point k and produces products i] ao-the probability that the total production expenses and economic losses caused by pollution will exceed the threshold value (G). Thus, we have the regional development planning model which takes into account pollution of river water and is represented as a stochastic programming problem with an objective probability function and chance constraints. We shall make several assumptions. Let all the components of the system of con straints and those of the objective function of the problem be independent normally distributed random variables (written as the argument u). Let atJ(uj) £ N(al:l, afj) denote the fact that ax3 is a normally distributed random variable with the mean value a:] and variance crfy For the complete model we have
Bdu) e N(B„a?(B)y,aUu)
e
N(al,afk(a));
CHAPTER 2. CHANCE-CONSTRAINED SPP
36
bW w) € # $ ,„^(^));^(^)G/V(^4) dL(">) £ N(dl, ,ff£,(d));Z?„MeW(2y, „^P));
v&*) e
N{Vqkl,erJ,-(V));7t(w)eJV(F» ,«*00);
&{<»:i 6 /V(F S \, ^ ( / ) ) ; W e i V ( F l t ,<&(F))i
«Mu 0 e #($* , 4 ( * ) ) ; < ? f t H e ^ - ,4(9)); <4(w ) € /v(c:„ ^ W J i n i i M e ^ , <*:*(«)); Vi(w) £N{Vk,>^(K));V;n(W)GyV(F,„, ^L(V'));
eN(vl,, rJ(V)); Vw(w) G JV(FW,fffc»(V));
^ »
*W(w) € JV(F«, ^ ) ) ; ^ M £ ^ „ ^ ( $ ) ) i **(« ) G N ( ^ , ^(F*));*fcHeiv(i;, ^ ( * * ) ) ; 7,\(* 0 e #(*&■,^(7));7f,HeiV(7^ ^ ( 7 ) ) ; K.
,(«) e /v(/ s , ^ ) ) ; i ' e ^ , <
!
(*));
Chapter 3 Two-stage Stochastic Programming Problems 3.1
Statement of a Two-stage Stochastic Programming Problem
Many problems in planning and management under uncertainty are treated and solved as two-stage stochastic programming problems. Stochastic problems with compensa tion for residuals in constraints are common in applications and receive more attention in the literature than any other stochastic programming problems. The solution of a two-stage problem is made up of deterministic and random vectors. In solving the problem, one has to select at the first stage a preliminary deterministic plan which should be adopted prior to realization of random conditions for the problem. A ran dom vector in the solution corresponds to the compensation plan for residuals which generally appear at the second stage subsequent to observation of the parameters realized. The manager wants to minimize the mean value of total costs incurred at the preliminary planning stage and at the second stage where the residuals in the constraints of the problem are to be compensated for. If the stochastic programming problem in a two-stage setting is solvable, the existence of a random vector for the residual compensation plan shall be guaranteed by properly selecting a preliminary deterministic plan. Suppose we have the problem: (C,I)^min,
(3.1)
A°X = B°,
(3.2)
AX = B
(3.3)
X > 0,
(3.4)
where C = {c3}, j = l,n;
B={bi),i
= l,m;
B° = {6°}, k = 17^7; 37
38
CHAPTER
3.
TWO-STAGE
STOCHASTIC
A0 = \\alj\\, k = l . r n , ;
j - T~n;
PROGRAMMING
PROBLEMS
A = ||a„||, i = l,ro;
j = 1,". Let the elements of the matrix A = A(UJ) and those of t h e vector B = B(UJ) and C = C(uj) be random variables. Then it may be reasonable to solve problem (4.1)(4.4) in two stages: prior and subsequent to realization of r a n d o m p a r a m e t e r s in the conditions of the problem. At the first stage we select a non-negative deterministic plan a;0 satisfying conditions (3.2). At the second stage we fix the realization ui° of a. chance event and its related values A(ui°) and B(uP). C o m p u t e t h e residual B(UJ°) — A(UJ°)X° which generally occurs in conditions (3.3) after the realization u)° € ft. Define the vector Y > 0 compensating for t h e residuals accorfing to the relationship D{u°)Y(u°) = B{UJ°) - A{u°)X°, (3.5) where D = \\d,i\\,i = l,rn,l = \,n\ is the compensation m a t r i x which is generally composed of random elements. Here it is assumed t h a t the r a n d o m realization u> observed at t h e second stage is independent of the choice of the preliminary plan x Consider the following mathematical programming problem: Find the n-dimensional vector x and the rti-dimensional vectors y(u>), w (E SI, which ensure nun Eu { ( C H , I ) + TM*(H,Y{U>))}
(3.6)
subject to A°X A(u))X
+ D(UJ)Y(LO) X > 0.
(3.7)
= B°, = B(w),
u>en
>'(^}>0.
(3.8) (3.9)
Here H is t h e given penalty vector depending on t h e components of t h e residualcompensation vector Y(OJ). E indicates mathematical expectation (or the mean val ue). Having defined the preliminary plan A' 0 we select components of the vector Y(u>) in such a way as to ensure a m i n i m u m penalty for offsetting t h e residuals in the conditions of the problem. In other words, having fixed the vector A" 0 ), one needs to solve the problem m m (H, Y(w)) \D{u)Y(w)
= B(w) - <4(u;)A'u, Y(u>) > o |
(3.10)
Problem (3.10) must have plans, otherwise the vector Y{ui) satisfying conditions (3.8) does not necessarily exist for every u> 6 ft. Problem (3.6)—(3.9) will be reffered to as a two-stage stochastic programming problem, and problem (3.10) as a second-stage problem. It is worthwhile to use the models and methods of solving two-stage (multistage) stochastic programming problems for the purposes of long-term planning and oper ational management, because the planned and managing systems are always subject to random influences and the two-stage models show low sensitivity to changes in the
3.2. ANALYSIS
OF A TWO-STAGE
STOCHASTIC
PP
39
parameters of the conditions of the problem, i.e., they are stable in some sense. For the vector x to be a feasible plan at the first stage, it is necessary that for all u> € f2 there exists a vector Y > 0 such that D{UJ)Y(LO)
= B(LO)-A{LO)X.
(3.11)
The additional implicitly given constraints of the form (3.11) appearing in the secondstage problem will be reffered to as induced and the conditions of the form (3.7) imposed on the vector X will be reffered to as fixed. Suppose the set K\ = {X : A°X = B°,X > 0} is composed of fixed constraints and the set A 2 = {X
: VOJ £
ft,]V
> 0,A[OJ)X
= B{u>) -
D(u)Y(u)}
is defined
by induced constraints. The set K = A'i n A'2 then is the set of feasible vectors X for problem (3.6)—(3.9); if X 6 K, then the vector X satisfies the fixed constraints A°X = B°,X > 0 and moreover the second-stage problem (3.3) has plans. For the purposes of further discussion we need the following result. T h e o r e m 3 . 1 . The set K of vectors X in the two-stage stochastic programming problem is convex. Proof. We have A = h\ D A 2 , but the set A'i = {X : A°X = B°, X > 0} is convex. For the fixed ui £ ft we define the set K%» = {A'|3Y(u>) > 0} to be such that A(u>)X = B(u>) — D(U>)Y(UJ) which is convex. This means that A'2 = flwsn^w, and with it A' = A'i n A2, are convex sets as intersections of convex sets. It has been established in [65] that with the random vector B[LO) and the deterministic matrix of conditions for the stochastic programming problem A the set A"2 and the set K are convex polyhedral sets.
3.2
Analysis of a Two-stage Stochastic Programming Problem
The set A of preliminary plans for a two-stage stochastic programming problem is given implicitly. In the general case, it is not clear how to construct the set A'2 efficiently. For some special cases, which are crucial in applications, the set A2 coincides with Rn, i.e., there are no induced constrains. We may assume that the matrix D has rank m; otherwise (3.8) can be replaced by the equivalent relation, in which case p rows are of the form OY = B - AX, where 0 is the m-dimensional vector composed of zeros and P is the number of dependent rows in D. In this case, they can be incorporated into the fixed constraints (3.7). We assume that the m x n matrix D has rank m and the first m columns are linearly independent without loss of generality. A s s u m p t i o n 3 . 1 . For every v 6 Rm there exists Y > 0 such that DY = v.
(3.12)
Lemma 3.1. [34] If Assumption 1 holds, then D has at least m + 1 coiumns, i. e.,
n > m + 1.
40
CHAPTER 3. TWO-STAGE
STOCHASTIC
PROGRAMMING
PROBLEMS
T h e o r e m 3.2. [34] For the system of equations DY — v to have a nonnegative solution for W G Rm, it is sufficient that there exists a nonnegative solution to the homogeneous system of linear equations (3.13)
DIT = 0,
such that 7Tj > 0 for j = l,m. Proof. The system of equations DY = v always has a solution. Suppose the solution is such that the first Y3■ ^ 0 for j = l , m and the rest are equal to zero, for if the relation a{D-n) + DY = v holds for any a, then taking a great enough yields a nonegative solution to system (3.12). Condition (3.13), with the exception of some special cases, is difficult to verify. Let n = m + 1, then the sufficient condition is J^jLV ^JD3 = 0. f° r if 7rm+i = 0, then the first m columns of D3 axe linearly independent, which contradicts the fact that the matrix D has rank m. Hence 7rm+i > 0, and so we have 771
TT
J -£>„+, = n - ^V" T , -~-= - -r - i - ^171+1 — 2 _i/r ^,-rJ- ^ "n-J ,' * ^J
;=1
(3.14)
Tm ^ m++1l
where 7Tj = -»—2—. This system of linear equations has a unique solution, i. e., if it is positive, then K2 = Rn The conditions of Theorem 3.2 are not only sufficient, but also necessary for the system DY — v to have a nonegative solution for any v G /? = flm But for A2 = Rn, it suffices to require nonegative solvability of DY' — u not for all 1/ £ flm, but only for v = B — AX with all A' £ A"i, and all w € fl, Such f do not necessarily run through all Rm Problem (3.1 )-(3.4) can be interpreted in terms of production planning, where A is the matrix of basic technological methods and D is the matrix of emergency methods determining alternative compensation for residuals in the system conditions. In this case, the conditions of Theorem 3.1 can be interpreted as follows. For any residual v € Rm to have a feasible compensation Y, it is sufficient that the emergency methods represent a "closed system", i.e., all the products (or services) produced by some production methods are consumed by the other methods. Purchase and sale of separate goods may serve as an example of such a system. Theorem 3.3. For problem (3.10) to have a finite solution, it is necessary and sufficient that the system of inequalities ZD
(3.15)
be solvable. The proof obviously follows from the dual LP theorem [14]. Indeed, if problem (3.10) is solvable and has a finite optimal solution, then its dual is also solvable, and vice versa. The constraints for the dual of problem (3.10) are conditions (3.15).
3.2. ANALYSIS
OF A TWO-STAGE
STOCHASTIC
41
PP
The conditions of Theorem 3.3 are economically meaningful.If the expenses con nected with the use of emergency technological methods in eliminating divergences are to be finite, it is necessary and sufficient that there be a system of estimates Z for products produced by emergency technological methods such that the value of output in these estimates (with products produced by the i-th technological method which is used at unit intensity) does not exceed the expenses associated with the use of this method at unit intensity. T h e o r e m 3.4. [33]. Suppose matrix D has m + 1 columns and satisfies conditions of Theorem 3.1., i.e., m
3=1
Then for the condition of Theorem 3 to hold, it is necessary and sufficient that m
j = l,m.
Y^Kjhj + hm+i >0, nj > 0,
(3.16)
3= 1
Proof. Necessity. Suppose the second-stage problem (3.10) is solvable, then the set of feasible solutions to its dual is nonempty. Suppose vector ZQ satisfies conditions (3.15) of the dual problem, that is, Z0Dj < hJt j = l,m-|-7.
(3.17)
Hence for 7r_, > 0 m
m
m
£ n}Z0D} < £ iTjhy, 3=1
Z0Dm+i
= - £
J=l
Zolr D
i i
(3'18)
3= 1
Furthermore, we have m
Z0Dm+i
=-*rZ0Tr3D,
< hm+1.
(3.19)
3=1
From conditions (3.18) and (3.19) we obtain the required result (3.16). Sufficiency. Suppose (3.16) holds and the objective function of the second-stage problem (3.10) is not bounded on the set of feasible solutions to the dual of the second-stage problem {Z\ZD
implies that the system
= l,2,...,m
(3.21)
has a unique solution Z0. By (3.20), we have Z0Dm+1>hm+1.
(3.22)
42
CHAPTER 3. TWO-STAGE
STOCHASTIC
PROGRAMMING
PROBLEMS
From the conditions of the theorem, and from relations (3.21), (3.22), we have rn
m
Z0Dm+1
= - J2nJhi
= -YJZQTTJDJ 3=1
> 'Wi
3=1
which contradicts the condition (3.16). This completes the proof of the theorem. Condition (3.16) is economically meaningful in terms of the planning problem. Suppose that technological methods form a closed system, then the expenses that are associated with the use of emergency production methods and are intended to compensate for residuals are finite if it is impossible to draw profit from an idle mode, i.e., when relation (3.14) holds. The work [34] shows that a similar condition of the absence of profit also holds for the cases where n > m + 1. But this condition then is only necessary. We shall now consider a deterministic equivalent to thw two-stage stochastic pro gramming problem (3.6)—(3.9) and show that it is a convex programming problem. The dual of the second-stage problem (3.10) is {Z,B-AX)->
max.
(3.23) (3.24)
ZD < H.
Suppose a. solution to problem (3.10) exists and is finite; then there exists a finite solution to problem (3.23)-(3.24) and optimal values of both functionals coincide [14]. Denote the functional value by <j>. One may state that <j>(X, A, B) is convex in X. Indeed, let Z(X, A,B) denote the point at which the maximum (3.23) is attained under conditions (3.24) for the fixed X, A, B. For any A'j and X2, for which the extreme value of the objective function (3.23) is finite, we have: Z (aXi + (1 - a)X2, A,B){B= ZiaXi
+(1
+(1 - a)(B - AX2)} < +(1
A{aX^ + (1 - a)X2) =
-a)X2,A,B)[a{B-AX1)+ aZ(XuA,B)(B-AX1)+
-a)Z{X2,A,B)(B-AX2).
Letbe the objective functional of the equivalent deterministic problem. Then <j> is a. convex function, since the non-negative combination of convex functions is a convex function. The convexity of the objective functional (f> implies its continuity in all interior points of the convex set A'. We have thus proved the following assertion. T h e o r e m 3.5. The deterministic equivalent to the two-stage stochastic program ming problem (3.6)-(3.9) is a convex programming problem. The last assertion serves as a theoretical basis for construction of numerical meth ods of solving two-stage problems. To consider the methods of solving a two-stage problem, we need to use the expression for the support functional to <j>(X) and es tablish a differentiability condition for cj>(X). Here the support functional to the
3.2. ANALYSIS
OF A TWO-STAGE
STOCHASTIC
13
PP
convex functional L, F(-q) at the point n0 <E M is taken to be the linear Functional if F(n) - F(rio) >{L,rj170) for all t] <E M. We shall provide assertions following [33] and [34]. Theorem 3.6. The functional E{C-Z'(A,B,X0)A}
=
= /„ {C{u) - Z* [A(w), B{LS),X0]
A(U)} dp
provides support to the objective functional of the equivalent deterministic problem at the point X0 S k. As is shown in [34], the objective function <j>{X) of the equivalent deterministic problem is everywhere continuously differentiable on the set K if the probability measure in the space A, B is absolutely continuous with respect to the Lebesque measure in the space A, B and certain conditions are satisfied. To examine the optimality conditions for the feasible solution A' to the first-stage problem, we require the vector C'x = E[C — Z*(A,B,X)A] and the linear form LXl = {CXl,X) = E[C - Z*(A, B,X)A]X. Following [33], we formulate the neces sary optimality conditions for the deterministic plan X to the two-stage stochastic programming problem. Theorem 3.7. If X' is a deterministic plan for the two-stage problem, then for any X g K Lx(X') < LX(X). (3.25) Proof. Let A'* be an optimal plan, and X a feasible plan for the two-stage problem. Then we have 4>(X") < 4>{X) E(CX"
+ Z*(A, B,X*){B-
AX')) < E(CX + Z"[A,B,X){B AX')) > E(Z'(A,B,X)(B-
E{Z*{A,B,X*)(B-
- AX)) AX*)).
(3.26) (3.27)
Subtracting (3.27) from (3.26), and taking into account the fact that Z*(A, B,X*) is an optimal plan for the dual problem, we obtain the required result (3.25). Following [18] and [33], we shall present the economic meaning of condition (3.25). The vector Z*(A,B,C) provides a solution to the dual of the second-stage problem and is the vector of estimates for products that prove to be in short supply or in surplus at intensities X of technological methods after the technological matrix A and demand vector B have been realized. These estimates govern the influence that the residual value may bring to bear on the expenses associated with the most economical elimination of residuals. The value m
^atlZ;(A,B,X)-c} 1=1
indicates the margin of profit that can be drawn from using the j - t h technological method at unit intensity under the assumption that the parameters of the problem are realized as elements of matrix A and components of vectors B and C, while the
44
CHAPTER
3.
TWO-STAGE
STOCHASTIC
PROGRAMMING
PROBLEMS
estimates of products are computed for the case where technological m e t h o d s are used at intensity A'.If t h e vector A'* determines an optimal preliminary solution to t h e twostage problem, then the total average probit t h a t is drawn from using technological methods at intensities A'* and computed in optimal (corresponding to A*) estimates is not smaller t h a n the total average profit computed in optimal estimates for any other feasible plan X. We shall formulate without proof the theorem of necessary and sufficient optimality conditions for a feasible solution to t h e two-stage stochastic programming problem. T h e o r e m 3 . 8 . Let X* denote an interior point of the set K and let the objective function 4>(X) of the deterministic problem that is equivalent to the two-stage problem be differentiable in the neighbourhood of the point A* The dual of the second-stage problem then has a solution Z*(A,B,X) such that Cx. if and only if X* is a solution
3.3
= E[C-Z*(A,B,X*)A]
= 0
to the two-stage
[33].
problem
(3.28)
Some Partial Models of Two-stage Stochastic Programming Problems
This section focuses on partial statements of a two-stage stochastic programming problem t h a t are widely met in applications. Here only components of the constraint vector B(oj) are random, while the other parameters in t h e conditions of the problem are deterministic. Suppose the matrix of emergency technological m e t h o d s has the form D = (I, —I). Problem (3.6)—(3.9) with such a matrix is called t h e simplest s t a t e m e n t of the twostage stochastic programming problem. Here the presence of a residual in t h e condi tions of the problem is taken into account by t h e simplest m e t h o d , i. e. by charging a penalty for underproduction or overproduction. Specifically, the penalty can be equal to zero. The overproduction penalty vector H+ (an m-dimensional vector) may take account of additional expenses on storage of finished products t h a t have not been sold and are still kept in warehouses. T h e underproduction p e n a l t y vector H~ (an m-dimensional vector) may take account of profits t h a t have not been completely collected because of the unsatisfied demand for t h e products. This s t a t e m e n t of the two-stage problem is discussed in many works, including [17], [32], [63]—[65]. The constructive results obtained for the simplest s t a t e m e n t make it possible to find an exact or approximate solution to the problem. Suppose the matrix D has the form (/, - / ) , where / is an m x m unit matrix. We shall divide t h e vectors Y and H into two parts corresponding to submatrices / and —/. Problem (3.6)—(3.9) then becomes (C,X)
+ E [ ( # + , Y+) + (H-,Y-)]
-> min
(3.29)
3.3. SOME PARTIAL MODELS OF TWO-STAGE
45
SPP
subject to A°X = B°, - IY- = B,
(3.31)
+
> 0, Y~ > 0,
(3.32)
AX + IY X > 0, Y The second-stage problem becomes [(H+,Y+)
(3.30)
+
-►min,
+ (H~,Y-)]
(3.33)
/ V + - IY~ = B - AX, Y+ > 0, Y~ > 0.
(3.34)
and has plans for any right-hand part B - AX. To make problem (3.33)-(3.34) solvable, it is necessary and sufficient that H+ + H~ > 0. We assume that H+ + H~ > 0 and exclude from consideration problems (3.29)—(3.32) for which(X) = -co for X € K. Evidently, there exist the induced constraints for the equivalent deterministic problem to problem (3.29)-(3.32). We shall now consider the dual of problem (3.33)—(3.34) m
m a x ^ ( 6 , - (A„X)) Zt\ - H~ < Z < H+
(3.35)
The optimal plan for problem (3.35) becomes Zi{{AuX)M)
= -K,
Zi{(Ai,X),bi)
= ht,
^
6s-(A,,A')<0,
if
bi-(Ai,X)>0,,
where ((A,, A'), 6,) is an arbitrary number on the interval [—h~, h*\ if &; — (A,-, X) = 0. Introducing the (A,, A)-convex functions f ((A„ A), 6.) = {min(/i+y+ + Ky-)\yf Vt > 0,yr > 0} = Z,((At,X),bt)(bt
- y~ = b, - (A„ A); - (A„A));
i = l,m
we reduce problem (3.29)-(3.32) to the separable programming problem n
m
^ C J X J + ^ v ' , ( A „ A ) ^ min, j=i
(3.36)
.=i
subject to A°X = B°, AA - ( A , A) = 0, A > 0.
(3.37)
When vector B has finitely many realizations, the simplest two-stage problem can be represented as min|(C,A) + ^ p r i / r ( w r ) J ,
(3.38)
46
CHAPTER 3. TWO-STAGE
STOCHASTIC
PROGRAMMING
where X and Y(uT) satisfy the constraints P{B(UJ) that AX+
= B(ur)}
DY{ux)
PROBLEMS
= pr,r = 1,/V, such
= B(u>i) }
AX+
DY{UJ2)
= B(w2)
AX+
DY(UJN)
= B(OJN)
(3.39) ,
x > o, ,y(w r )>o, r = T77v.
(3.40)
The corresponding dual problem is max {B(LOX)TZ1
ATZl
+
ATZ2
+ B(UJ2)TZ2
+ ... + B(uN)T
+
+
ATZN
(3.41)
ZN)
<
C,
<
P2H,
<
PNH.
T l
DZ
DTZ2 DTZN
(3.42)
If the variables of the dual problem are taken to be 2? Zr = —; pr
r = l,7V,
we have the problem max { f l B h ) T 2 ' + P2B(OJ2)TZ2
P1ATZl
+
P2ATZ2
+
+ ... +
+
PNATZN
T l
DZ
T 2
DZ
DlZTyN
PNB(uN)TZN}
(3.43)
< <
H,
<
H,
<
H.
(3.44)
The matrix structure of problem (3.43)-(3.44) makes it possible to solve this prob lem by the method of decomposition using the first constraints in (3.44) as "engaging" constraints and considering all others as a single unit. Suppose the vector B(u>) is discretely distributed with finitely many realizations such that the components 6, of the vector B(ui) may assume values b\ < b2 < . . . < bk'
3.4. THE TWO-STAGE
NONLINEAR
SPP
17
with probabilities p] , pf,. . . , p,', respestively. Introducing additional variables and constraints may then reduce the convex piecewise linear problem to the linear pro gramming problem. If the components of vector B(LO) are uniformly distributed over the interval [u,-, t>,], i. e., 0 ,bi <m hi — Ui
4>,(b,)
j. ,
r
i
(3.45)
, bi > v,
then the deterministic equivalent to the two-stage problem (3.29)-(3.32) becomes the quadratic programming problem. Problem (3.29)—(3.32) may also be reduced to the quadratic programming problem when components bt of vector B(u>) are all distributed exponentially, i.e., 0,
w.
bt < u,
l_e-Mk.-«0,
b,€[ui,oo]
(3.46)
The two-stage stochastic programming problem in which all random components of vector B(u>) have a continuous distribution function may be reduced to a quadratic programming problem. To do this, it suffices of replace random variables by weighted sums of uniformly distributed random variables
EM; f 0,
^M) =
6
v
(3.47)
K< <
^ £ , 6feK,vr].
I 1,
3.4
X>I = i
(3.48)
K > vTt
The Two-stage Nonlinear Stochastic Programming Problem
The two-stage nonlinear stochastic programming problem is considered to be a gener alization of the linear case in [50] and a special form in [58], where a special algorithm is provided. Let us consider the problem m i n S m i n h ^ . Y ) + ip{Y)]
(3.49)
subject to
g(X) +
h(Y)>b,
(3.50)
48
CHAPTER
3.
TWO-STAGE
STOCHASTIC
PROGRAMMING
PROBLEMS
where(Y) is a scalar function of n 2 -dimensional vectors Y, and g(X), h(Y) are vectors of demensionality m, whose components are scalar functions of their arguments, and b is a r a n d o m m-dimensional vector with known distribution. E designates expectation for distribution of vector 6. In the literature this setting of problem (3.49)-(3.50) is called a "here and now" situation according to Dantzig. At first, one takes a decision on X. Next, one waits and observes how random variables are realized. And finally one solves the nonstochastic problem by optimally selecting the vector Y Any further discussion will concern itself with the two-stage minimization problem (3.49)-(3.50), although analogous results can also be obtained under proper conditions for formulating a two-stage maximization problem. For convenience, denote 7(6, X) = n u n { y > ( * ) + t/>[Y)\g(X)
+ h{Y)
> b} .
(3.51)
Problem (3.49)-(3.50) then becomes (3.52)
min Ey(b,X).
Since problem (3.52) has a restricted m i n i m u m in the space X we assume t h a t there exists a convex set K such t h a t for every X £ K there is Y such t h a t g(X)+h(Y) > b for any vector b. Such X are called admissible. Define X(Eb) as a solution to the problem mm-y{Eb>X)=mm{
+ il>{Y)\g(X)
+ h{Y)>
Eb)
(3.53)
Moreover, we introduce the problem for all finitely distributed b < 6 m a x min 7 (fe m a x ,A') = m i n m j n { 9 ( A ' ) + j>{Y)\g(X)
+ h(Y)
> 6max}
(3.54)
Problems (3.53), (3.54) are called nonstochastic and can be solved by employing Kuhn-Tucker conditions [43] if the functions g(X) and h(Y) are concave. Problem (3.49)—(3.50) can also be solved directly for the partial linear case where il>(Y) and h(Y) are linear functions and vector b has a finite discret distribution 7r;, i = 1,1. In this case we may introduce vectors Y', i = 1,1 each of which has the some dimension as Y Problem (3.49)-(3.50) then becomes a deterministic problem in n.j + ln 2 unknowns and m x / constraints:
mm L(X) +J2*rf{Y,)\9{X) + *(>'*) > K ' = U } For t h e purposes of further discussion, we need to consider some results presented in the work [50]. Suppose we have the parametric nonlinear programming problem mm@(Z,a)
(3.55)
3.4. THE TWO-STAGE
NONLINEAR
SPP
49
subject to (3.56)
f(Z,a)>0,
where 0 is a scalar function of the vectors Z and a and / is a vector function of Z and a. Lemma 3.2.The scalar function a(a) = mmz {@(Z,a)\f(Z, a) > 0} is convex and continuous in a ifQ is a convex continuous function of[Z,a] and every component of function f is a concave continuous function of [Z, a]. A more general formulation of Lemma 3.2 is given in [43] for a two-stage nonlinear stochastic programming problem. Lemma 3.3.Let if : Rn -4 R and 4< : Rm -» R be convex functions. Suppose g and h are vector functions, g : Rn —>■ Rk and h Rm —>■ Rk, whose components are concave real-valued functions. If the functions ip, i\g and h take only finite values on the set Bh = {(X,Y)\g(X) + h(Y)>b}, then ct(b) = 'mi{x,Y)£BbW{X) + ^(Y)} where
is a convex continuous real function on W,
W = {b\be Rk,b< +
bsi = supX£Rnmij=Tjg]{X)
bSI}, supYeRmmiJ=Tjhj(Y)
The above assertions make possible the proof of the theorem below. Theorem 3.9. The sealar function a{b) = mm^(b,X)
= mmm\n{ip{X)
+ '4'{Y)\g{X) + h{Y) > b}
(3.57)
is a convex nonincreasing function in b if ip(X) and xj>(Y) are concave functions in their arguments. Corollary 3.1. The scalar function a(b) = min^ 7(6, .Y) defined in (3.57) is a continuous function in b if ip(Y) and h(Y) are convex continuous functions in their arguments. This corollary is proved by employing Theorem 3.9. The above assertions employ the convexity of functions in a random vector. We shall now consider the theorem which ensures the convexity of function 7(6, A') in admissible X. T h e o r e m 3.10. The function 7(6, A') = mm {V(X)
+ 4>(Y)\g(X) + h(Y) > b}
is a convex function in admissible X for any fixed b if ip and (/' are convex functions and components g and h are concave functions in their arguments. The proof is carried out by introducing proper notation and by employing Lemma 3.1 and Theorem 3.1. Corollary 3.2. The function Ef(b,X) is a convex function in admissible X if the conditions of Theorem 3.2 are satisfied.
50
CHAPTER 3. TWO-STAGE
STOCHASTIC
PROGRAMMING
PROBLEMS
These general results point the way to some theorems capable of giving upper and lower bounds for problem (3.49)-(3.50). We shall now introduce the "wait and see" problem of Madansky which differs from the formulation proposed by Dantzig in that at first one observes how a random vector b is realized and then solves the nonstochastic programming problem based on this realization, i.e., the problem Emm
= Ermnndn{tp(X)
X
X
+ 4>(Y)\g{X) = h(Y) > b}
T h e o r e m 3.11. Let X{Eb) be a solution to the problem minx 7(Eb,X). we have E-y (b,J((Eb))
(3.58)
Y
> mmE~f(b,X)
> Emmy{b,X)
>
Then
mm(Eb,X).
For the last inequality, it is assumed that(Y) are convex continuous functions, and components g(X) and h(X) are concave continuous functions in their arguments. Proof. The first of the inequalities is obvious, since minx £7(6, X) does not exceed the value of the function E-y(b, X) computed at some point X = X(E, b). If X is the solution minimizing £7(6, X) and if X(b) is the solution minimizing 7(6, X), then for the second inequality we have mm£ 7 (4,.V) = £ 7 ( 6 , X ) , Emm-y(b,X)
= E-y (b,~X{b))
Since the inequality 7 ( b , X ) > 7 (6.7(b)) holds for each 6, we have Ef(b,X) > E-y lb, X(b)j or minx E~f(b, X) > Emirix -y(b,X). by the properties of mathematical expectation. To prove the third inequality, one needs to employ the Jensen inequality which implies that EG(X) > G{EX) holds for the convex continuous function G depending on a random vector X. By Theorem 3.1 and Corollary 3.1, minx 7(6, X) is a convex continuous function in b, and hence E minx l{b, X) > minx y(Eb, X). This completes the proof of Theorem 3.11. An important result of this theorem is that the values of the objective functions in problems (3.49)-(3.50) and (3.58) can be approximately bounded below by solving the only nonstochastic problem, i.e., problem (3.53). If the vector b has a finite distribution, one may obtain some approximate upper bound for problem (3.49)(3.50), and hence for problem (3.58), by solving the only deterministic problem (3.54). T h e o r e m 3.12. If a random vector b has a finite distribution, i. e., —00 < bm\n < b < &max < +00, then min7(6 max ,A') > mmBf(b,X). Proof. By definition, we have: min7(& max ,A) = minmin {y{X) + i>(Y)\g(X) + h{Y) > 6 max } , A
A
1
3.4. THE TWO-STAGE
NONLINEAR
mm E mm {V(X)
51
SPP
+ 4<(Y)\g(X) + h(Y) > 6}
For the fixed X and 6, each value of V which satisfies g(X) + h(Y) > &max also satisfies g[X) + h(Y) > b. From this it follows that 7(& max ,A) > 7(6,A') and 7(6 max ,A') > £7(6, X). Hence, minA- "/{bmax, A') > minA- Ej(b, A'). The following theorem provides a lower bound for the values of objective functions in problem (3.1)-(3.2). T h e o r e m 3.13. // \p and ip are convex, components g and h are concave func tions, and £7(6, A") is differentiable in the point X = X{Eb), then min^ E~/(b, X) > £ 7 (b,X~{Eb)) + [X -Y(Eb)]T VEi (b,X~(Eb)) , where X~(Eb) and X are the so lutions which minimize j(Eb,X) and $7(6, A), respectively; and V is the column vector of partial differential operators
d
d
dx\' 32:2'
d \ ' dxnJ
The proof is carried out by employing Corollary 3.2. Since the solution A' of problem (3.4) appears in the estimation of the right-hand part of the inequality, this theorem serves practical purposes only if A' = X(Eb) or V.E7 (b, X(Eb)) = 0. Hence it follows (from Theorem 3.3 and 3.5) that E-j (b, X(Ebu = minA- £7(6, A). Referring to [50], we have a nontrivial example of the two-stage problem (3.1)(3.2) which provides numerical evidence in favour of the basic results above. The work [44] reformulates some assertions from [50], specifically, Theorem 3.3. Much thought is given to function spaces, and the values of upper bounds on functions are rejected. Generalized are conditions which yield specific results as special cases for stochastic linear and nonlinear programming. The work [58] examines the following problem: Find the vector Y 6 Rn which gives ™ n ( 5 ' ( r ) + X>P<(>''))
(3-59)
subject to f,(Y) < g,, 1 = l,m where S(Y) is a certain function of expenditure, Pi(Y) is a probability of non-fulfilment of the z-th condition in the system of canstraints and a, are involved expenditures, non-fulfilment of the condition f,(Y) < ,, is considered as a random event, the probability of each event G depending only on Y The problem 3.59 may be written otherwise: find the vector Y S Rn which provides min f S ( y ) + f > , p < f Y ) ]
(3.60)
subject to fi(Y) + A; < gi, 1 = l , m where P,{Y') is a probability of nonfulfilment of the i-th constraint, a, are corresponding expenditures and A, are random values with distribution laws F,-(£)i i = l , m .
52
CHAPTER 3. TWO-STAGE
STOCHASTIC
PROGRAMMING
PROBLEMS
The linear problem (3.60) is solvable as a simple two=stage problem. In the non-linear case the work [58] shows how to reduce (3.60) to a deterministic convex programming problem. Let us formulate the following theorem. T h e o r e m 3.4. Let S(Y), fi(Y), f2{Y),..., fm{Y) be convex functions and each function Fj(£) be concave on a certain segment [a]} bj] having the number £* assuming that (* = g} — fj(Y*), but Y* is the optimal solution for the problem (3.60). Then so that the problem (3.60) might be substituted for the equivalent convex programming problem it is sufficient to insert additional variables U\,u2, . ■ ■ , « m upon which are imposed the conditions f u} < 1 - Fjiaj) Ui > 1 - Ffa)
I Vi>Pj(Y), where P}{Y) = 1 - F}(rj(Y)), problem is of the kind
,
(3.61)
j=T^m
r3{Y) = g3 - f,(Y).
The objective function m the
m
S{Y)+YJaiUi
(3.62)
1=1
and the optimal vector U* = (Pl{Y*)1P2(Y*),. . .,P m (V"*)). The problem with m -f- I unknowns is solvable by convex programming methods, since according to the conditions of the theorem, the objective function (3.62) is convex near the optimal solution Y* But there appear difficulties due to 'ravening' of the problem inasmuch as at the crossing of a certain bound in the space Rn expenditures YALI aiPi(Y) start to increase fast due to the increase in probability of non-fulfilment of the z-th condition. The work [58] gives the example of a practical search for a solution, the number of variables being increased. In the wirks [3-6] there are generalized models for two-stage stochastic program ming problems in such a way that they include problems of stochastic optimal control over behaviour of dynamic systems. Consideration is given to the systems whose con trolled behaviour may be described by simple differential equations. The initial state of the system, law of movement, objective functionsl an constraints on phase coor dinates depend on random parameters. A preliminary decision how to control the system is made before observation of how the random parameters of the condition of the problem are realized. Divergencies in the constraints of the problem appear ing after mading the preliminary dicision are compensated by corrections. For the optimal selection of corrections at each realization of random parameters of the con ditions one should solve a new problem of optimal managenent — that is, the problem of the second stage. The mathematical expectation for the values of the lower side of the objective functional in the second stage problem, just as in the classical twostage stochastic programming problem, defines a penalty for correction which enters as a component into the objective runbtionsl of an infinitely-dimensional two-stage problem.
3.5. METHODS
3.5
OF THE SOLUTION.
EXAMPLES
53
Methods for the Solution of Two-stage Stochastic Programming Problems: Examples
A number of two-stage stochastic programming problems at the stage of choosing a preliminary plan are substituted for by deterministic equivalents, that is: convex programming problems. But traditionsl approaches to the solution of convex pro gramming problems are, as a rule, inadequate for the solution of two-stage stachastic programming problems, since the objective function and the set of feasible solutions for a problem in the general case are given implicitly. The vethods of solution which are considered in this paragraoh use the specific properties of the equivalent deter ministic problem and make it possble to compute a preliminary plan and obtain approximate estimations of a solution for the initial problem. In partial cases one can reduce a two-stage problem to a linear or piecewise-linear problem. It is natural that in this case there is no necessity to use the lower (given labour) consumming methods of solution. The works [24-27] investigated the generalized stochastic gradients method which makes it possible, according to the realization of random parameters in the conditions of the problem, to gradually specify the optimal first-stage plan. Following [24] we consider the principle of the generalized stochastic gradients method. Let the following be given: (1) A system of linear inequalities AX
(3.63)
+ DY>B(UJ),
X > 0,
Y > 0,
(3.64)
where A is the matrix m. xni, D is the matrix m x n 2 ; A', Y are accordingly ni and n 2 dimensional vectors, B(ui) — an m-dimentional random vector with distribution dP(oj) with constrained dispersion; (2) Two linear forms: LX(X) = ( C \ A ' ) ; L2(X) = (C2,Y) where C, is the rii-dimensionsl vector and C 2 the n2-dimensional vector. Denote as $ ( X , B(w)) the value min(C* 2 ,y)
(3.65)
DY > B(w) - AX,
(3.66)
Y > 0.
(3.67)
under the conditions:
Assume that for any X > 0 and B(u) there exists a solution of the problem (3.65)(3.67). Let E(X) = f$(x,B(u>))dP(uj). It is required to find mm F(X) = min \(C\X) x=o =
+ E(X)}
(3.68)
54
CHAPTER 3. TWO-STAGE
STOCHASTIC
PROGRAMMING
PROBLEMS
The function F(X) is convex but not necessarily continuously-differentiable. Find the expression for the internal normal of basic hyperplane to the body {X : F(X) < F(X0)} in the certain point X° Consider the dual problem to the problem of the second stage (3.65)—(3.67): find G{X,B{LO))
(3.69)
= max[(X,B(w))-(X,A*X)]
under constraints D*X < C\
(3.70)
A > 0.
(3.71)
From the fuality theorem it follows that: G(X,B(w))
(3.72)
= $[X,B(w)).
The function G(X,B(u>)) is convex at X. Let X(X,B(w)) be the point at which is attained the maximum in (3.69). Assume the domain R truncated by constraints (3.70)—(3.71) to be limited, then one may choose X(X,B(ui)) as coninciding with one of the vertices R. In the case of the non-simplicity solution of the problem (3.69)—(3.71), X{X, B(u)) may be selected, for instance, in the following way: conditionally enumer ate the vertices R and srlrct X(X, B(u>)) to be equal to the apex with the least number of vertices which is the solution of the problem. Upon such a definition, X(X, B(u)) represents pieciwise the constant vector-function from X and B(u). G(X,B[u>)) is the piecewise-linear function from X measurable by the lebeshue-Stieltjes measure dP(u), Fix a certain point X = X° Then the relations: G{X,B{UJ))
> [X[X°,B(UJ)),B(CO))
- (x,A'X(x°,B{u)))}
,
G (X°, B{vj) = [(A (X°, B(co)) , B(w)) - (X°, A'X (x°, B( w )))] hold. Hence it follows that E(X) > f (X (X°, B(w)) , B(uj) d P ( w ) - / (X, A'X (X°, S(w))) dP(ui), B(X°) = I (X (X°, B(UJ)) , B{u>)) d P ( w ) - f (X°, A'X (X°, B(u>))) dP(co). E(X) - E(X°) >-j
(X - X°,A'X
(X°,B{L>)))
dP{u).
(3.73)
In according with the inequality (3.73), the mathematical expectation of the random vector C 1 - A*X(X°, B(LO)) is the unknown vector to the internal normal of the basec hyperlane to the body [X : F(X) < F{X0)}.
3.5. METHODS
OF THE SOLUTION.
EXAMPLES
55
Thus, calculation of the generalized stochastic gradient for the function F(X) is reduced to a calculation of the integrals (3.73) that with known dP(to) is reduced to the solution of the parametric linear programming problem. In the partial case, if B(ui) takes a finite number of values we get the algorithm described in [24]. However the solution of the parametric problem involves many difficulties. Besides, the vector B(LO) is often obtained by way of random samples and the distribution dP(ui) is not known a priori. Therefore, for the generalized case a rendom search algorithm of the following dind is proposed. Let the value X' 5 ' be obtained at the step S. Then the 5' + 1-th strp will be specified as follows: (a) select a random realization B(u)^ in accordance with dP(u>); (b) find A(XW, B(UJ)^I) solving the problem (3.69)-(3.71) with X = X<s\ B(co) = (c) find X{s+1) = m a x { 0 , X ( s ) + ps [C1 - A*\ [X{s), B ( w ) w ) ] } ,
(3.74)
where ps is the value of step. Movement of the point X ' s ' then goes on in a random direction whose mathematical expectation converges with the geberalized gradient descent direction. The essence of the generalized stochastic gradients' method is in the following. Let it be required to minimize the convex function F(X) = F(x\,... ,xn) from below. Consider the iterative process .Y|S+1)=IW+^W,
5 = 0,1,...,
(3.75)
s
where X^ is a random point, ps is the value of the step, f' ' is a random direction such that E{({s)\X^,X{2\...,Xis)} = Fx(X{s)), (3.76) where F(X) satisfies the inequality F(Y) - F(X) > (FX(X), Y-X)
(3.77)
Let mmF(X) = F(X*) > - c o . Without restriciton of the generality consider that the point X" is the only one. Theorem 3.15. Let OO
P. > 0,
£]/9 s = oo,
CO
^p^
E(\\(M\\2/XU,XW,...,XW)
(i) (ii)
Then \ixa\\X* - X^\\ = 0, s -» oo with a probability of 1. Proof of the assertion is given in the work [28]. Demonstrate that for the process (3.74) the conditions of the Theorem (3.15) are fulfilled. Actually from the assumption
56
CHAPTER 3. TWO-STAGE
STOCHASTIC
PROGRAMMING
PROBLEMS
that the domain R truncated by inequalities (3.70)-(3.71) is limited it follows that the random vector C'1 - A*\(X, B{uiU norm is limited. Thus, for convergency of the process (3.74) the conditions oo
Ps > 0,
oo
Y1PI<0°
^2ps = oo, 5=1
S—i
are sufficient. In other words, the iterative process of the generalozed stochaxtic gradients' min imizing convex from below function F(X) on the convex set K may be represented by the following relation X(*+U = W(X{.)
., = 0 , 1 , . . . ,
_ PtVlgA),
(3.78)
where A'' 0 ' is a random n-dimensional vector belonging to the set K; X*-0' is the initial point of the process; p is the value of the stop on s iteration; vs is the normalizing factor; £' s ' is a random vector whose conditional mathematical expectation according to A' 0 ', A' 1 ',. . ., A' s ' depends linearly on the generalized gradient Fx of the function F(X) in the point A(s>; E(PW>\X<1\...,XM)=a.Fx(xM)+mM.
(3.79)
Here as is some number; m^s'n — is a dimensional vector; W an operator-projector on the set K, is such that W(X) <E K and \\X - W(X)\\2 < \\Y - X\\2 for any Y £ K. The operator-projector thus is Y = W(X) and represents the solution of the convex programming problem with the quadratic objective function inf | | A ' - V | | 2
(3.80)
In some partial cases if the set K is explicitly given, the calculation of the operatorprojector W(X) and consequently the definition A' (s+1 ' can be essentailly simplified. Give several such cases following [25]: (a) if K = J A | A > 0}, the solution of the problem (3.80) is such: >"* = W(X) = max{0, A'} but the process (3.78) is written as: .Vls+I»=max{0,.V(s»-Ps^(s)};
(3.81)
(b) if K = \X\a,j < Xj < 6, j , the iterative process (3.78) may be represented as ( «„ 4+l)
=
*<S) - P»»4]> . h,
x]'] -
Pmu.^
«i < ^ { ]
x; -
< a,
- PM* < h> s)
Ps„4
> bj
(3-82)
3.5. METHODS
OF THE SOLUTION.
57
EXAMPLES
(c)if X={X\DX = g} = {X\(S'\X) = g,,i = l , m } , where vectors d^ , U .. . c((m) are linear independent rows of matrix then the process (3.78) be such:
, . . . ,
m
A<s+1> = W (A<s> - psus^)
= A<5> - ,,„.£<») - J^
\td\
(3.83)
.=1
(d) if A' = {X\g(X) < 0}, where g(X) is a convex differentiable funct ion, then the problem of calculation of the operator-projector will have as a solution: [
1 A',
' '
(3.84)
g(X) > 0,
and A' is calculated form the conditions X = X - A(dG(A)/dA'); A > 0; g(X) = 0. The iterative process (3.78) is of the kind: Y(s+1)
_ f *<■> -p*K&\
g(XW-p,i,M{M
< 0,
" I X,
(A'M - p,v£*
> 0,
(3.85)
and A' is calculated form the conditions
XM-p.u.p>=X
A>0;
+ \l&P-,
g(X) = Q;
(e) if A' is a convex polyhedron A" = {X\DX < d}, the calculation of the operatorprojector is reduced to a solution of the quadratic programming problem \\X -Y\\2
-^ mm, DY < d,
(3.86)
where the iterative process (3.78) pre-assigns at each step of the quadratic programming problem solution A ( s + 1 >= min
\\X{s)-psv£s)-Y\\2
(3.87)
DY=d
Consider the specialized method for the solution of convex programming problems taking into consideration the properties of the structure in deterministic equivalent to a two-stage stochastic programming problem. Let us have the problem (/> —> min
(3.88)
under the conditions: E(CX
+ Z*(A,B,X°)(BA 6 K,
AX)) < «/>,
(3.89) (3.90)
where K is a convex plyhefral set. The latter takes place if A'2 = Rn or if the matrix A is deterministic. In the cases when the set K is given explicitly, the problem (3.88)(3.90) represents a linear programming problem solvable by traditional methods. In
58
CHAPTER 3. TWO-STAGE
STOCHASTIC
PROGRAMMING
PROBLEMS
the general case the set K is convex and the present approach can be used only for the qualitative analysis of the problem. Let K be a constrained set and X ( 1 ) be the solution of the problem (3.88)-(3.90), then the second iteration the following problem is solved: V> —> min AX)] < >,
E [CX + Z*(A, B, X°)(BE[CX + Z*(A,B,X{1))(B
- AX)] < 0.
On K iteration the problem E [(CX + Z*(A, B, X^)(B
- AX)] < >,
i = 0,l,...,K-l,
X & K,
is soved where A''"' is the optimal plan on i iteration. In the work [27] it is demon strated that lirriA'->oo F(Xh) = min^sA' F(X) and \im.K-too Xh is not necessarily the optimal plan of the initial problem. Then consider the approaches to the solution of separate partial cases of the twostage problem. Consider first of all a complete two-stage stochastic programming problem with a random vector of constraints B such as (3.29)—(3.32). Let the second stage problem (3.33)—(3.34) be solvable, that is the condition
H+ + H- >0 be fulfilled. Insert two classes of distribution functions for the vector B =
(bl,b2,...,bmf
Definition 3.1. The function Fd = {F(0) = F(0d,... ,0m)}; Ft{(3,) = F( — c o , . . . / ? , . . . + oo) for any i = 1, m satisfies the following conditions: (a)
Fi((3i) = 0\—) \PiJ
1 - F,(Pt) =0 1-A \p,J
with
with
/?,->■-oo
& -» +oo
or
or
lim # F ( f t ) = 0; A-+-«
lim (3% (1 - FA/3,)) = 0;
0,-H-oo
(b) in any finite interval F;(/?t) has not more than a finite number of points of gap. condition (a) assumes the presence of first moments F,(Pi) for any i = l , m . The se cond condition (b) provides the possibility to integrate by parts the Riemann-Stieltjes integral according to the measure dF,(f3,), if the under-integral expression is contin uous. Definition 13.2. Class F° = {F((3)} satysfying the following conditions(a) F((3) 6 F\-
3.5. METHODS
OF THE SOLUTION.
EXAMPLES
59
(b) for any i there exist numbers yf, y] such that Fj(y°) =), Ft(y\) = 1. If F(f3) 6 F 1 and we introduce the definition x = (xi,---,Xm) = ( M i , X),. . ., (Am, X)) we may omit the constant item (H+,EB). Under these assumptions represents the problem (3.29)-(3.32) as ip{X,x)^mm (3.91) AX -
x
= 0,
X > 0.
(3.92)
The problem (3.91)—(3.92) is a convex programming problem, since the function ip(X,x) is linear on A' and convex on \ . In the work [68] the following criteria of solvability for the problem (3.91)—(3.92) are proved. Theorem 3.16. Let F(/3) e F1 and H+ + H~ then the problem (3.91 )-(3.92) has the following results in the solution: (1) is not solvable by force of the absence of constraints form above for the function ip(X,x) if and only if there is no vector «r satisfying the conditions TlA>C,
H+
<~H~;
(2) is solvable if and only if there exist a pair of vectors (7r°,x°) such that (a) Ft{x°) = _ ^ - ~ _ \ + for all i for which -h~ > hf; (b) n° is the optimal plan of the problem [min(7r,x°)|7r/l>C , ,/Y + < n > H~] (3) Function0 then the problem (3.91)-(3.92) is such that: (1) it is not solvable, since the values of the functions are not constrained from above ip(X,x) if and only if there is no solution for the system w A > C, H+ < ■K < —H~; (2) it is solvable if and only if the system IT A > C, H+ < w < -H~ has a solution. Theorems 3.16 and 3.17 make it possible to formulate some consequences. Consequence 3.1 If the problem (3.91)—(3.92) is solvable for some function F(P) e F° and the condition H+ + H~ > 0 is fulfilled then the problem is solv able with all F{/3) £ F° Consequence 3.2. If the problem (3.91-3.92) is finite for a certain function F{P) e F1 with given A, C, H~, H+, it is finite with all F(/3) G Fl Consequence 3.3. The problem (3.91)—(3.92) is solvable if and only if with F{P) e F1; H+ + H~ > 0 when the following conditions are fulfilled: (a) the solution of the system nA>C,H+ hf either F,(x) = 1 for great enough x or there exists a vector ir satisfying the conditions -ITA > C, H+ < n < —H~; n, < —h~;
60
CHAPTER
3.
TWO-STAGE
(c) for eacn i such that -h~ exists a vector ir satisfying the
STOCHASTIC
> hf, either F,(x) conditions
TTA>C,H+
PROGRAMMING
PROBLEMS
= 0 for small enough
\ or there
-H~,hf
< 7T,.
From duality theorems and Theorem 3.17, if the problem (3.91 )-(3.92) has a solution there follows the assertion: T h e o r e m 3 . 1 8 . So that the vector (A'°,x°) be the solution of the problem (3.91)—(3.92 it is necessary and sufficient that (1) X° > 0, AX° - x° = 0; (2) for a certain vector TT° satisfying the relations the conditions n°A > C were fulfilled: (a) x ° > 0 , if (7TQ,A)i = Ci;
(b)
Fi(x!)=
_\°-~_ fc {+, if-/»,->/»+
The above assertions allow us to solve two-stage stochastic programming problems of t h e t y p e (3.29)—(3.32) with a random vector of constraints B. Works [8], [10], [16], [35] are devoted to t h e investigation of two stage stochastic programming problems by the use of a generalized inverse m a t r i x . Mention should be made t h a t generalized inverse matrix D+ for the matrix D is called the solution for the system of equations DD+D = D, (3.93) D+DD+ + T
(DD ) +
(D Df
= D+,
(3.94)
+
(3.95)
= DD , +
(3.96)
= D D.
One may demonstrate t h a t such a system has the only solution for any matrix D. Besides, if the matrix D has dimensionality mxn, then t h e generalized inverse matrix D+ — (n x m) in the case if D quadratic nondegenerate m a t r i x D+ coincides with D"1 In general the case D+ = RxDRi where Rl, R2 are t h e solutions for equations RDDT = DT and DT DR = DT accordingly. It is assumed t h a t for the complete twostage problem (3.6)—(3.9) which with any B(u>), u> 6 fi, X £ A', system of conditions DY = B — AX, Y > 0 is non-conflicting may be found the vector n providing the existence of a non-negative solution fro the system DY = B — AX: Y=
D+{B-AX)
(3.97)
+ Pn>0.
Under the conditions (3.97) the problem (3.6)—(3.9) m a y be represented as Z(X)
= (C,X)
+ Emm(H,
D+(B
- AX)
+ P„) -» m m
(3.98)
under the conditions P , > D+(AX A°X
= B°,
- B), X>Q.
(3.99) (3.100)
3.6. APPLICATIONS
61
OF TWO-STAGE SPP
The second stage problem (8.10) in that case is reduced to the selection problem r\ satisfying the inequality P„ > D+(AX - B) and minimizing (H, D+{B'AX) + Pv), that is, to the problem [mm(ff, P„)|P„ > D+(AX
(3.101)
- B)]
Then one may write {min„(ff,P„)|P„ >D+(AX-B)} +
= (mix„(w, D (AX
=
(3.102)
- B)\LOP = HP, to > 0} ,
The problem (3.101) containing random values in constraints is substituted for the problem (3.102) in which the set of plans is deterministic. After some manipulations the problem (3.98)-(3.100) is reduced to this: Z(X) = ( C +E{ma,xu,(uJ,D+{AX
HD+A,X)+ - B))\u>P = HP, to > 0} -> min
(3.103)
given that A°X = B°,
X > 0.
(3.104) +
Objective functions differ from Z(X) and Z(X) on constant item E(H,D B). The problem (3.6)—(3.9) in the form (3.103)—(3.104), as a rule, is not more con structive form the viewpoint of computing than any other form of notation of initial problems. In the work [8] it is shown that sometimes the usage of a generalized inverse matrix makes it possible to obtain convenient deterministic equivalents to a two-stage stochastic programming problem.
3.6
Applications of Two-stage Stochastic Programming Problems: examples
Traffic problems with random data are investigated in the works [7], [9], [11], [19], [51], [56], [57], [67]. It is usually assumed that the demand b3 at the j point of consumption is a random value. Let the demand bj be distributed continuously with a density ofbj(u>) then excessive expenditures of the value kj(y3 — 6j(w)) appear (coefficients ra; and k3 are defined depending on the concrete contexy of the problem) which are connected with the strong of non-realized parts of production, with their transporation, or, losses, appearing during their realization at lower cost. If into the objective function of the classical transport problem one builds a mathematical expectation of these expenditures, the stochastic transport problem will be mn F(X) = min j 5 3 c , - ^ +Y.ki
j y3 -
6
iM)
i
Pj(bj)dbiJi-
62
CHAPTER 3. TWO-STAGE
STOCHASTIC
n
PROGRAMMING
PROBLEMS
POO
+ 5>J-/
(3 105
(&»-%)¥*(&;)<%
-
Jy
3=1
'
)
)
under constraints ^xtJ=a,,
(3.106)
i = ujl,m,
3=1 771
£ > < , - % = (),
i = «l,n.
(3.107)
1=1
If as
/■+Oo
^ ==
/ + CO 6 i (cj)v? J (6 j )d6 J 6j(w)v,-(&,-)d6j /
denote the expected demand at j point of consumption, then the objective function of the problem may be represented as:
(3-108)
F(x) = 52ciXii + £>,-(% - H)+ J=1
hi 71
n
j3=1 =i
/-CO roc
+ N) /^ J ( & » - -!/)) ip,(bj)4bj.
To solve the problem, Williams [67] has worked out an algorithm which is based on the decomposition method due to Dantzih in nonlinear programming. In the work [21], Elmaghraby considered the generalized traffic problem provided that demand is random and allocation is optimal. To get an optivality criterion for the solution of the problem Elmaghraby modified the Kuhn-Tucker conditions. Let us investigate an analogous model, that of stochastic traffic problems in which the volumes of produciton are random, or such models in which neither production volumes nor demand in the consumprion points may have been forecast beforehand. Analysis of these models is reduced to the solution of convex or linear programming problems depending on the contimuously or descretely distributed parameters of the conditions of the problem. We give here the problem of the determination of the volume of produciton as iy is formulated in the work [29]. Suppose that the resources of i{i = l,m) kind are available and they are spent in the quantity XtJ for the production per unit of j(j = l,n) product. The given value of i resource is a, units; and the demand b} for each of the products is a random value with destribution density ipj(bj). Let x3 be the output volume of j product. In the case where x3 < bj(oj) there are expenditures of the value of mj(bj{io) — Xj) and where Xj > bj(u>) then the losses are kj(xj — bj(u>)). Denote p3 as the price and c, as the expenditures on production per unit of j product and j/, as the supply of the product at the beginning of the period considered. Then the stochastic linear programming problem will be; n
n
n
max F(X) = maxl ^ p3Xj — ^ CjXj — Y^ k3 x 3= 1
3=1
3= 1
3.6. APPLICATIONS
OF TWO-STAGE x X
SPP
(S3
j+y,
/ ' "'(xj +
yj-bjH^ib^dbj-
J — oo
Jl
3=1
J
)+J)
n
n
n
= m a x j ^ p ^ j - J2(kj + c,)z; - YKki 3=1
J= l
/•oo
X
/
+ Mi)*
J= l n
(&iH - *i - t/iVi(&i)d6i + ^ m ^ ^ -
%)}
(3.109)
J=l
under the conditions Y,xuxj
* = l,m,
(3.110)
J=I
x, > 0,
i = I7n.
(3.111)
If the problem is solvable then with fixed demand b3 it has a trivial solution i = bj {j = l,n)- Elmaghraby in the the work [20] considered the problem akin to the problem (3.109)—(3.110) and made the assumption that the vector b^ui) is destributed discretely, than is it tades values bj(u>t); t = 1,T3 with probability PJt. In that case the problem will be: x
min/pjf) =
{
C 71 71
(( = = (( ''
-6J(^())PJ( +
= min < J2 J ] + kj E (xj ~ J2 CjXj + k3 J2 (x, - bJ(ut))pjt + C X
3=1
under the conditions
1=1
+"ii£iif(&jM-a:j)Pit 2 j XijXj = a,, *j>o,
« = l,?7i,
y = I7»*,
(3.112) (3.112)
(3.113) (3.ii4)
where t* is defined from the relation bj(wf) < x3 < 6j(w t . + i). In the work [33] it is demonstrated that the solution of the problem (3.112)—(3.114) is equivalent to the solution of a linear programming problem. The work [52] considers the problem of aircraft flight planning as a two-stage stochastic programming problem. Suppose one needs to make an sircrafy flight sched ule for regular and additional journeys. Regular journeys are made between fixed points and are scheduled beforehand for each planned period. Additional cruises appear at random, the time and point of destination not being fixed. Aircrafy for additional journeys may be taken from regular lines. Different types of aircraft are specified with useful load, prolongation of flight and costs on different lines. Demand it is necessary to transport during a day cannot be completely forecast. Since the
64
CHAPTER
3.
TWO-STAGE
STOCHASTIC
PROGRAMMING
PROBLEMS
moment one gets information about the random parameters of t h e problem it is nec essary to re-allocate aircraft from t h e lines which are in less d e m a n d t h a n planned to t h e lines for which demand is higher t h a n expected. Re-allocation, in particular, may be effectuated at t h e way-stations. T h e problem is solved by calculating the minimum of t h e mean expected costs for t h e whole planned period. The flight programming problem is set up as a two-stage stochastic programming problem. At the first stage — until demands for additional journeys are known — the aircraft of each t y p e are allocated to t h e lines; and the n u m b e r of journeys for the aircraft of each type on each line is determined. At t h e second stage, after the observation of realization of the random parameters in t h e conditions of t h e problem, a re-allocation of aircraft from line is made. The conditions of the first stage constrain from above for aircraft of each type, t h e total number of flight hours on all the lines 1, m.
J2a43xi3 < a,,
(3.115)
T h e conditions of t h e second stage fix t h e fact t h a t for each t y p e of aircraft the total number of flight hours re-allocated from a given route to other lines does not exceed the sum of flight hours planned beforehand for t h e route. Besides, necessary balance relations for each line are usual for two-stage stochastic programming prob lems. If aircraft of the type i, the time of whose flight along t h e line j is equal to 0,-y hours, are re-allocated to the line k then journeys on t h e latter line will take a,-^ hours; and this flight by the line k will cause t h e cancelling of aljk/atJ glights by the line j . Constraints of the second stage under the given assumptions will be: ^-^
^IJK "ijk
E kjXij + E
l,m;
<j2 x;
j
E bikXijk - E
E [ h ^ )
k±,
k*,
-V,
=
d
h
(3.116)
1,71.
x
>}h +
yt(3.111)
j = 1, n.
T h e objective function of the problem of planning flights under risk and uncertainty is expressed as follows ^C.JX.J + E <
min
E E (ctffc
X,]k + °'J a„ I •"».)
(3.118)
+ J =E(#V^-V) I Insert the following denotations:
3.6. APPLICATIONS
OF TWO-STAGE
SPP
65
x%J: number of journeys during a planned period for the aircraft of i type initially allocated to the line j ; x%]k: number of journeys of the aircraft of % type taken from the line j and re-allocated to the line k; yj; non-satisfied demands (in cargo tons) for the freights on the line j ; yj: non loaded capacity of the aircraft (in cargo tons) on j line; a,j\ number of hours required for the aircraft of i type to fly the line j if aircraft from the very beginning was allocated to the line; at]k'- number of hours required for the aircraft of i type initially allocated to the line j so that it could fly the route k. Besides, atJk > an,; bt}: quantity of cargo tons taken during a flight by the aircraft of i type on the line
r, at: value of flight hours admissible during the plabbed period for the aircraft of j type; dy demands on cargo freights (in tons) for the line j ; c,3: cost of the journey for the aircraft of i type on the line j provided that aircraft form the very beginning were allocated to the line; djk'
cos
t °f the journey for the aircraft of i type on the line k if it was taken from the line j . Evidently, c,jk > c^; penalty for non-satisfaction of the demand on freight per cargo ton on the line
i; q ~ . penalty for under loading per ton of the aircraft on the line j . Thus, it is required to calculate the non-negative parameters xx], z,^, yj, yj min imizing the onjective function (3.118) upon the conditions (3.115)—(3.117) including that on variables Xij and x,}k there is imposed the additional requirement of whole numbers. For the more adequate modelint of the process of planning for aircraft fraightn un der risk and uncertainty it is natural to use dynamic stochastic programming problems in which daily changes in traffic demand would be sequentially taken into considera tion. Dynamic stochastic programming problems will be consider in the following chap ter of the present work.
Chapter 4 Multistage Stochastic Programming Problems 4.1
Formulations of Dynamic Stochastic Programming Problems
Multistage stochastic programming problems are extensions of two-stage problems. In long-term planning, desing and managemet there are many practical problems that cannot be adequately described by static models. Long-term planning for de velopment of economic systems, control of combat operations, control of technolog ical processes and other problems contain random parameters and require dynam ic stochastic models for their description. For these uses, in particular, multistage stochastic programming methods and models are intended. Replacement of static stochastic programming models with dynamic ones was initially made by developing a two-stage stochastic programming model. Models for stochastic programming problems and methods for their implementa tion are essentially dependent on the information about parameter values of problem conditions that is available when the next decision is to be taken. It is possible to distinquish multistage problems which require at each stage that the discrepancies resulting from implementation of problem conditions and earlier decisions (i. e. deci sions made at earlier stages) should be compenstaed in full. Other problems require at each stage that the probability of satisfying constraints should exceed some pre determined value or expectations for residual functions of conditions be bounded by predetermined numbers or by the functions of random parameters whose values have been realized at earlier stages. Multistage problems may have conditional or unconditional, stochastic or static constraints depending on the nature of actual processes to be modelled. A special feature of multistage problems with unconditional constraints is that decision making is based on information about joint distribution of random parameters at all stages. In multistage problems with conditional constraints, two cases are possible: only 67
68
CHAPTER
4. MULTISTAGE
STOCHASTIC
PROGRAMMING
PROBLEMS
information on realization of random parameters at earlier stages is available when decision is to be taken and full information is available on realization of random parameters up to conditions of the current stage, but the values of r a n d o m parameters of subsequent stages are still unknown. Between multistage problems with conditional and unconditional constraints there is a certain relationship. Optimal solutions to multistage stochastic programming problems can be obtained both in pure and mixed strategies. In the latter case, components of solution or statistic characteristics of distribution which constitutes a solution, are dependent on the values of random parameters of problem conditions t h a t have been realized by the time of decision making. Construction of dynamic stochastic models and development of methods for their implementation involve considerable difficulties, which, in particular, applies to mul tistage stochastic programming problems. This chapter deals with some issues of mathematical formulatios of multistage problems and procedures for constructing solutions for such problems. For the purposes of further discussion of formulations and analysis of multistage stochastic programming problems we need to define the following concepts. Suppose t h a t at the i-th stage ft, , i = 0 , 1 , . . . , n there are some spaces of elemen tary events u; Let Qk be the Cartesian product ft,, i = 1, k 1u>k = (u>i , . . ., uk) ; ft" = ft with the probability measure P given on ft and defined in t h e following way : if A C ft* , then pk(A) = p(A x flk+l x ... x ftn). We introduce the probability space ( f t , E , P ) with a suitable cr-algebra E and define Pk - conditional probability measure on ft1" Pk(A\ ujk~ 1 € B) =
Pk(A x B)
Pk(nk x B)
for any A C ft* , B C £lk-\. Denote by Xk the Cartesian product A',, i = 1,...,&; A'* = (a:1, ...,xk)g ft on t h e set X . We intoduce G° = G°h(wk) as some r a n d o m sets and ^ ( u / - 1 ) m^-dimensional random vector functions of a i t _ 1 t h a t are bounded and measurable. Denote by Bk some Banach space which includes bk(wk~1)
]T m,-dimensional ;=i
vector function. Finally, denote by Ewt(u(ujk)\ijk-1) conditional expectation u(wk) on the assumption t h a t realization of ujk~l is known. We will consider various formulations of stochastic programming problems by employing the notions and notations introduced above. Suppose we have the following multistage stochastic programming problem %K,.V")
-J-inf,
(4.1)
4.1. DYNAMIC
PROBLEMS
i,')
Eif>k(uk,Xnk>bk, k
X
(4.2)
e Gk , k = T~^.
(4.3)
For the problem to be fully formulated, it is necessary to point out whether constraints are conditional or unconditional, how the solution to this problem is determined(i. e. in pure or mixed strategies) and the class of measurable functions or distributions where a solution has to be sought for. In practical problems, where the meaning plays a central role, the solution can be computed at each stage as a determined vector, a decision rule of function of realized and observed parameters of conditions, or as a decision distribution (conditional distribution) Xh on the assumption that some information is available on the realized values of initial random data. Formulation of the problem and its inormation structure are specified by the information that is available when the next decision is to be taken. In the terminology adopted in [9], multistage stochastic problems are generated by chains of the form —observation — decision — observation — ... — decision decision — observation — decision — ... — decision We will consider various models of multistage stochastic programming problems by employing the classification adopted in [16]. The multistage stochastic problem with unconditional constraints has the form <^oK,-V")
f
I n *xk k
k
Vk{u\X
)dF„k,Xk>h,
(4.4)
(4.5)
Xk e Gk , fc = M
(4.6)
In the set of information structures corresponding to multistage probleme with conditional constraints, we separate out several classes that may be usefull in appli cations. For the problem with conditional constraints that can be solved in mixed strategies, the model (4.1)-(4.3) becomes j
^•,r)tf
u
.x.^m/
jh(uk-1), nkxxk Xh € Gk(ujk) , k = M
1
(4.7)
(4.8) (4.9)
The set of distribution functions Fx*|„* is a solution to the problem. It is generally agreed that the problem can be solved in terms of a posteriori decision distributions if Fv*| k is determined after realization and observation of random parameters of wk
70 CHAPTER 4. MULTISTAGE
STOCHASTIC
PROGRAMMING
PROBLEMS
with a posteriori decision distributions dependent on Xk~x and uk It is also said that the problem can be solved in a priori decision distributions if Fx^\uA is determined after realization and observation of uik~l, but prior to obsrvation of wk, with a priori decision distributions dependent on Xk~l and uk~l If the multistage problem with conditional constraints is solved in pure strategies, concrete definition of model (4.1)—(4.3) becomes if0(u>n,Xn)dFu«
/
f
(4.10)
->inf,
t
Vk(u
,Xk)dF^y.-i>ln.(uk-1),
(4.11)
Xk e Gk{uk) , k = ITn
(4.12)
The set of Xk functions of the realized and observed random parameters of problem conditions is a solution to the problem. The problem can be solved in a posterior decision rules if the decision is taken after realization and observation of u , in which case a posteriori decision rules become A'' = Xk{uk) The problem is said to be solved in a priori decision rules if decision is taken after realization and observation of tuk~1, but prior to observation of uik; in this case a priori decision rules become
Xk=Xk{wk~l). Problems (4.7)-(4.9) (or (4.10)—(4.12)) are usually referred to as multistage sto chastic problems in rigid setting if they contain no conditions of the form (4.8) (or respectively (4.11)) and decision is taken at each stage after observation of parameters of conditions and earlier decisions. There is a relationship between domains of conditionally and unconditionally con strained problems. Below is given an extension of findings in [8] to a partial linear multistage stochastic problem. We will consider this statement by following [21]. Let U denote the set of feasible solutions to the multistage unconditionally con strained problem U = {xk 6 Gi x ... x Gn\E¥k(u>k, Xk) > bk, k = l , n } and let V[fe„(cj"-1)] denote the set of feasible solutions (decision rules or decision distribu tions, a priori or a posteriori) to the conditionally constrained problem. T h e o r e m 4 . 1 . The sets U and V are related by U = {Xn £ V[b(u>n-l)}\E~b{uJk-1) = bk , k = 1777} Proof. Introduce notations: V = Xn G V[bn{iOn-l)}\Ebk{uk-') Let Xn eV
= bk,k = I7r7.
This means that Eu*y>k{uh>Xh) = Euk-,{Eukifik{uk,Xk)\uk-1}
>
4.1. DYNAMIC
PROBLEMS
71
> Eujk-,bk(Lok-1) = bk; k = (l,n),i.e.Xn n
Now let X
€ U.
e U. Define M*"* - 1 ) = E^{ipk(cuk,Xk)\Lok-1} k
h
+ {bk - E^
k 1
< Eu„{^{u1 ,X )\u - }, h y
<
k = T^. k
By the definition of bk(uj - ) we have Ewk-,bk(tu -') = bk. Therefore Xn € V, which proves the statement. Corollary 4.1. With the same functions ifk(ujk,Xk) and sets Gk, k = T/ri , domains of feasible solutions to problem (4.4)-(4.6) and (4.7)-(4.9) or (4.10)-(4.12) (depending on whether the problem is solved in mixed or pure strategies) coincide if and only if Ebk(u>k~l) = bk. The above statements permit reformulation of qualitative results and sometimes computational methods developed for one class of problems in order to provide mean ingful investigation of the other class. A relationship between decision distributions and decision rules can be useful in handling practical problems. If the function ipo is convex and components of vector function ipk are convex in X at each u>, the optimal values of the objective functional attained on decision distributions can also be attained on decision rules. Convexity of ipo and) and pjt(w ) bounding in absolute value ip0(ujn,Xn) and all components ipk(u>k,Xk), respectively. Then the optimal a pos teriori decision rules of multistage stochastic problem determine the same value of objective functional as optimal a posteriori decision distributions. The proof of Theorem 4.2 for a single-stage stochastic problem is given in [16]. Transition to a multistage problem requires more cumbersome notations and compu tations. The statement of the theorem also holds for Borel functions if, even though no assumption is made as to reqularity of measure FWixMultistage stochastic programming problems with conditional constraints can be replaced by a system of problems corresponding to separate stages. Let us consider a problem of the form (4.10)-(4.12) that can be solved in pure strategies. Domain of the i-th stage problem corresponds to the set K, = {X, € G ? | 3 [ y i + 1 e G ? + 1 ) . . . , y „ € G £ ] ;
£^(ty,x,)l^-i]>^,~1),
72 CHAPTER 4. MULTISTAGE STOCHASTIC
PROGRAMMING
Eu,Jlpi(uji+s,X\yi+l,...,yi+s)\^'-1}
PROBLEMS (4.13)
>
> 6 t + s (^ + s - 1 ), where VCJ* +S - 1 , ...,un-L,
S =
l,n-i}.
G° is projection of G, onto a cutting plane determined by components of vector A',. Existence requirement for vectors y,+s, S = l , n — i satisfying (4.13) is equivalent to availability of induced constraints in the classical two-stage problem. The necessary and sufficient condition for solvability of problem (4.10)—(4.12) is provided by K1 ^ $ (objective function (4.10) is assumed to be bounded below) if, additionaly, R\ ^ $, then also K, ^ $,? = 2,n. The objective function Q,(Xt) of the i-th stage problem is conditional expectation ipo(u>u, Xn) on the assumption that at the earlier stages (i.e. prior to the i-th stage) the set of random parameters oj' _ 1 has been realized and the decisions constituting the set Xi-i have been taken, while at subsequent stages (i. e. following the i-th stage) optimal decisions X*+1, ...,X* have been taken: Q.(-v.) + B w B , u „-,(w B ,Jf,_ 1 I A„A7 + 1 ,...,jr;)
(4.14)
Thus, determination of an optimal decision rule at the i-th stage of multistage stochas tic problem amounts to solving the programming problem inf. Q,(X.)
(4.15)
A posteriori decision rules now become A, = Xi(u>'),y,+S = y,+s(uj'+s, while a priori decision rules become
5 = l,n — i,
X, = A , ^ - 1 ) ; y 1+s = y ! + s ( ^ I + s - 1 ) , s = 1 ^ = 7 . n
nn nn If the objective function is separable , i. e. ip
Qi(Xi) = Eut.]u.-i{tp0(u,i,Xi)
+
'
Q:+1(LO\X')}
where Q r K - ' . A ' - 1 ) = inf. ^ . - . { v ^ V . A " ) + Q * + 1 ( ^ , A " ) } , A, t A,
i = 1, n — 1, with i = n Q:L"-\X^)
=
inf
^ - . ^ ( u ^ Y " )
Along similar lines problems of separate stages are constructed for multistage stochastic problems that can be solved in mixed strategies (in decision distributions).
4.2. QUALITATIVE
4.2
ANALYSIS
73
Qualitative analysis of Multistage Stochastic Problems with a Posteriori Decision Rules
Suppose problem (4.10)-(4.12) can be solved in a posteriori decision rules. With the properly defined norm, the set of vector functions bk(uk~l) of right-hand sides of con straints (4.11) forms Banach space Bk, in wich case each vector function 6"(a)n_1) G Bn corresponds to its own problem (4,10)-(4.12). Denote by S(bn(ton~1)) the lower bound for values of the objective functional (4.10) depending on the right-hand side of constraints (4.11). Then S(bn(uin'1)) is the function on Bn. Suppose the measure Fw™ is continuous. We have the following statement. T h e o r e m 4.3. With the X-convex functions tp0(utn,XH} and — tpk(uk,Xk); k = l,n the functional S(bn(uin~1)) is convex. With n — 1 , convexity of S(bt) follows from Liapunov theorem of vector measures [31] in much the same manner as, e. g., in [12]. To validate the statement for n = 1, no assumption is required for convexity of ip0 and —tpi. For n > 1, however, convexity of 5(6 n ((jj" _1 )) requires some assumptons concerning the structure of objective function and constraints of problem, e.g. , convexity of ip0(uf\Xn) and — y>k(u) , Xk), k = n l,n. Although convexity of'', X ) and — cpk(u> , X ) in x constitutes a sufficient condition, it is not a necessary condition for convexity of S(bn(u>n~1)). Theorem 4.3 permits construction of a dual problem to (4.10)—(4.12). We have ht(Xk,J'1Xk-1)=
sup [(\k,ipk{iJk,Xk-l,Xk))-Sk+i(wk,X%
Hk(Xk,uk-\Xk-1) Sk{uk-\Xk-x)=
(4.16)
= j hk(\k,uk-1,u>k,Xh-1)dFut^t-i,
svipihM^'1))
- Hk(Xk,wk-1,Xk-1)],
(4.17) /c = M
(4.18)
At>0
where 5 „ + i K , I " ) = M u " , r ) ; A i = {Xkj}; k = T^n~;j =
T~^.
The scalar product in Euclidean space is symbolised as (•,•), while the symbol > is taken to mean the ordering in non-negative orthant. Let S{bn(un~>)) denote the closure of function 5(6"(w n_1 )) in the norm of space Bn, i. e. 5(6"(u;" _1 )) is the largest function that is semicontinuous below on Bn and does not exceed S(bn{wn~1)). We have the following duality theorem. T h e o r e m 4.4. There is a. relationship 5j = 5(6"(w"" 1 ))[21]. Here bn(ujn-1) is not a variable parameter, but the set of functions on the right-hand sides of (4.11). The above statement permits formulation of a posteriori decision rules for problem (4.10)—(4.12). Thus we have a sufficient optimality condition. T h e o r e m 4.5. Suppose there exist vector-functions
X " K ) = {Xl(^),...,xn(u>n)} e G„(wB),
74 CHAPTER 4. MULTISTAGE STOCHASTIC
PROGRAMMING
PROBLEMS
A"^"" 1 ) = {A ll A»(w 1 ),...,A n (w^ 1 )} > 0, satisfying t i e relationship (4.11) (as equalities for A,-j(u> hk(Xk,LJk,Xk-l(ujk-1))
) / 0), and
= (\k,Vk(iuk,Xk(ujk)))-Sk+1(^,Xkyi
fc=T~»,
(4.19)
where 5 n + i (wn, X") = ^>o(w", X") , while the other are computed by means of recur rence relations (4.16)-(4.18), and Xn(ujn) is the set of optimal a posteriori decision rules to problem (4.10)-(4.12). Proof. Our proof follows [21]. From relations (4.16)-(4.18) we have: n
,*»( u ))dF^ <
/
Hfc x . . . x f i n
-\xk~> -1\
=
X* " V " 1 ) ) ^ „*-! +
k
hk{\k{ujk'
= -
^dP^u-i
ft*
+ J(h(0Jk~l):
t
k
,*''(w*)))^,^+ (A k (« fc - 1 ),b fc (w fc - 1 )) <
= -Hk{\k{u>k-1),wk-\Xk-1(u;k-1)) <Sk(LOk-\Xk-l(uk-1))]k
= n, n-1,...,
1.
When k = n, the first inequality is replaced , by definition, with equality. With k = 1, we additionally have Si < 5(6"(o; n ~ 1 )), but on the other side, for k = 1 I
ifin(un,Xn{un))dFan
=J
r2i X...x£2 n
n
,Xn(ujn))dFur,
tpo{uj
> 5(6"(u ) "- 1 )).
fin
Therefore fipo{<»*,X*{u>n))dFy* = 5(6 n (w n - 1 )), nn i.e. A""^71) is a solution to problem (4.10)—(4.12), which proves the theorem. The statement we have proved can be used to construct optimal a posteriori de cision rules for multistage stochastic programming problems with conditional chance constraints. Suppose we have the problem Eipo(ton,Xn)
-> inf,
P{XK 6 Dk(u>k,Xk-1)\u>k-1} k
k
Xk(u )£Gl{u ),k
> ak(ojk~'), = T^,
(4.20) (4.21) (4.22)
4.2. QUALITATIVE
ANALYSIS
75
wherek ! ) belonging to the interval [0, 1] (problem (4.20)-(4.22) is a special case of problem (4.10)-(4.12)). We will employ the following notations: ak(uk,Xk-')=
sup §k{u}k,Xk);Ck(ujk,Xk-1) x„eDt = {uk e nk\ck(uk-\wklxk-i)
uk(\k,uk-\xk~l) Vk{\h,uk-\Xk~l)
= {cuk € Slk\ck{uk~1,uk,Xk-1)
where Uk(Xk,ujk-\Xk-1) respectively,
and Vk(\k,uk~\
\l(uk-\Xk-1)
< ak(uk-\u>ktXk-l)
Hk{\k,uk-\Xk~l)
(4.23)
+ xk}
+ \k}
(4.24) (4.25)
Xk~1) are measures of the sets Uk and Vk
= inf{A > 0\ak(u>k-1) < Uk(Xk,uJk-\Xk-1)};
$,_1(^-1,A'fc-1) =
hk(Xk,u>k ,Xk~1)
= svLp$k{uJk,Xk); xk < ak(uk-\tok,xk-1)
(4.26)
Hk(Xt[uk-\X^l),uk-l,Xk-l)-ak{Ljk-1)Xl(u,h-l,Xk-1); (4.27)
= Jhk(\k,Uk-1,wk,Xk-l)dFu,tlu^; n* = max{a t (u;' ,Xk~l)
+ Xk,Ck{uk ,Xk'1)}.
(4.28) (4.29)
Theorem 4.5 permits verification of the following statement. Theorem 4.6. [21] Let (p0(uin, Xn) denote the function bounded below and suppose conditions (4.21)-(4.22) are compatible. Then the lower bound of functional (4.20) ipg(u}n, Xn) = — $„(^", X"). This number is determined from the solution of recurrent system (4.23)-(4.29) for k = n, n — 1,..., 1. To write out optimal a posteriori decision rules for multistage stochastic prob lem with conditional chance constraints, we introduce at each (k-th) stage the sets Lk(ujk-\Xk-1) and Nk(uk-\Xk-1) . The set Lk{uik-1 ,Xk~l) is an arbitrary subset of k 1 k 1 measure ak(ui ~ ,X ~ ) - Vk{X'k)uk-\Xk-1) of the set Rk(wk-\Xk~1) = = Uk{Xl{uk-1,Xk-1),wk-\Xk-1) vk{xi{ujk-\xk-l),uk-\xk-1). The set Nk(uk-1,Xk-1) completes Lk(u>k~\ Xk~l) to Rk(cjk~\ Xk~l). n n In these notations X (u> ) determines an optimal set of a posteriori decision rules of problem (4.20)-(4.22) if and only if at each stage f e 6 fc (w k , Jf*- l ),**(u;* ) Jf*(w*)) = ak{uk,Xk-1(u^-1)), k k 1 k k 1 k if ukeUk(Xl(u -',X - ),u> -\X - )\Nk(uJ -\Xk-i) Xk(iok) : £$k(uJk,Xk(uJk)) = Ck(ujk,Xk-i(Lok-')), if Uk eUk(\i(uk-1,Xk-1),L>k-l,xk-1)uNk(uJk-\xk-1).
f43m {
'
'
In planning, design and management there are many problems that can be for malized as extremum problems. Solution of deterministic extremum problems is not
76 CHAPTER 4. MULTISTAGE
STOCHASTIC
PROGRAMMING
PROBLEMS
always brought to a form convenient in practical applications. Above all this applies to situations where operational decision choice is required, because solution algorithms of deterministic problems are extremely laborious. In stochastic extremum models, the decision process can be extended in time and divided into preliminary and operational phases. In the preliminary phase decision rules or decision distributions (solutions) for stochastic problems are prepared. Here no rapid information about initial data is required. One has to know the structure of problem and some statistical characteristics of random variables. Another approach to stochastic problems has been devised in recent years. This is the iterative method of constructing decision rules or decision distributions by successive observations of realized random variables in order to provide development and flexible improvement of decision rules or decision distributions where little or no statistical characteristics of initial information are available. We will consider this matter and provide generalized schemes of stochastic approximation.
4.3
A Priori Decision Rules in Multistage Stochastic Programming Problems
The preceding section shows that general principles of duality theory can be applied to function spaces for the purposes of qualitative analysis and construction of a pos teriori decision rules of multistage stochastic problems. However, development of a priori decision rules is much more complicated. At present there are no complete mathematical investigations that may form the basis for construction of a priori de cision rules. This section deals with the matter and directs the way to construction of a priori decision rules for multistage stochastic problems The initial focus is on separate models of multistage stochastic programming problems for which optimal a priori decision rules can be constructed Let us consider the multistage stochastic problem with conditional chance constraints
Elf2cj(uJ~')xJ I ->max, P{Anxlai, P{A^)xx
P{Anl(ion-')X]
(4.31) (4.32)
+ A22(u;2)x2 < 62(w2)|w1} >
+An2{u>n-i)x2
+ ... +
> o n f w - 1 ) , ^ = XjiiJ-1)
Ann(wn-1)xnn-l}> > 0,j = T ^ ,
(4.33)
where c} and XJ are vectors of dimension n,-; 6; and a; are vectors of dimension m-' components a; belong to interval [0,1]; A;, are the m; x n3 matrices; operator P is
4.3. A PRIORI DECISION
RULES
77
utilised by rows. Problem (4.31)-(4.32) is equivalent to the problem
£JE«»V"1)si|-»-mH)
(4.34)
J2 / ! , > * " > , < Fr\l - ^(w*- 1 )),! = M
(4.35)
3=1
Z j V " ) >0,j = T^
(4.36)
-1
where F,z = P{fe,(u') < zju;' } is the conditional distribution function for compo nents of vector bx with the a)' - 1 fixed , while F~x(t) = s\ip{y\F,(y) < t} and both latter equalities are componentwise. Under certain assumptions of constraints on problem (4.31)—(4.33) it is possible to derive explicit expressions for its a priori decision rules. Model 1 is based on the following assumtions : 1. Cj = CjUJ3'1 are vector with nonpositive components. 2. AtJ = Aij(aJ ,_1 ),! / j are matrices with nonpositive elements. 3. Af, = /l,;(i _1 ) are diagonal matrices with negative elements arranged along the diagonal. 4. Components F~l(.) are bounded below Vu;1-1, i.e. inf F~l(l -at(u*-1))
> - o o , i = I~ri
We have the following statement . Theorem 4.7. [19] Under assumptions 1-4 optima.! a priori decision rules of problem (4.31)—(4.33) become xt = - A - I [ J F r 1 ( T - a i ) ] ( - ) , s,V-i) = -AZ1('^l)lir1P-*("i-1))
(4.37)
- £ A^-^x*}^
= T^
(4.38)
J=I
wiere [a](_» = max(0, — a) (component-wise maximum). Assumption 3 can be weakened and replaced by the following assumption. 3'. Matrices Ati(w*_1) are rectangular, rank of An is equal to the number of rows in the matrix: elements of matrix Au and those of pseudoinverse matrix Af, are nonpositive and, additionaly, the relations CibS-'W - At(^-l)A,^-l)j
=0
(4.39)
hold for all W;_i (/ is a unit matrix). If assumptions 3 are replaced by assumptions 3', then in the optimal decision rules of problem (4.31)-(4.33) inverse matrices A~x have to be replaced by pseudoinverse matrices A%. From conditions (4.39) it follows that the objective functional of the n-th stage problem will remain uchanged if u> = 0.
78 CHAPTER 4. MULTISTAGE STOCHASTIC
PROGRAMMING
PROBLEMS
Model 2 is characterized by the following assumptions: 1. Aij = i4^(w 1 - 1 ),i / j are matrices with nonpositive elements. 2. An = Aii(w*_1),« = l,n are diagonal matrices with positive elements arranged along the diagonal. 3. urf.F}- 1 (T-a,-(w*- : t ))>0,*=I7n Under the above assumptions the domain of the fc-th stage problem is nonempty for an arbitrary set xk~l, hence the fe-th stage problem does not induce additional constraints for the problems of earlier stages. Model 2 can be described by the multistage stochastic problem. Let us introduce notations C(+'(^-1) = max{0,ft(^-1)},(7„_fcK^-1) =
= Cn-h{J*-h-1) - Y, £ . - * + i , » - ^ - * - {citli^-'-^A-^^A^,^}
;
k = 0 , n - 1. _1
With k = 0,e n (w" ) = Cn(u> ~ ); / ^ ( w " " * - 1 ) is a diagonal matrix whose di agonal elements are equal to 1 if the corresponding components Cn-k (uin'k-1) are positive , and are zero otherwise. We have the following statement. Theorem 4.8. Under assumptions 1-3 optimal a priori decision rules of problem (4.31)-(4.33) become n l
<-* = <-k(
~
an.k(^-k-1))-
A n ^_ l i i ( W * 1 -*- 1 )x;(^'- 1 )],fc = 0 , l , . . . , n - l
(4.40)
3=1
For the proof of this statement see [19]. Assumption 2 of the degeneracy of A,i can be dropped and replaced by a less strict requirement that follows. 2'. Matrices Aft are rectangular; rank of An is equal to the number of its rows; elements of matrix An and pseudoinverse matrix A^t are nonnegative and, additionaly, for all O J - 1 C , ( W S - 1 ) ( / - AtlAu)
= 0
Under the above assumtion, expressions for optimal decision rules (4.40) are pre served if inverse matrices A',1 are replaced therein by pseudoinverse matrices Af{. Constructions of decision rules for multistage stochastic models 1 and 2 are not related to the linearity of respective problems and can be extended to some of the special multistage nonlinear stochastic models with conditional chance constraints. Let us consider the problem n
EY.i'oj^'1,'^]) i=l
->max
(4.41)
4.3. A PRIORI DECISION
79
RULES
p{£0*V~S*i)<Mw*)l(w t ~ 1 )} > ^ ( ^ - 1 ) , xk = xk{ujk-i)
> 0,/c = I~r7
(4-42) (4.43)
The following requirements are said to be satisfied. 1. ipojiu3'1 ,XJ),J = l , n is the monotone nonincreasing functional in xf, foiiiJ-1,0)
2. 0,j(ui*
=0,VCJJ-1
, Xj),i ^ j are the monotone increasing vector functions in xr
i / ^ - S o ) = o,Vw'^j,i = M , j / i
3. i/>„(u/-1, x,),i = l , n are the monotone decreasing vector functions in x, which permit existence of inverse vector functions >~V~\y) = raf{*|&.-(w,"'l»*) = y},i/> t! (a,'-\0);Vu>-\< = M 4. Components of vector functions F~1(.) are bounded below. Then we have the following statement T h e o r e m 4.9. [19] Under assumptions 1-4 optimal a priori decision rules of problem (4.41)-(4.43) become X"n = i>n{-[Fr1(7
-<*!)](-)},
(4.44)
AV = A-*(^-1) = 0.T1 V 1 , -[#•-'(/ - a,^'" 1 )) - £ l^(w,-1,*;)](-)},i = 2T^. (4.45) We shall now consider a special class of multistage stochastic programming prob lems where each stage induces additional constraints n
£{EC' J K" 1 )^}->max, P{AlXl < b^1)} P{-AiXl
Pi-An-tx^i
(4.46)
> au
+ A2x2 < b2(u2)l(u1)} > Q 2 ( ^ ) , + Anxn < b^u^K^-1)}
> c^""1),
(4.47)
Anxn > 0 Here operator P is used in a component-wise manner. We make the following as sumptions. 1. Rank of matrix A,,i = l , n is equal to the number of its rows. 2. CJa;' l - 1 )/l+<0,C„K- 1 )(/-A+A„) = 0,V^-1
80 CHAPTER 4. MULTISTAGE STOCHASTIC 3. r,n_i(ujn-2)
= in::
i
F-'HI ~°n ( w - I ) ) >
PROGRAMMING
PROBLEMS
-oo,Vu;n-2
4. Fixed constraints: for separate stage are consistent. We introduce notati ons
a n _i(w n " 2 ) = C n - i ( w
Ci(w'' "')
} == C,{u>'~ )
J,
+ *U~V « - i [ C i + i ( w ')](+)Ai
i = 1,71-2, [«](+) = m a x ( 0 , a ) ;
/«(w*- ')
1 is the diagonal matrix with elements [C(w t_ " )](+) arranged along the diog-
onal;
hi = I - hi',* = 1,»— 1; 8
/in-iCw- ) = - M ( ^ ' ) = - mf F - 1 ( 7 - a „ ( t y 1 - 1 ) ) ; / . 1 ( W '-
1
) = -inf[JF!;11(/-aJ+1(W'))-^+1(^)],J = l , n - 2 .
We have the following statement. Theorem 4.10. [21] a) For problem (4.16)—(4.17) to be solvable under assump tions 1-4, it is necessary and sufficient that Ci(iJ-l){I
- AfAA = 0,Vw'-\i = I7n, Ff 1 -
W
> 0
b) Optimal a priori decison rules for a solvable problem of the form (4.16)—(4.17) under assumptions 1-4 become
x;(u;"-1) = o, X . V ' - 1 ) = /j,(w-1)A*<(wi-1) + / « ( w < - , ) [ f ; - 1 ( 7 - a j ( w t - 1 ) ) + +,4,_ 1 .Y,"L 1 ( w - 2 )], l = l , n - l .
(4.48)
Optimal decision rule of the form (4.48) is nonunique. The general form of optimal a priori decision rules of problem (4.46)-(4.47) can be derived by adding to the righthand sides of (4.48) for the i-th stage an expression of the form (/ — Af Ai)y{ui'~l) where jfc(w,_1) is an arbitrary vector Furthermore, it is possible to costruct a priori decision rules of problem (4.46)-(4.47) by rejecting assumption 2. In this case X*(uin ) is nonzero; expression (4.48) is preserved for X*{u>'~l),i = l , n — 1; only expressions for C,u/ _1 undergo changes. The schemes for constructing a priori decision rules utilized the structure of special classes of multistage stochastic programming problems. Two general approaches to construction of a priori decision rules are described below. Solution of multistage stochastic problem of a general form reduces to analysis
4.3. A PRIORI DECISION
SI
RULES
of a simpler structure (e. g. of a single-stage problem with a posteriori decision rules or a two-stage problem) which provides an optimal plan to restore a solution to the original problem. Let us consider a multistage stochastic programming problem iWo(wn,*n)->-sup,
(4.49)
BuJo/5} > Mw*- 1 ),
(4.50)
Xk EGk,k
(4.51)
= l,...,n.
Let L, and Wt denote the classes of measurable function constraints mapping respec tively ft = ft" and ft1-1 into Rn' ( n, is dimension of the optimal solution X, at the j-th stage ), while $ ; = (>: L, -4 Wi). We will consider an arbitrary set i\ g $, and set up a correspondence between multistage stochastic problem (4.49)—(4.51) with a priori decision rules and two-stage problem with the same objective functional (4.49) and with the same first stage constraints as in (4.49)-(4.51) and with the second stage constraints X,eG f e 0 ,fe = 2,...,n
,
(4.52)
E^oK.ff.r))-^;^.!")} > o where, by definition,
^(Xk)
= {XuMX2),....,MXk)}1k
= l,....,n;
G°k is projection of Gk onto the coordinate cutting plane determined by components of vector Xk- We have the following statement. T h e o r e m 4.11. [21] Let Xu = const; Xt, = Xlr(ujn),i = 2,..,n be an optimal solution of two-stage problem, then x; = Xi.,x;
= Mx»),i
= 2,...,«
(4.53)
are optimal a priori decision rules for the original problem (4.49)-(4.51) The role of two-stage problem from Theorem 4.11 (an"equivalent" original mul tistage problem) can also be played by other stochastic models of a simple structure generated by problem (4.49)-(4.51) for wich methods of constructing decision rules have been developed. Analogs of Theorem 4.5 can be formulated for such models. Another apporoach to construction of decision rules for multistage stochastic prob lems follows from the relationship (see[30j) between stochastic programming problems and lexicographic optimization problems. Lexicographic optimization is a special scheme of vector optimization with the emphasis on vector components with smaller numbers. It is generally said that inA
troduction of lexicographic ordering (R , >) into the /c-dimensional Euclidean space A
Rh can be written as y > y' if for any pair of elements y and y' from R" for all i < k such that j/i < y[ there exists j < i for wich ?/, > y'}. The lexicographic optimization problem is written as
82
CHAPTER
4. MULTISTAGE
STOCHASTIC
Z* = A m a x { U ( X ) \ X
PROGRAMMING
PROBLEMS (4-54)
€ Q)
where U(X) = {U,(X)} is the fc-dimensional vector function. [30] shows t h a t the solution set Q* for problem (4.54) (if it is nonempty) is determined by t h e recurrent procedure Q0 = Q Qi = snp-1{Ul(X)\X
eQt-l},i
= l,...,k
,
(4.55)
Q* = Qk In [30], the procedure (4.55) is taken to define lexicographic optimization. We now relate the t-th constraint of the fc-th stage of problem (4.49)-(4.51) to the set Gkt(ujk-\Xh)
= {Xk\EuHujk-14>kAu\Xk)
> bki(ujk-l)7
XseGs,s
= l,...,k}.
(4.56)
Let Kjt(o)* -1 , Xk) be t h e vector whose i-th component is t h e characteristic function of t h e set Gfc, and let Uk{Xk) = jE* -1 Kfc(w fc_1 , Xk). We will order vector functions Uk by stages, and by a priori significance of constraints within these stages. Introduce the vector function U(Xn) = (U1(X1),-,Un(Xn))T and consider the lexicographic optimization problem A max U(Xn)
(4.57)
where Un is t h e space of measurable functons on H * - 1 If problem (4.49)-(4.51) has a solution, t h e n problem (4.57) also has a solution t h a t is optimal in t h e sense of (4.49)-(4.51). However, the solvability of problem (4.57) does not imply t h a t the ordinal problem has a solution. T h e solvability of problem (4.49)-(4.51) follows form t h a t of (4.57) only when all components b u t the last one are unity on the lexicographic maximum. Another lexicographic interpretation of the problem (4.49)-(4.51) can be obtained when conditions (4.50) are required to hold not for all, but for all most all realisations u>k~1 Let us introduce the n + 1 vector function U(Xn,yn) = {Uk} whose components are determined by the relations Uk (Xk,Yk)
= - H ^ i ^ - , ^ ( a / , A'*) -bk
< y - > ) - ak ( c ^ - 1 ) ||, fc = M
t/„ + 1 (A r n ) = £ ^ o K \ A ™ ) . : 1
(4.58) (4.59)
Here ||a / t (tj' " )|| denotes the Hilbert valuation of the function space ak(u>k~l). Lexicographic interpretation of the multistage stochastic problem changes when the norm ||a(w)|| is defined in other ways. Let us set up a correspondence between problem (4.49)-(4.51) and lexicographic optimization problem
83
4.3. A PRIORI DECISION RULES \maxU(Xn,Y«),
XkeGk,Yk>0
(4.60)
If problem (4.49)-(4.51), in which condition (4.50) holds for almost all w, is solv able, then problem (4.60) is also solvable and its solution determines optimal a priori decision rules for the original problem. Let us consider the lexicographic optimization problem U(X) ->• A. max.X € Q, ^4'61) T where U{X) = {Ui(X),..., Un(X)} , and Q is an arbitrary set. We define on the Cartesian product Q x Q the function [/(X,Y) = ; £ > * s i g n
(Uk(X) - Uk(Y))
(4.62)
where {o^} is an arbitrary set of parameters for which an>0,ak>
J2 a„k=
l,n - 1,
(4.63)
fcft+i
{a*} can be taken, e.g. , as the sequence ak =
■,_, ,fc= l,n. here
L/(X,K) = - / 7 ( y , X )
(4.64)
We have the following statement. Theorem 4.12. [21] For the point x* G Q to be a solution to problem (4.61), it is necessary and sufficient that U(X,X')
< 0 , V A ' e Q.
(4.65)
Corollary 4.2. Let X' be a solution to problem (4.61), then U(X,Xn) = 0 if and only if X 6 Q is also a lexicographic optimum of problem (4.61). x Corollary 4.3. IfU{X, Y) < 0, then X < Y We have the following statement. Theorem 4.13. [21] Problem (4.61) is solvable if and only if maxmmU(X,Y) A-£Q Y£Q
Theorem 4.14.
= 0.
Problem (4.61) is solvable if and only if minmaxc/(.Y,V') = 0. YtQ
T h e o r e m 4.15.
(4.66)
(4.67)
XeQ
Equality min max U(X, Y) = 0 YeQ X€Q
is valid if and only if the function U(X, Y) has a saddle point in Q x Q. The last assertion permits utilization of the familiar minimax computation meth ods for constructing decision rules of stochastic programming problems.
84 CHAPTER 4. MULTISTAGE STOCHASTIC
4.4
PROGRAMMING
PROBLEMS
Duality in Multistage Stochastic Programming
In the construction of a posteriori decision rules for multistage stochastic program ming problems, dual problems(A-problems)can be used under sufficiently general as sumptions about the structure of the original model. In constructing a dual problem for multistage stochastic programming models, one has to repeatedly utilize exten sions of the minimax theorem. John von Neumann proved the minimax theorem max min f(x,y)
= min max
f(x.y)
for the simplest case, where f(x,y) is a bilinear function on X x Y, while X and Y are finite-dimensional simplexes. Wald and other authors generalised von Neumann theorem to the sets X and Y belonging to some infinite-dimensional spaces. Knezer, Fan Tsi, Berg and Nikaido extended the minimax theorem to concave-convex func tions that are semicontinuous in x and y, respectively. The most general results in this respect were obtained by Sion. Let us introduce the following notations: 1. The function f(x, y) on X x Y is quasiconcave in X if for any y € Y and for any real c the set s|/(s,y) > c is convex. 2. The function f(x, y) on X x Y is quasiconcave in Y if for any x 6 X and for any real c the set Y\f(x,y)■ XQ, and let F(x0) be the image set of the point xa. 4. The function f(x) is upper semicontinuous in A"0 if L C F(X0) and lower semicontinuous in A'0 if F(X0) C Z. If f(x) is upper and lower semicontinuous, F(x0) = L, and the function is continuous in A'0. The function is upper(lower) semicontinuous on the set A' if this property holds for all x 6 X. 5. The function f{x,y) on X x Y is upper and lower semicontinuous if f(x,y) is upper semicontinuous in x for each y g Y and lower semicontinuous in y for each x e X. The following generalizations of the minimax theorem belong to Sion [37] and Fan Tsi [11] and are given without any proof below. Theorem 4.16. [37] Let X and Y be convex sets, one of which is compact, and f{x,y) is the function on X x Y that is quasiconcave- convex and upper semicontin uous, lower semicontinuous. Then sup inf f{x, y) = inf sup f{x, y) xexv^r v£Y xex
(4.68)
4.4.
85
DUALITY
Theorem 4.17. [37] Let X and Y denote arbitrary sets and the function be concave-convex on X x Y If for any
f(x,y)
C < infy sup f(x, y)(C > sup inf.Y f(x, y)) y£ xex xexv£ there exists a finite subset A C X(B D Y) such that for any y g Y(x 6 X) there is x* £ A(y' e B) satisfying the inequality f(x*,y) > c(f(x,y*) < c), then the relation ship (4.68) is valid. Theorem 4.18. [37] Let. X(Y) be compact, Y(X) an arbitrary set, and f(x,y) a concave-convex function on X x Y If f(x, y) is upper semicontinuous in x for each y € Y ( lower semicontinuous in y for each x € X), then (4.68) is valid. Theorem 4.19. /[ll]. Let X and Y be arbitrary sets, and f(x, y) a real, almost periodic function on X x Y Equality (4.68) is valid if and only if for an arbitrary t > 0 and for any finite sets A C X and B C Y there are x0 S -V and y0 € Y such that f(a,y0)f(x0,b) <e(a(EA,be B). Specifically, if f(x,y) is concave-convex, then (4.68) is valid. In what follows, gener ation of X-problems will involve the following statements. Lemma 4 . 1 . [11]. Suppose there exists the function X*(u>) 6 X(u>) which provides for each UJ a solution to the problem sup
f(u, s(w)) = /(w, x*(u))
rM6Jf(«)
Then
/.
sup
sup
f(uj,x(ui))dp=
/
f(uj,x(uj))dp
Lemma 4.2. [11] Let f(x,y) be quasiconvex on Y, and for each y G Y there exists the function xj:(y) = sup f(x,y). Then ip(y) is quasiconvex in y on Y xex We now turn to construction of A-problem that is dual to the multistage stochastic programming problem below inf /'i>0(uj0,Xn)dp
(4.69)
n
/
- h M^Xk)dpi_ x
i j
^,
M > lL4 i.( w.k-X ) , t = l.nj
(4.70)
XkeGk{u>k),k=l,n fc,n
k l
Denote A = {\k(u> ~ ),..., i>k(ujn,xn-k,\n-k+1'n) : 4,k+i(u,xn-k+1,\n-k-n)
(4.71)
-1
A7l(a>" )} and introduce recurrently the functions
= (\n-k,bn_k)-
sup *v_k€G„_*
[(\n-k,i>n.k)
- 4%
(4.72)
86 CHAPTER 4. MULTISTAGE STOCHASTIC
PROGRAMMING
PROBLEMS
where ^f\un,Xn) = 4>0(un,XH). T h e o r e m 4.20. [16] Let the function ij)a{ujn,Xn) determining the objective functional of multistage stochastic problem (4.69)-(4.71) be quasiconvex in Xn £ Gn for each ujn G H, and let the components of vector function ^k{uk, Xk), k = l , n , determining the constraints of multistage stochastic programming problem be quasiconcave in Xk G Gk forallujk 6 ttk. Then the functions ^ ( w » , Xn'k, Xn~k+1>n), n k _t k = l , n are quasi-convex in x ~ £ G " for all w" G H a n d >"-*+!." > 0 and concave ifl
^n-i+1,7. > Q
for
y wn
M
g
ft
d Xn-k
M
e
Qn-k
Proof. Let us consider
^ 1 ) K , A ' " - 1 , A n ) = (A n ,fe n )- sup [(A nji /.„)-V'o], xn<=G„
Definitions of ipg ' and Lemma 4.2 imply the statement of the theorem for k — 1. For fc > 1, the statement can by readily proved by induction. We now provide the main result. T h e o r e m 4.21. [16] Let the functions xp0(con,Xn) and ipk(u)h,Xh) determining the conditions of multistage problem (4.69)—(4.71) satisfy the conditions of Theo rem 4.19, and the corresponding Lagrange functionals satisfy one of the generaliza tions of von Neumann theorem. The dual problem (X-problem) then can be written as sup [i>i0n\uJn,Xn(ujn'l))dp (4.73) - n Proof. In the adopted notations, recurrence relations can be written in accordance with Lemma 4.1 and minimax theorem as follows S. K - \ A — ) = sup f{(Xn,bn)*»>0i
sup [ ( A " , ^ ) - ^ ] } ^ . , ^ A'„SG„
= sup / " 4 1 ) K ! X B - 1 , A „ ) ^ _ 1 > ? n _ 1 ( ^ - 2 , X ^ 2 ) = sup [(An_i,V>„-i) - sup / 4'(o] dpl_{\}dplz\
ip / { ( A n _ i , 6 n _ : ) 1>0"(
A
Xn-^G^.i
/
= sup A„_,„>0
{(An_lC,„-l) -
J
SU P
"50O
SUp [(A„_ 1)1 /,„_ 1 )-V'0 1) ]}*n-2 =
A'„_,6G„_, A'„_,gGn
42X"n,Xn-\K-l,n)dpl_2.
/
In a similar manner we obtain
5„_,(u;"- t - 1 ,.Y"- t - 1 ) = sup
/
{(A„-fcA-*) -
sup
[(A„_fc,Vv-fc) - V'o* 0 ]}^" ifc-i =
=
4.4.
DUALITY
87
= sup MH1v,*"~*.A""fc,B)n,\n{u>n-i))dp V»>o-{ - a Hence problem (4.73) is dual to problem (4.69)-(4.71). The conditions of Theorem 4.19 guarantee the concavity of %n\un, Xn) in A" > 0 for all un 6 ft. This completes the proof of the theorem. So if the problem of the form sup
/
[{kn-k, 4'n-k) - 4
(4.74)
(by the stipulated properties of functions V't and ipQ this is the convex programming problem) is solved at each stage, then the above statements enable one to obtain ex pressions for the decision rules Xyxik of problem (4.69)—(4.71) depending on uik,Xk~L and Eu,k+i,„\k,n, that is Xk = fk(uk,Xk-\Ew^,„\k-n)
(4.75)
where £„*+«.» A*'" =
/
\k'ndpnn
Eliminating A from expressions for ?/>Q 0 and solving A-problem (4.73) yield the required expression for the decision rules of the original stochastic programming problem through substitution of optimal values of \n(u>n~l) into (4.75). We employ the obtained results to construct a dual problem to the multistage stochastic programming problem of the form: n
£ „ V o K , Xn) = E„ Y\XjD^)x,
+ a,(u>')] -s- in}
(4.76)
i-i
k
££[**("*. **)l<"*_1] = EAY,
h("k)Xi I"*-'] < M"* _1 ), k = l,n.
A
(4.77)
3=1
Here D,(<J/) is a symmetric positive-definite matrix of dimension n, x n; for almost all realizations of random parameters: Ak}(u> ) is a random matrix of dimension m i x t i j j a;(a/) and bk{wk-i) are random vectors of dimension n3 and mjt, respectively. Ad ditionally, we assume that the matrices Dt(u>') and Ak3(uJk) satisfy the requirements under which all the integrals in (4.76) and (4.77) exist. The function rj?a(u",xn) is convex and the vector functions \ffc(w*,xk) are linear. Furthermore, the conditions of
88 CHAPTER 4. MULTISTAGE STOCHASTIC
PROGRAMMING
PROBLEMS
Theorem 4.16 are satisfied. Hence the conditions of Theorem 4.20 are satisfied and A-problem can be constructed for problem (4.76)-(4.77). Employing (4.72) , we write the expression for i{'0 (w'% Xn — 1, A„): r$\u>n, Xn~\\n) + Y.(XnArnXt
= -Kbn
= -Xnbn + mf JUKAnX,
+ X[DtX,
+ XfDiXi
+ X? DnXn + anXn).
+ a,X{\nAnnXn
+ a,Xt) = (4.78)
i=l
The unconstrained quadratic minimum in the last term of (4.78) is attained at the point xn, where the gradient is zero. Hence, in accordance with (4.74) and (4.75), we have Xn(ujn) = - i f l . - V ' l W ^ ' l A , ^ " ) +
an(un)}T
(4.79)
Substituting the values of Xn(uin) into (4.78), we obtain 4l)(uj\Xn-\\n)
+ an)D-l(\nAnn
= -[Xnbn + l/4(\nAnn
+ an)T} +
T l - l
+ YLi^A^X, + XfD.X, + a,X,). 1=1
Employing the recurrence relations (4.72) and performing transformations, we obtain a dual problem to the multistage stochastic programming problem (4.69)(4.71)
sup £ U 4 " V . A") = ~ x inf El J2 [A,-6,-+ ,\>o
+ l/4-(
Y,
>°
A , A ^ J + i + an_J+1)D;lJ+1(
Y.
J=i
KAi,n-j
+ l+an_J+1)T]
(4.80)
i=TL — j + 1
Let A (a;""1) = {Ai,A2 (w 1 ),...,A„ (u"" 1 )} be a solution to problem (4.80). Then we may obtain optimal decision rules for the original multistage quadratic stochastic programming problem 1 T
*£(«*) = - l / 2 f l ; V R
52\,{u>-1)Alk(u>)
+ ak(uk)
(4.81)
3=k
Let us consider an interesting special case that has been suggested by E. V. Zoy and investigated in [22].
4.5.
89
APPLICATIONS Suppose the objective functional of A-problem (4.73) can be represented as 0(A") = / ^ ( A B + 1 ( w - ) , A > " - 1 ) ) d p 1 n
where t^ is a concave differentiate function with respect to components of A"(OJ™_1) and vector function A n+1 (w n ). Consider the finite-dimensional concave programming problem sup ^{Jn+1,(in)
(4.82)
where A n+1 = / An+1 (u>n)dp,(j,n = {filt ...,fin},i = l , n are the m,-dimensional vectors, n Let pn be a solution to problem (4.82). The structure of the functional V>Q is subject to the auxiliary condition /(VAn+1^o"(An+1(u;"),^),A„+1K)-A„+1)dp = 0 a
(4.83)
We have the following statement. T h e o r e m 4.22. [22] under the assumptions ma.de for the structure of the objec tive functional T/'O , the deterministic vector fi — decision of the Unite-dimensional convex problem (4.15) is a solution to problem (4.72). In general, the deterministic optimal solution is not a unique solution to the problem. Optimal solutions (under the assumptions made) are all the sets of functions A"(aj™_1) > 0 that safisty the equation
Eu
7(n >(A,
t+i > " ) ■
K(un ')) = W >(A,i + i i
LI
) .
~:„, A posteriori decison rules at the fc-th stage xk = Xk(u ) determine an optimal solution at this stage depending on realizations of variables of the problem up to the fc-th stage inclusive. But the optimal evaluations of the fc-th stage conditions Xk = Xk(u>k~l) depend on realizations of random initial data up to the k — 1 stage inclusive. This means that to choose decisions related to estimations of conditions we may restrict ourselves to the information of earlier stage without waiting for the current information to come. J.
4.5
.1 __
j-U „
/.
i L
„-4-,
'. k.A: \
J„J.
Applications of Multistage Stochastic Programming Problems
Many problems of management, planning and design under risk and uncertainty are described by multistage stochastic programming problems. So the problems of longterm planning for economic systems and control of combat operations, issues of ex periment planning and control of space objects can be viewed as multistage stochastic programming problems with statistic, probability and rigid constraints.
90
CHAPTER
4.
MULTISTAGE
STOCHASTIC
PROGRAMMING
PROBLEMS
T h e chance-constrained programming problem can be exemplified by the planning problem of oil production, processing and storage as presented in [2], [36]. We will consider the model by following [21]. T h e planning period can be divided into N stages. We introduce t h e following notations : Xj -the reqired oil o u t p u t at the j - t h stage; Cj -oil production costs per ton at t h e j - t h stage ; Tj -oil transport costs per ton at the end of the j - t h stage ; yj -reserves of oil produced at the end of the j-th stage ; y = l / 2 ( y J _ 1 + yj) -average oil reserves at the j-th stage; q} -oil storage costs at the j-th satge ; Sj -oil sales profit per ton at the j - t h stage ; bj -demand for oil at the j - t h stage ; Suppose the distribution function of random demand for oil produced b3 = bj(ui) is known. T h e total costs of oil production , t r a n s p o r t a t i o n and storage at the j - t h stage are expressed as (cj +r])x] + q]y]. The objective, function can be taken to be the expected profit m a x i m u m for the entire scheduled time
+ rj)xi-qiyi]
\,
7= 1
At each planning stage the available oil reserves are bounded below by a sp ecified value of ymin, while storage capacities are! characterized by ymax. Oil reserves at the end of j - t h period then are J
j
yo + E Xi - ^2 hi 1=1
1=1
Since the earl ier demand (at the j — I —m) is assumed to be known at the start of the j - t h stage , the planning problem of oil production, processing and storage can be formulated as the chance-constrained programming problem. This formulation is characterized by two types of constraints: the constraints determined by the demand for oil >af]
(4.84)
P J Vo + I > < - hi) 2/maxlV< -A-l [ > " f
(4.85)
P\VO + X > . - b . ) > y m i n | 6 i , . . . A - i and those associated with storage capacit ies
where aj and a1- are assumed to be given. In the above notations, the planning model for oil production and storages is designed to maximize the target period profit and can be represented as
4.5.
91
APPLICATIONS N
max E I ^2{s3b, - (cj + r3)x, - qjyj\ ) ,
(4.86)
P{yo +a=i - h > ymin} > aj 1 P{yu + Xj -h< !/mM} > a\(2)
(4.87)
P{y0 + Xl +x2-b,-b2> y mln |6,} > a{2l)(b,) P{yo + XT. + x2 - 6i - h2 < ymax|&i} > 4 2 ) (6i)
(4.88)
P{yo+ E(xi-bi)
> y mirl |6i,..., 6/v-i} > a $ ( 5 i > - . & N - i )
/V
-P{yo+ E ( ^ . - hi) < y max |foi,...,6 W -i} >
(4.89)
a$(bu...,btf-i)
So the optimal solution is determined by a priori decision rules of chance-constrained programming problem. [36] describes a stochastic optimization model for processing oil products and provides a solution algorithm based on the random search method [10]. We will consider multistage stochastic programming problem. The deterministic analog of the model below is discussed in [6], [35]. Let us formulate the planning problem for operation of industrial enterprise in the setting where some parameters are random. The stochastic model can be employed in sectoral planning, and possibly for planning at a higher level. Below are given the basic results obtained in [7]. We employ the following notations: Ak — the m x n matrix of production methods for the fc-th planning period ; bk — the m-dimensional vector of demand for the products produced by the en terprise in the fc-th period ; c —the n-dimensional vector of production method costs in the fc-th period ; xk — the set of possible utilization rates for production methods in the k-th period. Based on economic and engineering forecasts, Ak, bk, ck, xk are determined for periods k = 0,1...., N - 1 The deterministic version of statistical approach to production planning problem is described by the traditional programming model mm(Ck,Xk), AkXk
= bk,k = 0,l,...,N k
(4.90) -1,
(4.91)
(4.92) x-* e X A special feature of the production planning model is that it takes into account the products that are still held in store due to some discrepancy and can be realized in the subsequent planning periods We introduce the residual vector yk+1 for the fc-th period Y1 = b° - A°X°
92
CHAPTER
4.
MULTISTAGE
STOCHASTIC
PROGRAMMING
PROBLEMS
Let y1 < 0, i.e. overproduction has built up stocks T h e residual in the system of constraints for the second planning stage then becomes
Y2 = BxYl +bl
-AlX\
where Bk is the diagonal matrix of dimension m. Nonzero elements 6 | of matrix Bk show what part of t h e i-th product put in a warehouse at t h e beginning of the fc-th period is still suitable for sales at t h e end of this period for Yk < 0 and what part of the demand for t h e p r o d u c t t h a t has not been satisfied in the k — 1 period refers to the fc-th period. If operation of the industrial enterprise is planned in such a way t h a t for realization of a r a n d o m parameter ui the demand is satisfied, i.e. constraints are imposed on the residual Y < 0 (almost everywhere ), and storage costs of finished products in t h e fc-th period are represented by the m - t h vector gk (which is also r a n d o m ) , t h e n t h e linear multistage stochastic model for long-term production planning takes the form E ( ] T CkXk\
Yk+i
_
-AkXk
Bky^ k
DX
k
-+ min,
+ bk,k = 0,1,...,N
< 0(almost Vu G $7)
Y k
- E ( £ qkYk)
k
= d ,X
k
> 0 , f c = 0, {,..., N -1
(4.93)
-1
(4.94) (4.95) (4.96)
Constraints on utilization rates of production methods are specified by the rela tions (4.13). T h e m a x i m u m principle which can be applied to solution of deterministic discrete dynamic problems is discussed in [1], [24]. T h e stochastic case is discussed in [22], [23], where only t h e necessary optimality condition for solution was obtained. In complex situations, decision generally is selected under risk or uncertainty. Time shortage does not permit a delay in decision making until complete information on problem conditions becomes available. Sequential collection of information, however, enables one in time to correct and slowly improve the decision. T h e rational pre liminary decision, and intermediate corrections, must take into account t h e predicted values of statictic characteristics of random parameters t h a t have not yet been re alized. Formal models for analysis of complex dynamical systems are provided by multistage stochastic programming problems with conditional or unconditional, sta tistical, probability or rigid constraints.
Chapter 5 Game approach to Stochastic Programming Problems 5.1
Game formulation of Stochastic Programming Problems
In practical problems of stochastic programming, joint distribution of random pa rameters is not always adequately determined. Replacement of random parameters with mean values, if these are known, and computation of optimal solutions to the deterministic problems obtained are not necessarily justified, because adequacy of tha model to the phenomenon under study can be violated by averaging parameters of problem conditions. In such cases game formulation of stochastic problems may prove to be effective. A decision maker, who has no adequate information on the decision situation, admits occurence of various outcomes resulting from each decision he may take. In actual practice, occurence of various outcomes is governed by some unknown natural laws. For this reason the decision maker may admit that the true laws of the Nature are the least favourable for him. This means that the decision maker regards the situation as if he is faced with not the objective, but unknown Nature, but with a conscious opponent seeking situations that are the least preferable to him. In this respect the Nature can be regarded as a conflict participant. From the above reasoning it is clear that the stochastic programming problem can be viewed as a two-person zero-sum game [3], [4], [13], in which the players are: the decision maker and the Nature. The reward function is determined by the objective function and the amount of penalties for violation of constraints. Construction of the reward function reflects in some form the information available to the manager. In the setting, the payoff function determines whether the stochastic programming problem will be solved in pure strategies or in mixed ones. The manager's optimal strategy determines a solution to the stochastic problem. The game approach can be successfuly applied to analysis of stochastic programming problems where there is
CHAPTER 5. GAME APPROACH
!)4
TO SPP
little or no information on joint distribution of random parameters. Suppose the stochastic programming problem is generated by the following extremum problem max(C*,A),
(5.1)
AX < b
(5.2)
A >0
(5.3)
where A = (atJ);b = (bi);C = (c ; ); A = (xj),i = l,m;j = l,rc. Although the notation (5.1)—(5.3) is completely determined when the values of parameters are determined, it becomes indeterminate and requires additional clari fications if the elements atJ of matrix A, components 6, of constraint vector b and components c3 of vector of linear coefficients(or some of them) are random values. To formulate the stochastic programming problem, one has to find out what is meant by the objective function, how the constraints must be interpreted and what are the strategies (pure or mixed) in which the solution to the problem must be computed. Let random values be the nature state functions u; £ ft, where ft is the set of all possible states of the Nature. As in [4], the stochastic problem can then be viewed as a, two-person zero-sum game in normal form G = (M, N,g) Player 1 is the decision maker whose strategies are the vectors X > 0 belonging to a properly chosen set M C E+(E£ is a nonnegative orthant of n-dimensional Euclidean space). Player 2 is the Nature whose strategies are the sets of triples (A(to), b{io), C{LO)). The set ft determines the set N in the Euclidean (mn + m + n) space corresponding to a feasible range of variations in the elements aij(uj),bi(u>), C,(CJ) of the problem conditions. The decision maker seeks to maximize the average payoff during the game, while the Nature (his opponent) selects its states so as to minimize the average payoff to the decision maker. For each state of the Nature and each choice of strategy A' by player 1, the payoff function is given as the sum of appropriate values of linear form and penalty for violation of constraints n
g [A, A{w), 6(w), C M ] = £ c^)x3 j=l
m
+£ > «=1
I n
\
£ a, J (^)x J - 6 , H \j = \
,
(5-4)
I
where tpt(Z), i = 1, m are continuous nondecreasing functions satisfying the condition cpi(Z) = 0 if Z < 0. In actual practice, the specific form of penalty function
5.1. GAME FORMULATION
OF SPP
95
In the case of infinite games, if the opponents have continuum-equivalent sets of pure strategies and the payoff function is continuous and concave-convex , then the game has optimal pure strategies. But representation of the game as the one on unit square does not reflect specific features of the structure of pure strategy sets and may result in the loss of many properties of the payoff function. Preservation of the natural structure of strategy sets does not guarantee the existence of optimal pure strategy, since according to [4] the set of pure startegies for Player 1 is required to be a subspace of some linear space, whereas in the problem under consideration the space is not linear, besause one of the linearity axioms (existence of the opposite element) is not satisfied. So in the majority of cases the game model of stochastic programming problem does not guarantee the existence of the solution to the original problem in pure strategies. There are meaningful formulations of extremal problems, e. g. , the problems related to recurrent situations the solution of which can be discussed only by extending the domain of problem definition and interpreting the feasible optimal solutions to mean not only deterministic vectors, but also vector distribution. In such cases the reachable maximum of the objective function can only increase, while the reachable minimum can only decrease. Mixed strategies are assumed to make sense in the meaningful terms of the prob lem. Denote by S the set of mixed strategies for a decision maker, i. e. the set of distributions Fx of vector x defined over M and by F the set of mixed strategies for the Nature, i.e. the set of joint distributions FAbc of matrix A(u>) with vectors b(w) and c(w) defined over N. When the distribution of part of the parameters is known, one has to consider only those mixed strategies in which the distribution of these parameters coincides with the known one. Suppose they form the set F C F. In the adopted notations, the game formulation of stochastic programming problem can be stated as follows: it is required to compute mixed strategies F* 6 S and FAbc € F such that max min
FxeS FAbc€F
/
g{(, a,0,-f)dFI(()dFAbc(a,P,'y)
J MxN
= min max FAbc^F Fi£S
/
g{(,a,/3,f)dFx{()dFAi,c{a,P,j)
=
=
J MxN
= J 5(£,a,/?,7)dF;(0<^;t>,/3,7)MxN
Under sufficiently general conditions (the compactness of the sets M and N) there exists F% 6 S and F*Ahc € F on which the value of the game is reached [3]. If the distribution function FAbc is known in advance (problem under risk), then the set F consists of this single element.
CHAPTER 5. GAME APPROACH
')(:,
TO SPP
If the joint distribution of parameters FAbc is not known (problem under uncer tainty), then F = F is the set of all possible distributions defined over N and the solution Fx of stochastic problem defines the mixed strategy. Each finite two-person zero-sum game has a solution in mixed strategies, and hence original stochastic problem has a decision distribution. For the infinite game, the existence of optimal mixed strategies alone is not enough for determining decision, because one has to know whether these are strategies of finite order. [3] shows that if the set M is compact, while the set N is convex and compact and the payoff function g(X, A(uj),b(u)),c(uj)) is continuous in its variables and convex in (A(Lj),b(u),c(u>)) (or concave in X), then Player 1 must have the optimal mixed strategy of order not higher than {(mn + m + n) + 1]. In addition, if Player 2 has the jD-dimensional set Y of Player l's optimal strategies (i. e. pure strategies), then Player 1 has optimal mixed strategies of order not higher than [(mn + m + n) —p+1]: k
F'x-
= l,A,->0j 1=0 i=0
\\ii==00
where k = [(ran + m + n) — p + 1]. IX{>) is mixed strategy of first order (i.e. pure strategy) which permits selection of A''' with probability 1, where as the first order mixed strategies /A*,4*,C* with some [A*, b*, c") £ Y are all optimal for Player 2. Since the payoff function is of the form
g[X,AH,&H,cH] = £ C j H x J + ^ > !
^2a.ij(bj)xj
- bi(
its convexity or concavity is determined by choosing the payoff fonction. Suppose the joint distribution FMC of random vector (A,b, c) belongs to the dis tribution set F: define for all x £ E* and Fx £ F:
g(X, F) = EFIJ2 CjXj + Y, VJ, £ at]x3 - b{ where ipi(Z),i — l , m are continuous nondecreasing functions that satisfy the condi tion
5.2. SPECIAL
CASES OF THE GAME G{E+,F, G)
97
{F* : Fx € F})). Let the functions tfi(Z), i = 1, m be integrable {with respect to Fx € F) for all X £ B+. Then sup
= min
mm g(X,Fx) x£F
F
X€Et
convex and uniformly
sup g(X. Fx) xeE+
FxeF
The equivalent assertion for concave penalty fuiictions is g iven below. T h e o r e m 5 . 2 . If the set F is convex and compact, c , are uniformly integrable (with respect to Fx € F) and the functions ?, (Z) are un.iformly (in Fx * E F) integrable for all X E E *, then
inf max q(X, Fx) = max inf
q(X.Fx)
where g(X, Fx) = Epx< 12 Cjij + Yl Vi( T. aijX3 — 6») >, which corresponds to rel J=l
j=l
.= 1
placement of the original problem max(C, X), AX < b, X> 0. by the equvalent problem min — (C, X), AX < b, X >0.
5.2
Special cases of t h e G a m e G(E+, F,g)
Let
= v,Z+, Vi > 0,i = 1, m
where Z+ = \(\Z\ + Z). We will consider the game G(E+, F,g) for some special specifications of the set F. Since the conditions imposed on random values of stochastic problem are natural, the random values in practical problems generally have a restricted feasible range and possess the mathematical expectation. We now consider some special cases of the game G and point out some ways to its solution.
98
CHAPTER
5.
GAME
APPROACH
TO SPP
In the game G corresponding to the stochastic problem, we assume t h a t a;,-,» = l,m,j that
— ]. , n are constant values, &,, i = l , m , are independent r a n d o m variables such
0i,0i
E(bt) =
and Cj,j =■ l , n are random variables in which E(c3) = "/3 are determined for VFx G F. Denote ? by A{I,J, k} an arbiicrary partitioning of t h e set { l , 2 , . . . , m } into three parts, one or two of which can be'■ empty. Denote : by Ai/,v = 1,M their partitions for which t h e sets
{*>0:_
n
TL
n
< £ aijxi
< Pi', i € J „ ;. £
a
i j x j > A ! >'
e
3=1
3= 1
1
^ r (5.5) J
are nonempty. Let
(5.6)
*{X)=mmg{X,Fx)
Since 6; are independent random values, (5.6) can be! transformed t o the form Vi m a x EFx -E F €F •=i
*(*) =
X
3=1
I
(l>
««, -
&,
)
\
where t h e terms of the :second sum can be estimated as follows r / „
max
FX£F
%
\ +i
/ _
\ +
? = >< X > » J ^ - #
E V J - M
\
1 _
X
+ (! - i) E "ijXj ~ Pi. (5.7)
where
Thus we have m
TI
(
/ n
\
/ n
A
v(x) = 3E= 1-m - E"< «- \£ «*■*; - fl + (i - A,-)\ i =El <w --Ai=l 1 3=1 / In accordance with the partitioning, we have for i g Jw :
/
n
ME«w-ft \i=i
\+ /
+(I-A,-)
/"
\+
\3=i
/
E^-/?]'
=
-- A> I E ahxi - 0i I + (! - A>) I E a'ix3 - 0i E "Hxi - W - W - 0'i)] = E ai}x} - 0,. 3=1
j=l
\+l /J
■ (5-8)
5.2. SPECIAL CASES OF THE GAME G ( £ $ , F, G) for i £ J„ : At
1 n 1 2_^, 0.t]Xj --Pi)
\+ /
\J = 1
99
\+
/ > + ( 1 -- A , ) \Y.a'Jxi
■-Pi
\3=l
/
= K I E aax3 - 0, for i £ ku :
^Vtavxi-0i)
+(!-<*;) E w A "
=0
which follows from the condition /?; < 01'. Let us write for the arbitrary partition i/\&(x) :
*(*) = £ 7 ^ - 1 > .
*(.Y) = - E
E%*i - ft - E "A E <**, - A'
-7; + E »j + E »>xj )xi + E ^ft + E v^iPii=I„
3=1 \
<=J„
/
i=I„
i=J„
Introduce the notations
?„, = —7j + E t / , a ^ + E "t^tOyjfci/ = E "'A + E ^A-ft; f = l,Af,j = l,n,G =te„,),A: = (*„)■ If £(6;) = 0t instead of 0{ < E{bi) < 0{,i = l,m, and all the other assumptions are retained, 0, is replaced by 0{ in the definition of A, and kv. In the game G(a,i,. . . , a,n, &;), i = \,m , we assume that mutually independent random vectors are such that a't] < an < aVj.alj < a £ ;
Pl]
E{k1) = 0%] = \{0l]+0'i); i = l^rn, j = 1, n. and c,,j = l,n are random variables such that E(CJ) = jj
100
CHAPTER 5. GAME APPROACH
TO SPP
In addition, let A„ = (/„, J„,kv,), v = 1,M be those partitions {1,2, . . . , m } for which the sets
[x > 0
It
71
•*j > ft
K,^ > ft, 2
6 J t .;
n
E a i j ^ < ft>2 6 fc.>} are • nonempty As in the precedi ng case, for an arbitrary partitioning v we obtain the function .V) and int roduce notations: = ~7j + E
% rZv —
ud
'
^'rf«ji
E "'# + 9 E ".-/Sii f = Tim.
Finally, in the game G, Lat (A, 6, c) be a random vector such t h a t o ; 3 <
ct'i),
ft, < K < p-y. %
j = l,n,i
Suppose A = (lyfJu),!/ sets
< % ,
=
l,m.
= \,M are all those partitions {1,2,. . . , m} for which the
X > 0 : £ or" x, > (]•, i £l,.:J2
oe>,x, < # , i € J , 1 ;
are nonempty. Then, in much the same way as before, for the arbitrary partition u we obtain the function ^(X) and introduce notations:
qVj = -7j + Yl "■°;;;^ = Y2 "■■#;" = u f j = TTR. ie/„
.6/.,
Since the above special cases reflect only those partitions for which respective sets are nonempty, M < 3 m , M< < 2 m , where m is the number of constraints for the original problem. For the above cases the relations [ sup minig(X, Fx) = min sup
XeB+Fx€F
FxeF
XSB+
g{\\Fx)
5.2. SPECIAL
CASES OF THE GAME G{E%, F, G)
101
are valid and X is the solution of the game G = (E+, F,g) if and only if {X,Z0)is the solution of the linear programming problem max Z, QX + Zt< K,
(5.9)
X > 0, where e = (1,. . ., 1). In the special cases, the game solution of stochastic .linear pro gramming problem is equivalent to the solution of deterministic linear pro gramming problem which generally has a larger dimension than the original stochastii; problem. Suppose the payoff functions are convex and have the form f,(Z+),l=L^
where Z+ = \{\Z\ + Z), and for 'i(X) = min we have ^
FX£F
f {X) = min EFx < E c J s i ~ £ / * ( E atJx3 - 6, 1 From the independence conditions of vector components b = (&i, 62, ■ ••, bm)follows 1 the independence of functions / , ( Z + ) . Then we have n
m
Since by the condition, f,(Z+)
\
' I n
* W = EH X J-Ejgfj E Fx
I
-MEa'Jx>- M |
are convex functions, the inequality (5.10)
EFx{f,(Z+)}>f,{EFx(Z+)}
holds for each Fx 6 F. In view of the relations (5.7), (5.10), and by the nonnegativity of fund ;ion values
Mz+), m
{
\
( n
+
\ +\
( n
W < £ w - L / ' ' W E v r M +(i-Ai) rjZoijXj-bix or, in view of partitions, (5.11) *(*) < E w - E /■ (£«w -ft) - £ fi Ucti*»*i - |ft)) ]=i
;eJ*
\j=i
/
-eh
\
3=1
,1
Denote the right-hand side of (5.11) by 4>j(x). The function $(2) is concave on £+ and concave on each set determined by (5.5). These sets correspond to partitions Ai/,i/ = l , m and form disjoint partition £ + . From this it follows that V(X) < min ~4>v{x). v _ '
Kv<M
"
'
102
CHAPTER 5. GAME APPROACH
TO SPP
Maximization of * ( X ) is then equivalent to maximization of some concave function 4>o{Z) subject to 4>0(Z) < ${X), i. e. UZ)
< ~4>y(x)
denote
~4>AX) -MZ)
= 4>„{x,z).
This brings us to the deterministic convex programming problem (5.12)
ma,x(f>o(Z), 1,M
<j>„{X,Z) >0,f-
(5.13)
X 6 Et If (X*,Z*)
(5.14)
is the solution of this problem, then X* is the solution of the game
G = {E+,F,g). The implication of the duality theory for convex programming problems is that for (X*,Z*) to be the solution of problem (5.12)—(5.14) it is sufficient that there be numbers £/*, E/J,..., U^ satisfying the conditions
U:MX\Z-)
= O,
ut > o M
L{X\ Z\ IT) = sup[L(A', Z, U') = MZ) + £ VfaiX,
z
)\
where L is the Lagrange function. In this case the differentiability requirements are not imposed on the functions
MZ)zndMX,Z). In the above special cases we determined some matrix Q. By this matrix one can say whether there is a finite optimal pure strategy for Player 1 in the game Suppose there is such a strategy. Then problem (5.9) has a finite optimal solution and duality theory permits construction of the following problem: minA" T t/,
(5.15)
QTU > 0,
(5.16)
M
E u» = !.
(5.17)
Uv > 0
(5.18)
5.2. SPECIAL CASES OF THE GAME G(£+, F, G)
103
By the duality theory, the optimal solution set for problem (5.15)-(5.18) is nonemp ty. But the conditions (5.16) and (5.18) then imply that the matrix Q contains no strictly negative column (necessary condition) and has a nonnegative row (sufficient condition). To check the sufficient condition there is no need to construct the matrix Q for all partitions, it suffices to consider only one of them: A 1 = (/ 1 ,J 1 ,A- 1 ) = { ( l , 2 , . . . , m ) , 0 , 0 } If the corresponding set {.V > 0} is nonempty and qh > 0 for Vj = Y/n, then there is an optimal pure strategy. Implementation of the game approach to stochastic programming problems in cludes several stages: Stage I. Game formulation of stochastic programming problem. The governing condition in this stage is provided by the available information on the stochastic programming problem. The initial information enables one to deter mine: a) how the solution of the problem is to be interpreted; b) the opponents' strategy sets; c) the function of penalty for violation of problem conditions; d) the payoff function. Stage II. Analysis of the resulting game. In this stage, one has to consider the existence of the value of the game and that of optimal strategies. Stage II focuses on the following situations: a) pure strategies versus pure strategies; b) mixed strategies versus mixed strategies; c) mixed strategies versus pure strategies; d) pure strategies versus mixed strategies; for finite and infinite games (with uncountable number of strategies). Although transi tion to the game on unit square for obtaining "pure strategies versus pure strategies'' situations guarantees the existence of the game sulution in pure strategies, it may disturb the conformity of the resulting game to the original problem. In the "pure strategies versus mixed strategies" situations, the choice of convex payoff functions is equivalent to that of concave payoff function for the original problem and sufficient condition for existence of player l's optimal pure strategy are furmulated for the case of convex payoff functions. Stage III. Method of solving the game. In the general case, the simplex method and Brown-Robins iteration method are applicable to finite games. If the payoff function is continuous, Brown-Robins method can be applied for solving infinite games. This method is effective where the games with large numbers of pure strategies are involved, whereas any other method requires substantial computations. The main drawback of this method is the low convergence rate of iterative process. The current methods for improving convergence are related ro specific problems.
104
CHAPTER 5. GAME APPROACH
TO SPP
In the general case, application of these methods may substantially complicate the solution procedure and make it inefficient. For Stage III in this chapter, we considered some methods of solving infinite games. It was shown that, for the convex payoff function, the infinite game can be reduced to the deterministic convex programming problem.
Chapter 6 Existense of Solution and its Optimality in Stochastic Programming Problems 6.1
Dual Stochastic Linear Programming Problems
The issues of necessary and sufficient conditions for optimality of the solution and its existence play a very i m p o r t a n t role in stochastic programming problems. We will concentrate on some properties of dual stochstic linear programming problems. Below is given t h e result ("weak duality t h e o r e m " ) which was proved by Tintner [8]. Consider t h e problem min Z = min xes(w) x<=s(w) where S(w) = [X
: A(LJ)X
G(u)X
> 0 , u e n } with r a n d o m parameters and its
< b{u),X
dual problem with t h e functional denoted as g = b(ui)Ty. Let gl, be t h e value of this function for ui £ fi at t h e l-th extreme point of the dual problem. We have t h e following theorem. T h e o r e m 6 . 1 . Let the point 1-J: be the optimal extreme point for UJ0 £ H and let L^0 be the corresponding
value of the objective
region for this solution.
Then for any u £ W*0 there is an extreme
of the dual problem
function.
Denote by W^0 the
such that g^ = L% and there is a neighborhood
poinnt
stability (e.g., 1)
0{oJ) of point LJ
such that for any ui G 0(uJ) J
7~k
Proof. At t h e point us t h e direct problem has t h e solution t h a t can be reached at the e x t r e m e point L-^. Therefore its dual problem also has a solution. Let lj be the 105
106
CHAPTER
6. EXISTENCE
AND
OPTIMALITY
optimal extreme point of the dual problem. By t h e duality t h e o r e m , 7* = Zi^
J g^j
To prove the second part of t h e theorem, suppose the opposite is t r u e , i. e. there is no neighborhood 0(uj) in which the extreme point lj is optimal, although it is optimal in t h e point uJ itself. Note t h a t for any u> 6 W*0 the dual problem has a solution because the direct problem has a solution in the same domain. Consider an arbitrary neighborhood 0{p) € W% of the point UJ. By the optimality of extreme point lj
(6.1)
gl-gl>o
for any I = 1 , 2 , . . . , La, the L^ is the total n u m b e r of e x t r e m e points in the dual problem. On the other hand, under our assumption,
I ~I < 0
(6.2)
Two cases are possible for some u) € 0(u>). 1. T h e number of points, in which the relation (6.2) is valied, is finite Therefore these points can be enumerated as i = 1,2,. . . , p. Denote by pi(w, Ui) the distance between t h e point 57 and the point a>; and find p = mill,-p,-(57, u^). T h e required neighborhood, in which the equality Z* = glu is valid, can be taken to be the set 0(T3) = {UJ :p(UJ,uj)satisfying the inequality (6.2) is infinite (countable or uncountable). Since the space D, is a topological space, t h e r e exists a system of neighborhoods of the point 57. Denote by On(to) t h e neighborhoods of this system. In each neighborhood On(tJ) there is at least one point u>, satisfying t h e conditions of (6.2). Consider Dn{ul) = On{U) - O n+1 (S7). T h e r e are two possibilities: a) the number of neighborhoods Dn(Uf) containing t h e points CJ,- is finite; then there is a number A^ such t h a t for all n > N 0n(uj) contains infinitely m a n y points U»J; from these points it is possible to construct the sequence w,-n connverging to the point w, i. e. toln —¥ u. By the continuity of t h e function g1 of w, g' —> g1-; > 0, it ' " n—foo
follows t h a t g —g —)■ a1—— ql- and hence, under t h e assumption t h a t the terms of the sequence are smaller than zero, whereas by the relation (6.1) its limit is greater t h a n zero. This contradictioon proves t h e theorem for the case under study; b) the number of neighborhoods Dn(uf) containing t h e points u>, is infinite; this also permits determination of the sequence of points u>,n converging to w. T h e proof can be completed in much the same way as in the preceding case. In the previous chapters, dual problems were investigated for various models of stochastic linear programming problems and in separate cases analytical criteria were formulated for optimality of solutions. Below are given some findings concerning the issues of existence and optimality of solutions in problems of general form.
6.2. OPTIMALITY
6.2
AND EXISTENCE
OF THE SOLUTION
107
Optimality and Existence of t h e Solution in Stochastic Programming Problems
The issues of existence of the solution and its optimality for stochastic programming problems are discussed in Hanson [1]. We will consider some of the results presented in this work. Suppose the vector Y £Y C Em has a scalar distribution ij){Y), that is continuous and twice differentiable. Let X 6 X C En be some vector minimizing the mean value of the scalar function <j>(X, Y) under two types of constraints: 1) the mean value of the function g(X, Y) in nonnegative for g € Er; 2) the function h(X,Y) is nonegative for h G Es. Both types of constraints are important from the point of view of practical appli cations. For example, the former constraints can be viewed as long-term contractual requirements where assumption is made for deviations which, however, in the aggre gate cannot be vioalated. The latter constraints may reflect some physical constraints of the system which cannot be violated. Let us formulate the following problem min /4>{X,Y)4'{Y)dY
(6.3)
Y
subject to Jg{X,Y)4>{Y)dy>0,h(X,Y)
>0,X
G X c En,
Y
where, g, h are continuous twice differentaible functions given on S = X x Y. For this formulation of the problem, [1] provides formulations of duality theorem and theorems of necessary and sufficient optimality conditions for vector X. We will provide the proof of sufficiency for optimality of the vector. Note that the conditions stated in the theorem are too strong, which substantially restricts the range of problems under study. T h e o r e m 6.2. Let the above-stated function (X,Y) be convex in X and the above-stated g(X,Y) and h(X,Y) concave in X. Fulfilment of the following conditions is sufficient for X to be the minimizing vector of problem (6.3): 1) there exists A < 0; 2) there exists ft(Y) < 0; 3) 4>x{X,Y)^{Y) + \gx(X,Y)iP(Y)+fi(Y)hx(X,Y) = 0; 4) Xg(X,Y) = 0; 5) /i(Y)h(X, Y) = 0, where <j>x = (-§£-;...; j$;),gx,hx are the matrices of rele vant dimensions composed of partial derivatives, while \,fi{Y), are the r-dimensional vector and S'-dimensional vector function, respectively. Proof. Let X* and X be some feasible solutions. Since <j>(X,Y) is convex in X, continuous and differentiable, all the conditions (l)-(5) will be utilized for X from
108
CHAPTER 6. EXISTENCE
AND
OPTIMAL1TY
the theorem. Thus we have: j <j>(X*, Y)i>.{Y)dy > j (4>(X, Y) + (X' - X)x(X, Y)) 4>(Y)dY = Y
Y
= j 4>(X,YMY)
- (X* - X){\gx{X,Y)MY)+t±{Y)hx(X,Y)\
dY
Y
which, by the convexity of functions Xg and /j,h, is not less than > j(X, Y)HY)dY Y
- j Xg{X\Y)
- g(X, Y)^{Y)dY
Y
- j n(Y)(h(X\
Y)-
Y
-h(X, Y))dY = j{X, Y)xP(Y)dY - J Xg(X\ Y
Y)4>(Y)dY-
Y
- f n(Y)h{X*,Y)dY
> /
Y
cj>(X,Y)$(Y)dY.
Y
The latter exactly means that X is the optimal solution. The duality theorem below provides a means of establishing lower limits for the objective function values in problem (6.3). Theorem 6.3. TfJ>(X,Y),4>(Y),g(X,Y),h(X,Y) are the functions defined in Theorem (6.2) and if X is the optimal solution in the direct problem (6.2), then X is also the optimal solution in t i e dual problem: max j (4>(X, Y)i>(Y) + Xg(X, Y)i>(Y) + f*(Y)h(X, Y)) dY Y
subject to j (MX,
Y)4>{Y) + \gx(X,Y)il>(Y)
+ n(Y)hx(X,Y))
dY = 0
Y
> < o , K y )wiere t i e optimal X and ~p(Y) satisfy conditions (l)-(5) from Theorem 6.2 and tie optimal values of objective functions in both problems are equal. We will consider another existence theorem that has been proved by Stoddart [7]. This theorem is applicable to a sufficiently general system from which the problem (6.1) can be derived as a special case. Consider the probability space (A,S,n). Suppose the probability measure fj, is ordinary with respect to some topology over A. Let R be the closed space in En, and U the closed convex set in Em. Consider the fixed measurable mapping r : A —y R. The real-valued function f(r, u) on R x U is linearly bounded below in U if the inequality f(r(u),u(u)) > p(u) + ug(ui)
6.3.
109
INVESTIGATION
holds for some integrable function p and bounded integrable function g on A. Consider a number of real continuous functions ;(r, u), ihj{r,u) cm R x U each of which is linearly bounded below in V. Let F be a class of all integrable mappings u : A —>• Em such that 1) U(u>) e U for almost every OJ G A; 2) /gi(r(u>),u(u>))dij. < 0 for each i; A
3) ft.,-(r(w), a(w) < 0 for almost every u £ A and for all j . Furthermore, let <j>(r, u) be some real continuous function on R x U. The solution uQ 6 T is then said to be optimal for <j> in T, if/(it) = f d>(r(ui, u(u>))dn has a minimum in the point U0, i.e., the inequality I(u0) < I(u) holds for all u 6 F. Note that we have the system that is equivalent to Jenson's system if A C E„, r(w) = w and ji = J tpdX, where ip is probability distribution toward Lebesgue measure A on En. The proof of the existence theorem for the nonempty F is based on the assumption of closure and compactness under weak convergence in L\ = Li(A,S,fi). That the function I(u) is semicontinuous below under this convergence is sufficient for existence of the optimal solution in F. T h e o r e m 6.4. Let <j>(r,u) be the continuous real function on R x U that is linearly bounded below and convex in U. Suppose f \u(w)\d)j, is bounded on V and A
e-absolutely ofF. Then, if F / 0
I(u) = J (j>(r(u),u(w))d}i A
has a minimum on T. The key condition in this theorem is provided by the boundedness and e-absolute continuity of / \U(u>)\dfi on T. A
In Stoddart [7] two theorems are formulated to meet this requirement under some conditions. This is true when some real function(u) on U is bounded below and is such that f}U> as soon as \U\ -> co on U, and if f ip(u)dn, is bounded on T. \U\—foo
j^
It is also proved that Theorem 6.4 is true even without assuming the key condition if we take p,-(r(w),u) > i>{u), where ip(u) has all the properties stated above.
6.3
Investigation of one Stochastic Programming Problem
Consider the following economic problem. Suppose the production of products has input values X = ( s i , a 2 , . . .,£„) and provides flexible responses to such random factors as demand, weather and failure of equipment. That is, there are plans for production under various conditions whose implementation is random and continuous.
110
CHAPTER 6. EXISTENCE
AND
OPTIMALITY
We need to find the optimal set of input values X by minimizing the mean value of some cost function under the following constraints: 1) average long-term contractual obligations (supply of raw materials, production of products, profit, etc. ) must be generally met, although short-term requirements can be violated; 2) general physical constraints of the system cannot be generally violated and must be satisfied at each random realization of problem conditions. Suppose the functions presented in the problem can be nonlinear. Then we have the problem formulated as in Hanson [1]: min /(X,Y)ip(Y)dy {Y}
subject to (6.4)
j g(X,Y)j>{Y)dy>Q {Y}
(6.5)
h(X,Y)>0,Vy
where <j>{X,Y) is some scalar function of X 6 En and Y £ Em,g(X,Y) and h(X,Y) are the vector functions from Er and Es, respectively, cj>(X,Y) and components of g(X,Y) and h(X,Y) as well as probability distribution ip(Y) are twice continuously differentiable on S = {X} x {Y} C En x Z5m in all their variables. As indicated above, Hanson provides necessary and, under convexity of the func tion <j>(X,Y) and concavity of components of vector functions g(X,Y) and h(X,Y) in X, sufficient conditions for optimality of vector X These conditions seem to be too strong, because in the system M~X,yWY)
+ Xgx(X,y)4>(Y)
+Jl(Y)hx(X,y)
= 0,Vy
(6.6)
each summand depends on the vector-parameter y. We will show that these condi tions can be weakened without loss of generality by elementing y from some of the summands. By the properties of functions, g and if), problem (6.1) can be rewritten in equivalent form: find vector x which carries minF(i)
(6.7)
subject to G(X)>0,
fc(jf,y)>o,vr,
(6.8)
(6.9)
where the function F(X) and vector function G(X) are computed by the distribution {Y}, integrals / <j>{X, Y)i>{Y)dy and / g(X, Y)^{Y)dy, respectively. F{X),G(X) {V}
(Y)
and h(X, Y) are also twise continuously differentiable. In this problem, the vec tor function h alone is dependent on the vector-parameter Y To reduce problem
6.3.
111
INVESTIGATION
(6.7)—(6.9) to a parametric nonlinear programming problem and derive necessary and sufficient optimality conditions for vector X, the following conditions are said to be satisfied: if r + 5 > n, then at most n components of functions G and h can vanish at any point of the set S. The components Gj(j = 1, J) and hp(p = 1, P) vanishing at some point of the set S are such that the matrix dGj_ dGj_ dhy_ dhp_ 5 X ; ; ' " ; dXi' dX^']~dX~, has a maximal rank. Below is given the lemma which permits derivation of necessary optimality con ditions for vector X. L e m m a 6.1. Suppose: 1. there is a region gr{X) > 0,r = 1, M, where gr are reaj functions defined over {X} C En ; 2. gr(X) for any r is twice continuously differentiable on {X} C En; 3. there is a vector X such that gr(X) = 0,r = 1, m; m < n (*) gr{X) > 0,r = m + l , M ; 4. ./;x~(l, ■ ■ ■ , m) =£ 0 for m inequalities of the form (*), i. e. the Jacobian has a maximal rank at the point X. Then for any i 6 l , m there is a point X1 such that ft(X') > 0, whiie gr(X') = 0 for any r = T~m, r ^ 1 and gr{X%) > 0, r = m + 1,M. Proof. 1. Since the functions gr are continuous, for any r 6 m+ 1,M and any e r > 0 there is a vector SCr > 0 such that VX 6 \X - X~\ < Ser,\gr{X) - gr(X)\
< er.
Hence, since gT(X) > 0,Vr = m + 1, M, it is possible to choose Str such that for any
xeSSr
_ 0 < gr(X) - e r < gr(X) < gT{X) + er.
If one has to select min<5..1
=
°roi0,
where 5raia is the l-th coordinate of the r-th vector (the coordinates are all strictly greater than zero), and construct the n-dimensional vector whose coordinates are Jr„i0, then for any x from such 5roi0 neighborhood the inequality gr{X) > 0 holds for any r = m + l,M. The reasoning below applies only to vectors X from the neighborhood denoted as 5{X). 2. Consider gr(X),r = l , m on S(X). Fix some i€ l , m and consider the sys tem of m — 1 equalities gr(X) = 0 for any r G l , m , r ^ i. For convenience, the functions are enumerated to fix i = m. From the m-th Jacobian (say, the first m columns of the matrix composed of partial derivaties) we delete row m and column,
112
CHAPTER 6. EXISTENCE
AND
OPTIMALITY
in which an element -J^- is such that its cofactor is not equal to zero. This choice is made possible by the nondegeneracy of the m-th Jacobian. The nonzero cofactor of element jf^- is the Jacobian ./jf(l,..., m — 1) of functions gT(X),r = 1, m — 1. Since the functions gr{X) are continuous in the neighborhood S(X), gr{X) = 0,r = l , m — 1 and ^ r ( l , . . . ,m — 1) is a nonsingular matrix, by the implicit function the orem there exists an ^^-neighborhood of the point X = (x m , z m + i , . . . , i n ) C ^n-m+i e n t; r e ]y contained in i X of a suitable dimension such that for any point X = (xm, xm-i,.. ., i n ) from i(X) there are unique continuous functions ipi(X),..., ^ m _ i ( J ) , possessing the following properties: a) Z r = fr(X~), r = l , m - 1; b) for any -Y € J(A') the value A' r ,r = l,m — 1 computed from the formula Xr = fr(xm, xm-i,. . . , i „ ) together with the components of vector X form the vector X satisfying the equations gr(X) = 0; c) in £(X) the functions <pr(X),r = l,m — 1 are differentiable and the derivaties -g^-1 for the given /, I = ra,n are the unique solution of the set of linear equations: m l
- dg'dVr
dg, .
S^;^ -^il = 1'm-L =
( }
**
m
m
3. Let us show that for gr(X),r = l,m there is X such that gm(X ) > 0, and all gr(Xm) = 0 for r = l , m — 1. Suppose there is such Xm Let us fix all the coordinates Xm+l
=
x
m + l j -^m+2 — xm+2
i - - -j ^n — ^ m
in the neighborhood i(X). For the coordinate xm = xm there is the neighborhood (x'm,x^) which is entirely contained in the neighborhood £(X). For any xm € ( i ^ , i j and fixed coordinates £m+i i • • ■ > xn there are X\ = fi(Xm,
Xm+i, . . . ,
X-rn — l — ^Pm— 1 \X-m, X
9i{ )
= gi{f>l,¥2,-
£ m + l , ■ ■ • , XnJ
■ ■,fm-l,Xm,Xm
such that i = 1, m — 1. If these coordinates are substituted in gm(X), any xm 6 (x'm,x'^) we get the inequality
Xn)
+ u
. . . ,Xn)
= 0
then under the assumption made for
7 V^m-l i ^ m i 3-771 +15 ■ • ■ ) ^ n ) ^
0.
Since the last n — m coordinates are fixed, the changes in the functions ip% are all dependent on the changes of the xm coordinates only. Hence for any xm 6 (x' , x" )
6.3.
113
INVESTIGATION
there is gm < 0, while gm(X) = 0 is valid_at the point X. In this case, the total derivative with respect to xm at the point X must be zero, i. e. the equality dfi
dxi dxm
,
dg™d
dgm
d^>m^ _
9xm_i dxm
dx2 dxm
dgn
d
(* * * )
must be valid. Comparing the relations (**) with (* * *) yields the set of m equations in m - 1 unknowns. The vector 3x m '
' dxm
is uniquely determined from the set (**). (**) is a solution of (***) only if (***) is a linear combination of equations from the set (**), Since the coefficients form the Jacobian, the relation J y = ( 1 , . . . ,m) = 0 is valid, which is inconsistent with the condition. Hence the above assumption is valid and there is Xm such that for r = l , m — 1, gr(Xm) = 0 and, additionally, gTn(Xm) > 0. Since the function gm was obtained by renumbering, the statement of the lemma holds for any i = l,m. The above lemma can be obtained as a direct result of the implicit function the orem by introducing an artificial variable. The following assertion yields the necessary conditions for existence of the solution in Hanson's problem. T h e o r e m 6.5. For the vector X to be the solution of problem (6.4), it is necessary that there be the r-dimensional vector A < 0 and s-dimensional vector function u\(Y) < 0 such that
f,,i,tl0,(wWifl
- "su^wjsaa. o (6,o) \G(X)
u.(Y}h(X,Y)
= 0 = 0>Vy
(6.11) (6.12)
where FX(X) = ^ ^ i s a vector a n d G * W = ^5T^. M X . Y) = ^ x ^ are the matrices of partial derivatives computed at the point X. If in problem (6.4) the function 4>{X ,Y) is convex, while components g}(X,Y) and hi(X,Y),j = l,r,/ = 1,5 are concave in X, then the necessary conditions are also sufficient. Proof. Necessity. Instead of problem (6.4) we will consider the equivalent problem (6.7)—(6.9). Let X be the optimal solution of problem(6.4). Then it is also the optimal solution of problem (6.7)—(6.9). Introducing residuals, the theory of Lagrange multipliers can be utilized to show the necessity of conditions (6.10)—(6.12). The nonpositivity of A and fJ,(Y) can be obtained by the Taylor expansion theorem. Let { = ((1(1 ■ ■ ■, (h f and r,(Y) = {f]f(Y),..., ^(Y))7 be r-dimensional vector and 5-dimensional vector function such that G(X)
- e = o,
114
CHAPTER 6. EXISTENCE h(X,Y)-v(Y)
AND
OPTIMALITY
= 0.
Let us construct the Lagrange function: L (X, A,/*(y),£, ij(y)) = F(Jf) + A (G(X) -0+
fi(Y) (h(X, Y) -
V(Y)).
The extreme points of the objective function in problem (6.7)—(6.9) are among the stationary points of Lagrange function. Partial derivatives are taken with respect to all the variavles vanishing at the point X. Thus we have dLdF(X)+x8G(X)dh(X,Y)^
OX OX
OX OX
OX OX
(6.13)
OX OX
| ££ == Gm Gm -- ee == 0,0, ((*)g * ) g == 2A.6 2A.6==0,* 0,*= =T7rT7r
MX y) vhY) 0l (
= ** } ^) = 2MYMY) = 0'l = zrs-
oik) = ' -
r)T.
Comparing equation (*) with (**) permits elimination of variables £,• and rj\(Y). We get the conditions A,G,(.Y) = 0 and {X\(Y)h,i(X,Y) = 0 or £ A 8 G , ( X ) = A G ( Z ) = 0,
YlMY)hl(X,Y)
= ^(Y}h(X,Y)
= 0.
1=1
The last two expressions and relation (6.13) are the conditions (6.10)—(6.12) from the theorem. We will show the nonpositivity of A and fi(Y). From the conditions (*) and (**) it follows that if Gi(X) > 0, then the corresponding A, = 0 also for hi(X,Y) in a similar manner. If some component of vector constraints (6.8)—(6.9) vanishes at the point X, we consider how the vector X changes to some adjacent feasible point X" Let us employ Taylor theorem to the first order with the residual term of a higher order e(X", x) = e. We have
F(X*) + v(Y)h(X; v(Y)h{x; y)-{F(X) + fi(Y)h(X F{X*) + + XG(X") XG(X') + - (F(X) + + \G(X) XG(X) + ^(Y)h(X, t >Y)) - ) ) == y) : (X* ~ X) d
If,;
= (** -X)( -m Hence
\ \
OX OX
F(X*) - F(X) = -XG(X')
dG
(X)
+
dh(X,Y
Xd-^l+ KYf-^) ' 8x Ox
Ox
- XG(X) - fi{Y)h{X\
= -[XG(X)
+
l
h~ +e
= ,
]
y) + »(Y)h(X,
v{Y)h(X\y)-S\.
We now take the feasible solution X* so close to X that:
y) + e =
6.3.
115
INVESTIGATION
1) £ has no effect on the sign of expression in square brackets and 2) all zero" components of the constraint vector (6.8), except the z-th component, and all zero components of the constraints vector (6.9) keep their values unchanged. This choice of X* is possible, because the conditions in problem (6.7)—(6.9) satisfy the preceding lemma. The signs of expressions [F(X*) - F(X)] and [-X,Gt(X*)} are equivalents. Theorefore, if by choosing X", Gi(X') > 0 and X supplies a minimum to the functional (6.7), that is, F(X") - F(X) > 0, then A < 0. Since any index was taken from zero components, the entire vector A < 0. In a similar manner it can be proved that )J.(Y) < 0. Sufficiency. Suppose there are A < 0 and /i(V) < 0 such that the conditions (6.10)—(6.12) are satisfied. The continuous X-convex function <j>[X,Y) integrates over the region {V} to produce the A'-convex founction F(X). This also applies to the concavity of components of vector function G{X). When the conditions (6.10)— (6.12) are satisfied, by the properties of functions F,G and h, we have the following system of relations:
F(X) > F(X) + (, - I ) M 1
= F(X)
U9^!
-{x-X)
> F(X) - A (G(X) - G(X)) - v(Y) (h(X,y) = F(X) - XG(X) - v(Y)h(X,y)
+
«Y)BJ&*>) >
- h(X,y))
=
> F(X).
Comparing the beginning and the end of the system of relations, where X is any feasible solution, we find that X is the point of minimum. Hence the set of equations (6.10)—(6.12) provides n-\-r-\-s specific, necessary and, in the case of convex programming, sufficient conditions to solve problem (6.4) in terms of problem (6.7)—(6.9). The number of unknown components of vectors X, A and fi(Y) is also n + r + s, in which case the inequalities A < 0 and fi(Y) < 0 for each Y change problem (6.4) to a typical nonlinear programming problem. In contrast to the conditions of Hanson problem, here the augend and addend from the conditions of (6.10) are no longer dependent on Y The vector function /J.(Y) is still dificult to determine. We will give the theorem which, under the conditions stated in problem (6.4), follows from the optimality of vector X and new necessary conditions for optimality. T h e o r e m 6.6. Suppose X is the optimal solution under conditions of problem r (6.4), and let
dF(x) OX dx
! ,dG(X) OX dx
Q
(*)
for some y = Y\ Then /i(Y) = 0 for any Y Proof. Let X be the optimal solution in problem (6.4). Then there are A < 0 and fi{Y) < 0 such that the necessary conditions (6.10)-(6.12) are satisfied. Since the system (*) is independent of Y, there must be
p[Y)—w~— = 0,Vy
116
CHAPTER 6. EXISTENCE
AND
OPTIMALITY
From the necessary conditions of (6.10)—(6.12), and from the nonpositivity of ji{Y) it follows that if h,(X, Y) > 0, then i*(Y) = 0, and if hi(X, Y) = 0, then m{Y) < 0. If for the last i there is at least one component w(Y) < 0 with some Y, this would mean in accordance with (*) that the linear combination of gradients for the functions hi(X,Y) vanishing on ~X must be zero, i.e., the gradients must be linearly indepen dent. This assertion seems contrary to the condition. Hence fj,i(Y),i = l , s are all identically zero for any Y Note that the conditions (6.10)—(6.12) permit formulation of the dual problem. It can be shown that for the covexity of F(X) and concavity of components of G(X) and h(X, Y) there is a theorem that is equivalent to the Kuhn-Tucker saddle-point theorem. In this case Slaytor's condition follows from Lemma 6.1, because for nonzero components of constraints (6.8)—(6.9) there is a neighborhood of the point X, at which these components are strictly greater than zero, and for the other components it is always possible to take a linear combination of suitable points X' from Lemma 6.1 to be the interior point for the convex region of admissibility. In applications, however, this approach to solving problem (6.4) involves compu tational difficulties and, additionaly, the imposed conditions are very strong. Conditions (6.10)—(6.12) generally provide only stationary points in which case for each stationary point one has to determine (i(Y) for all y 6 V The solution of the system of equations (6.10)—(6.12) depending on the vector Y generally involves serious difficulties. Furthermore, the range of problems becomes narrow if the functions are all re quired to be twice continuously differentiable with respect to all variables.
6.4
Definition of the Set of feasible Solutions in Hanson's Problem
Utilizing the specific features of constraints as in (6.9), we may, in some cases, elimi nate the dependence on parameter Y from problem (6.4) and define the set of feasible permanent solutions FJ- The set of feasible solutions S for problem (6.4) then is the intersection of the set of feasible X from constraints (6.8) with the set of feasible permanent solutions F] from constraints (6.9), i.e.
S = Uf){X:G(X)>0}. Consider some specific cases, where the analytical character of the functions hi(X,Y) or the economic sense of random values allows one to find the set F] f° r constraints of the form (6.9). Let us consider I I = {X : X € fl {h(X, Y) > 0}} = | A ' : H(X) = r min } h(X,
Y)>o\,
where H(X) and h{X,Y) are ^-dimensional vector functions. If the set \\ is convex, while Gi(X) are all concave, then the nonempty intersection of S also is convex. But
6.4. FEASIBLE SOLUTIONS
IN HANSON'S
PROBLEM
117
in the general case T\ is not convex, and components of H{X) are not continuous functions in X. In all cases the set of feasible solutions S is assumed to be nonempty. Let L be some subset of indices i = l , s uniting the constraints from (6.9) by some common property. In contrast to the constraints in Hanson's problem, the only requirements here are that hi(X, Y) are all continuous for any i — \,s and the condition X > 0 is satisfied at all times. Let \\{L) be the set of feasible permanent solutions for constraints with indices / 6 L and let n(L) be the set of feasible solutions for the other constraints of the form (6.9). 1. Let L be set such that for any / 6 L : hl(X,Y)
=
h'1(X)h'2(Y),
i.e., h'(X,Y) can be represented as the product of two analytic functions, if L = Ly U L 2 U L 3 it follows that: a) if for all l\ g Li the inequality /i2' (V) > 0 holds for any Y, then the relevant constraints hll{X, Y) > 0 can be replaced by constraints hll(X) > 0; b) if for all l2 6 L 2 the inequality h2(Y) < 0 holds for any Y, then the relevant constraints h'2(X,Y) > 0 can be replaced by inequalities h'2(X) < 0; c) if for any I3 € L 3 there are 3/1,3/2 such that h2{y\) > 0 and h2(y2) < 0, then the set of feasible solutions is determined from the condition h^(X) = 0. So we have
Yi(L) = [ x : x e n (Mi(A-)>o)}n n(-Y:A'e n (M 2 po
(6-14)
2. Let L be a set such that for any / G L there is h'(X, Y) = fej(X) + k'z(X) can be represented as the sum of two analytic functions. In this case, the relevant constraints with indicies from (6.9) can be replaced by the following conditions: h[(X)>-mmh'2(Y),Wle-L
J[(L) = L
.Xef]
[h[(X)
> -rmn4(n}|
(6-15)
3. Let L be such that for any / € L there are linear inequalities, i. e., AX — 6 > 0, where A is an I x n matrix and some or all elements of matrix A and vector b are random and finitely distributed, i.e., there are optimistic and pessimistic limits for changes in the values. These random elements form the vector Y Introducing A~ and b+, where random elements are replaced with their pessimistic and optimistic
118
CHAPTER 6. EXISTENCE
AND
0PT1MAL1TY
limits, respectively, the set of feasible permanent solutions LI(L) can be determined for the group of constraints L as follows: Y\(L) = {x '■ AX ^ fe+}> w n i c h follows from the set of inequalities A(u)x > AX > b+ > b(u), where the X satisfying the (•') (•) M inequality (*) satisfies the inequality (**) as well. 4. Let L be such that for any / € L,h'(X,Y) are quadratic in X, i.e., these functions can be represented as h'(X, Y) = XTH,X
+ p,X - b'; I € L
where some or all elements of matrix Hi, vector pi and component b ,1 £ L, are random, finitely distributed values. They determine the components of vector Y So for the constraints I 6 L we have: U(L) = Ix
X : X G fl
X
: X ef]
{minh'(X,Y)>0
min [XTH,X
+ p,X - b, > 0
■■Xtf) l ^ f E E A i i ^ y + E p W - 6 ' ^ 0 ) } }
(6-16)
Since the function between the inner brackets is linearly dependent on the com ponents of vector Y, the minimum can be achieved on the boundaries of the feasible region, in which case x,x3 > 0 and
U(L) = \x . x e n (EXX-**; + tp'f*i -*t > o Alternatively, U(L)
= \X
I
:X e n
V(£L
{A'T//"X + P r ^
> 6f}[
(6-17)
J
Ifffj""are all nonpositive definite matrices, then the nonempty set n ( L ) is convex. 5. Let L be such that for any / € L h'{X, Y) is concave in X. Then it immediately follows that the nonempty set Yl(L) is convex because the relation h'(X, Y) > 0 holds for any Y Here the problem is to determine boundaries for t h e convex region of ]1(L). If there exists a boundary or part of it, then it can be determined from the condition h'{X,Y) = 0. Note that in all cases we have n = n ( £ ) n n ( £ ) - If II and subsequently the set S are composed of a number of incoherent sets of feasible X, then problem (6.4), has to be divided into subproblems with the same objective function (6.7) given on these sets Sk(k = 1, K). In problem (6.4), the minimum can be obtained from the relation F(A) = m i n t e T ^ m i n S t F ( A ) .
Chapter 7 Stability of Solutions in Stochastic Programming Problems 7.1
Stability of Solutions in Stochastic Linear Programming Problems
The issues of solution stability in mathematical programming problems are addressed from various points of view and in a variety of ways. The literatures on solution stability choose the conditional extremum as a random point, the optimal basis as a set of vectors, and the optimal value of objective function as a random value. Such a choice determines introduction of different stability concepts. In some literatures, the stability issue is dealt with in stochastic and parametric terms as well as in terms of the error theory. The solution stability is crucial for stochastic programming problems, because here the parameter values are random. We will consider some issues of existence of solution stability regions for stochastic LP problems. Let us introduce the following concepts: 1. Feasibility region. Consider a fixed realization of random event ui0 6 fi. Suppose we have a deterministic LP problem of the form min C(u>o)X.
(7.1)
Let lk(k = 1,2,..., Kwa) denote the extreme points of the convex set S(LO0), each of which is produced by intersection of n cutting planes, where KWa is the total number of extreme points when LO0 € fi random parameters are realized. The region Vj^ C fi is called the feasibility region of the point If. if for any w 6 V^o the intersection of the cutting planes producing this point determines an extreme point of the corresponding set S(ui). 2. Optimality region. Let l^0 be an optimal extreme point of problem (7.1), i.e., for any k / k0 Z* > Z*°, where Z* is the objective function value of problem (7.1) 113
120
CHAPTER
7. STABILITY
OF SOLUTIONS
IN SPP
at the extreme point I}.. The region W** C Vk° is the optimality region of the point lko if for any u G W* Zkw > Zk" (ft = 1,2,... ,ft„;ft = k0). L e m m a 7 . 1 . The function Zk is the continuous function of ui for any extreme point Ik. Proof. The functional value at the extreme point Ij, of the convex set S(u>o) for the fixed realization o>0 G 0 is Zko = c\x\k + c°2x°2k + . . . + c°nx°nk, where xf are coordinates of the extreme point. Denote by xok nonzero values of x° ■ These values are determined from the relation DokXok = 6°, where D0 is the relevant basic matrix. By the nonsingularity of this matrix, the values x?k are continuous functions of a°- and 6°, respectively, Z* is a continuous function of u> G Q. We will prove that for any u0 6 fJ there exists a stability region of Wua such that for all u> G Wwo the corresponding problem (1.1) has the same optimal basis. T h e o r e m 7 . 1 . Let
zt0 > zli (ft = I X : ; k ± k0) at the point w0 E H. Then there exists a neighborhood 0(OJQ) of the point u>0 such that for all u> G 0{ui0) C tt
Zt > ZkJ
(k = I X ; k ± ho)
Proof. At the point u>o E fl Zl - Ztl > 0.
(7.2)
By Lemma 7.1, the functions Zk and Zk° are continuous in w, hence their difference also is a continuous function. From the properties of continuous functions it follows that there exists a neighborhood O(uio) such that for LO G 0(U>Q) the inequality (7.2) is preserved, i.e., Zk > Zk°. [1] shows the ways of obtaining boundaries for parameter variations in LP problems to preserve the optimal basis. [17] and [20] prove the existence of stability regions at any point w € SI.
7.2
e-stability of solutions in t h e mean
Unlike Tintner,N.I.Arbuzova and V.L.Danilov [1], [2] considered one fixed point in space J7, i.e., the point To whose coordinates are expectation for random parameters of problem conditions. The authors obtained conditions under which the optimal basis of problem continues to be optimal for all LO G fi except the subset of a specified measure e < 0. DEFINITION. The solution of stochastic LP problem in the mean is called stochas tically stable in module e (e-stable) if the optimal extreme point of problem in the mean continues to be optimal for any realization - of random parameters except the set of measure e.
7.2. e-STABILITY
OF SOLUTIONS
IN THE MEAN
121
N.I.Arbuzova and V. L. Danilov concentrate on a special case of stochastic LP problem, where the vector b(u) alone is random and, additionally, its components are indipendently distributed. Denote by 6; the expectation of a random variable 6j(w), and by of its variance. DEFINITION. The hyperplanes, whose intersection produces the optimal point of problem in the mean, is called labelled. Let the boundary hyperplanes of the con vex set 5(57) be all enumerated, and let ki,k2,. . . ,kn be the numbers of the labelled hyperplanes. The extreme point is determined by intersection of the hyperplanes, and its mo tions are dependent on realizations of the vector b(uj). The set of n-dimensional Eucledian space, in which the point may fall with probability greater than (1 — (l//2))™, is an ellipsoid with its center at the point corresponding to the optimal solution in the mean. Here / is determined from the inequality (1 — (l// 2 ))* 1 > 1 — e. The axes of this ellipsoid of concentration are obtained in the following way: at the point 57 G f! we consider the intersection of n — 1 hyperplanes from n labelled hyperplanes which is geometrically a straight line. On this line, we separate on either side of the ellipsoid center the value toa^ , where kj is the number of the labelled hyperplane, which does not belong to this intersection, and <7t is the mean square deviation of the random variable on the right-hand side of this hyperplane equation. From the independence of components bj(u) and Chebyshev inequality follows
p {n |6s(u) -l~\ < to} = T[p{\b>(") - £ | < to} >II (l - jj) = (l - £ f (7.3) The value of / can be determined from the inequality
/
1
1,m
I - -P)
>l
~
-e,
(7.4)
where e is a sufflcently small number. DEFINITION. Constraints of the form a,.x < 6, — loi(i = l,ro) are called lower constraints, and those of the form a^X < b, + la, (i = l,m) are called upper con straints. Suppose the intersection of lower constraints and a positive hyperoctant is not empty.This means that any realization of vector b(u>) (except the set of measure e) there is at least one feasible solution. Using the above concepts, we formulate without proof the sufficient condition for the stochastic £-stable solution in the mean. T h e o r e m 7.2. For the solution in the mean to be e-stable, it is sufficient that the ellipsoid of concentration does not intersect lower constraints except the labelled ones. For the convex programming problem, the e-stability signifies the constancy (with probability 1 — e) of the basis of labelled normals, by which the vector is decomposed with positive coefficients. Following N.I.Arbuzova, this definition of e-stability is hereinafter referred to as basic.
122
CHAPTER
7. STABILITY
OF SOLUTIONS
IN SPP
DEFINITION. Suppose that for edl wmode the set of optimal solutions H(u) belongs to the &-neighborhood of the set Hip), i. e., for any X°(u), there is X°(U) such that \X°(LJ) - X°{ZJ)\ < S.
It this case, the stochastic programming problem is called solution-stable with char acteristics e, S. Suppose the problem is basis-stable mod e and has a unique optimal solution for u = U, Then it is possible to find a distribution of the optimal solution X°(to) = = (I 1 (UI), X2(UJ), . . ., Xk{u>), 0 . . . 0) over the set of measure 1 — e, and a distance dis tribution \X°(u>)-X°(uJ)\. For any S there is S\ = p{\X°{u) — X°(ZJ)\ > S}, and the problem becomes solution-stable with characteristics s + E\, 5. In some practical problems, the final criterion for stability is provided by devia tions of the optimal value of the objective functional. For this reason we introduce the stochastic stability concept of convex programming problem, in which the issue stability is considered from this point of view. DEFINITION. For any ujmode, let \C{LO)X - C{u)X\ < 5; then the problem is called functionally stable with characteristics e, S. If the problem is e-feasible, 5-stable, then by the continuity of the objective functional in all its independent variables there is S such that the problem is e-functional, S-stable. Following [3], we formulate some propositions. T h e o r e m 7.3. For the solution in the mean X*(U)
= (x*(Z3),x*fi),...
,xl{u),0,...
,0)
to be e-stable, it is sufficient that Y17L\ of < ed2, where d is a minimal distance from the point U to the boundary hyperplanes of stability region W-g C fi. Proof. If DZJ is the k x k basic matrix corresponding to the solution in the mean, then the nonzero components Xj(tj)j = l,k of vector X*(LO) are determined from the relation X"(UJ) = D^ b(ti), where b(iJJ)T = (&i(w),..., bk{ui))T Additionally, in the stability region, the solutions in the mean coordinates of the intersection point of labelled hyperplanes x](u>) are determined from the same relation %{u) = or
m
D^b(u) D
^ » =B-1)'+J^.H,
J = hk,
(7.5)
where A is the determinant of matrix D, and D, ; is the minor of this matrix element d^. The stability region of the solution in the mean W^ coincides with the feasibility region of this solution in space 0. On the other hand, the intersection point of labelled hyperplanes X(OJ) is no longer feasible if it lies on hyperplanes of the form
7.2. e-STABILITY
OF SOLUTIONS
123
IN THE MEAN
di.x = b'(i > k) or x3 = 0(j < k). By the relation (7.5), this is equivalent to the fact that bi(u>) satisfies the inequalities k
k
6 , H A - £ a t J £ ( - l ) ^ ' D , A M = 0, i>k, i=l
(7.6)
e=l
E(-1)' +J .D^,H = 0, ,'
(7.7)
!=1
Hence the stability region Ws of the solution in the mean is determined by the intersection of m hyperplanes of the form (7.6)—(7.7) in the m-dimensional space fi. If the equations (7.6)—(7.7) are normalized and contain b{ instead of 6,-(w), then they provide the distance from the point Xo of space R to the hyperplanes determined by (7.6)—(7.7). Denote these distances by di,d2,. .., dm and define d = min,d\
(7.8)
<e.
Show that 23™-i erf < ed2 implies (8.8). By Chebyshev inequality and properties of expectation, - i i , we have
E['\b{u)-b?\ I1 |2 1 T 1 \b(u))-b\ Pi 'J |6(w)-6 >d\ >d\ < - J d22 1 d
E ( E , (&.-(«) v \ 22
-tf)
d rf
Er =1 E (6,-(W) --5)'.
Em, °
d?
d22
-£'
This completes the proof of the theorem. In [3], the necessary and sufficient condition is derived for the e-stability of the problem ma,xCX,AX 0 with the random first column of matrix, i.e., the problem max CX, Pl(uj)xl + P2x2 + ... + Pnxn < B;
__ Xj > 0,j = l,n.
(7-9)
The elements of column Pi{u>) are random independent variables an(u>),a21(uj),. .. . . . , amj(u>). Suppose problem (7.9) has a unique optimal solution in the mean X*(UJ) with a positive component x\(lo). The proposition is to transform variables Y = 1
xx22
xn <£n
1 / 2 == — ; • ■• •■; ;yn~y\ --== —; — ; --2/2 yn =- — X! Xi X1 Xi X\ X\
124
CHAPTER 7. STABILITY
OF SOLUTIONS
IN SPP
Then (7.9) becomes a fractional programming problem: f m a x Cl + CW + ■ ■ ■ ± C ^ Jl I [ -By, + Pm + ... + Pnyn < - P i H ;
(7.10) y, > 0, j = l,n.
In problem (7.10), the column containing free terms of the set of conditions is random. The solution of (7.10) in the mean exists because X*(UJ) > 0, y" = ip(X*); the e-stability of (7.10) implies the constancy of the set of labelled indices with probability 1 -£.
The equivalence of £ stability is proved for problem (7.9) and (7.10); the necessary and sufficient condition for e-stability is derived for problem (7.10). Let us consider the dual problem to (7.1) with a random vector 6(w) max YTb{u),
(7.11)
where T = {Y : YTA < C{Y < 0)}. It turns out that the e-stability condition we have obtained for the direct problem also ensures the e-stability of the solution in the mean Y for problem (7.11). This assertion follows from the lemma. L e m m a 7.1.Suppose the problems (7.1) and (7.11) have a unique optimal solution for all UJ 6 fi (except for the set of measure e). Additionally, let the solution in the mean X of the direct problem be e stable. Then the solution in the mean X of the dual problem is also e-stable. Proof. Let WZJ C O be the e-stability region of the solution in the mean of problem (7.1). The problems (7.1) and (7.11) have a unique optimal solution on the set WZJ. The point of intersection of the hyperplanes labelled with numbers i j , . . . ,ik,ji,- • ■ ,ji is the optimal extreme point for any n = k -f 1. By the continuity of the solution n = k + l, the duality theory implies that the labelled hyperplanes in the dual problem are the hyperplanes designated with numbers ik+\,. ■ ■, im, ii+i, ■ ■ ■, j n - The set of labelled hyperplanes for any w S W„ in the direct problem is the same, and hence the set of labelled hyperplanes in the dual problem is constant. Since the optimal solution of problem (7.11) corresponds to the intersection of these hyperplanes, the solution in the mean Y is e-stable. We will now consider some special cases of determining solution stability regions for linear programming problems with random data. In problem (7.1), only the vector b is assumed to contain random components. Suppose the problem is solved at the point u>i G $7. Denote the optimal solution
jf(w l ) = (ii(wi),...,*;(«!), o,...,o). In this special case of LP problem, the stability region coincides with the feasibility region. The feasibility region is determined by the set of inequalities 6,HA-^a„X(-ir 3=1
v=\
+ 1
^^H>0,
^>k;
(7.12)
7.3. STABILITY
125
OF SOLUTIONS
B-ir+^&.H > o, j< k. i=i
To ascertain the belonging of the point w2 to the stability region WUI is to find out whether the vector coordinates b(ui2) satisfy system (7.12). But such verification requires substantial computations. One may try to find the subset of stability region (7.12) which allows to ascertain geometrically whether the vector b(u>) belongs to the stability region. Denote by d a minimal distance of the point u from hyperplanes of the form (7.12). Let us solve the equation m
J2(b,(^)-b1(i01) + af = d for a, i. e., we have a = (d/y/m). Then the rectangle [&,-(wi) —i 6 fi the optimal extreme point is the point l0 with (x°,..., i ° ) coordinates. To find out under which realizations of the random vector c(w) this point is optimal, we consider n adjacent extreme points whose coordinates are denoted as (x{,..., x3n), j = 1, n. Then the stability region is specified by the set of inequalities s?ci(w) + . . . + x°nCn(oj) < sici(w) + . . , + x{cn(u).
(7.13)
As before, it is possible to find in this stability region the set which allows to ascertain geometrically whether the vector c(ui) satisfies system (7.13). A more general case is considered. Let o»i € fi be a fixed point; then hyperplanes are of the form at,(ui)X = b^Wj), i = 1, m. Find all the extreme points of the convex set determined by the intersection of these hyperplanes with a positive hyperorthant. Denote the coordinates of these points by {x\{u>\),..., x3n(cji)), j — l , ^ , where kui is the total number of extreme points. Let lB = (x°(u>j),.. ., i°(wi)) denote an optimal extreme point. The stability region then is determined from the system of inequalities: clx<{(u + ... + cnxl(Lo)
j =IX-
(7-14)
This system differs from system (7.13) in that involved here are not only the adjacent points, but all extreme points. Additionally, in (7.13) the coordinates of extreme points are the same for all w G fl, whereas in (7.14) they vary with realization of random data.
7.3
Stability of Solutions to Stochastic Nonlinear Programming Problems
In applied problems of stochastic programming, random parameters have optimistic and pessimistic boundaries for their variations, i.e., random variables are distributed
126
CHAPTER
7. STABILITY
OF SOLUTIONS
IN SPP
in a finite (continuous or discrete) way. In such cases it is possible under certain conditions to determine the range of the objective function and localize the set of optimal solutions for all realizations w € fi. Suppose we have a stochastic programming problem, in which the objective func tion f(X,u>) = f(X) is a continuous deterministic function, and the functions defining the problem conditions gi(X,u>), Vz = \,m are quadratic or linear in X with nonpositive definite matrices Hi(cj) for any realization to € fi , i. e., we have the problem min f(X)
)
J \ i subject to XxTHj(u>)X + p,{u>)X - bi(u>) > 0, i = 1, m ( T n subject to Xv H-*i(u)X + pi(ui)X-bi(uj)>0, i = T~^ >
X > 0,
(7.15; (7.15)
J
where all the elements of matrices H,(UJ), vectors p,(oj) and componentsfe;(u;),i = l,m can be random. Let us introduce the following sets: S(u)=
{X :XTH1{LO)X+P,{OJ)X>
S~ = [X ■ XTH~X T
+
S+ = {X : X H,
+ p~X > b+,
X + p+X>b-,
bi{u>), z > 0 , X>0, X>0,
»=T~m},
i = I~^} , i = T~^},
where in fl/, pf, bf, H~, p~, b~, i = 1, m random variables representing realizations of vector LO are replaced by optimistic and pessimistic boundaries, respectively. We have the following statement. T h e o r e m 7.4. If the set of permanent solutions S~ is nonempty, then the fol lowing relation holds for the objective function of problem (7.15): min f(X)
< m^fiX)
< mm f(X).
(7.16)
Proof If is sufficient to show that S+ D S(u) D S~, whence it follows that the minimum of continuous function in some set is less than or equal to the minimum of this same function in any part of this set. If X € S~, it follows that for all random realizations u> and with Vi XTH+X+p+X +P," 1
> XTHi{u)X + pi(u))X > XTH~X+ (...) (..) > K > bAu) > b~, X > 0. (.) (,.) (...)
From this it follows that any feasible solution from S~ satisfying inequality (*) also satisfies inequality (**), i. e., 5(w) D S~. Additionally, any feasible solution from the set S(u>) satisfying (**) also satisfies (* * *), i.e., S+ D S(LO). Finally, S+ D S(u) D S~
7.3. STABILITY
127
OF SOLUTIONS
and if S~ ^ 0, then the expression (7.16) from the conditions of the theorem is meaningful. A symmetric theorem can be formulated for the max problem. Corollary. Suppose we have a quadratic programming problem, i. e., in problem (7.15) f(X) = f(X,uj) is a random X-quadratic function with a nonnegative definite matrix H0(UJ) for VCJ
= XTH0(UJ)X
f(X,u)
+poM-Y + 6o(w)
where among the elements of H0, po and b0 are random components with finite dis tribution. Then the expression (7.16) from Theorem 7.4 becomes f-(X-)
= mm f-(X)
< mm f{X,u)
< min f+(X) = f+{X+),
where X~ and X+ are optimal solutions for respective sets S+ and S~, while f~(X) and f+(X) are objective functions, in which random data are replaced by their pes simistic and optimistic boundaries, respectively Proof. In accordance with the statement of Theorem 7.4 the system of mutual inclusions S+ D S{u) D S~ is valid. If X (w) is the optimal solution of problem (7.15) for realization of u>, then for any ui f-(X-)
= X-TH;X-+P;X-+bv QT
= X {U)H0{U)X°{LO) T
( X » )= + Po{to)X°(u)
< X+ H+X+
+ bo(w) <
+ p+X+ + 6+ =
f+(X+).
whence comes the required result. Introduce the set S~= [X : XTH~X
+ p~X >bf,X>Q,
i = TTT^}
Theorem 7.5. Let the functions gi(X,u), i — l,m in problem (7.15) be concave or linear, the set of feasible X S~ nonempty, and the set S+ bounded. If the objective function of problem (8.15) is concave and deterministic, then for any to there is an optimal solution X°(u>), which does not belong to S>, and the set of all such optimal solutions satisfies the conditions
{x»}
C A S = S+\ S~
Proof. Since the minimum of the concave function f(X) on the boundary is not less than the minimum of f(X) inside the convex region of feasible solutions, by the continuity of f(X) and by the closedness and boundedness of the region of feasible solutions for each realization of u there exists a boundary point X°(UJ) such that f(X°(uj)) < f(X(u)), \/X(LO) e S(u). This means that for each realization of w at least one of the inequalities gi(X,ui) > 0 or Xj > 0 becomes an equality, i. e. either X0T{io)Ht{uj)X°{u)
+p,(u)X°{u)
= b,(u,)
128
CHAPTER
7. STABILITY
OF SOLUTIONS
IN SPP
for some iu or z°Aui) = 0 for some j w . Such X°(u>) does not belong to S>, because t h e set 5< includes none of the X which could transform some constraint of problem (7.15) into a equality. But 1 " ( C J ) £ S+, because S+ includes all feasibility regions for any realization of w, and hence
{*V)}Vw o s+\ f If in Theorem 7.5 t h e objective function of problem (7.15) is convex, while the absolute m i n i m u m does not belong to any feasibility region S(OJ) for any UJ, then Theorem 7.5 is also valid. T h e proof can be carried out in much t h e same way as above. Theorems 7.4 and 7.5 are generalizations of the findings in [21], where only linear functions in t h e original problem are considered. The s t a t e m e n t below refers to a special case of stochastic programming problem. T h e o r e m 7.6. If in problem (7.15) all the functions are linear in X and only the vector b(u) has random independent components that are finitely and continuously distributed, then the set of optimal solutions {A°(w)} V u / from Theorem 7.5 is convex. Proof. Let S(UJ) be t h e region of feasible solutions for realization of a;. Let X°(LJ\) and X°(u>2) be optimal solutions for any two realizations, u>i and u>2, i- e. /(A'V,)) (*(«*)),
V A > , ) e 5(Wl),
/ (x0(u2))
VA>2) e
From AX"(ui) > 6(wi), M > A, 0 < A < 1 t h a t
< f (A>2)), 2
)
a
> b{to2), X {ujx)
S(LO2).
> 0, X°{u2)
> 0 it follows for any
A ( A A ' V i ) + (1 - A)X°(«4)) > A6(w,) + (1 - A)6(w 2 ), X° = \Xa{w1)
+ (1 - A)X°(w 2 ) > 0.
The vector X° is a possible realization because of t h e independence and continuous distribution of random variables. This vector is optimal because by t h e linearity of functions in problem (7.15), the objective function gradients and normals to hyperplanes of constraints are constant, and the optimal point X° moving from A°(u'i) to X (w 2 ) continues to be optimal as long as it is feasible, while t h e feasibility for vector X° is proved for any A, 0 < A < 1. Suppose t h e functions of the original problem are all q u a d r a t i c in _Y with nonpositive definite matrices fij(w), i = l,m and nonnegative definite matrix H0(u>) for Vw, i. e., we have the quadratic programming problem for each realization: twnf(X,u>)
= min (xTH0{w)X
+p0(w)X
+ b0{u))
(7.17)
subject to
( XTHt(u)X+Pt(co)X>bt(uJ), \ x>o
t=T^
y7A8>
129
7.3. STABILITY OF SOLUTIONS
wherejhe elements H{{u>), p,(uj) and b,(u), i - 0 , 1 , . . . ,m, can be random variables. Let h'ti, p* and 6; be the mathematical expectations, and let <j\ ,) and b,(w), i = 0 , 1 , . . . ,m, respectively. Introduce the following sets and notations: S H = {.V : XTHt(uj)X
+ Pi{u)X
> bi(u), i = T~fR,
X > 0} ,
S+ = {X : XTH+X
+ p+X >b;,i = T^,
X > 0} ,
S~ = {.Y : XTH-X
+ p~X > b+,i = 17^,
X > o} ,
T
A„iA = {X : X e {S+\ {X : X H~X Jt„ = \ mm h{X)\
+ p~X > bf, il^i,
= mm (xTH+X
fa = I min+ Mx)} = min+ {XTH°X
X > o}}} ,
+ p+X + 6+) ,
+ PoX + 6o) ,
where H+, #7", p+, p~, and 6+, 6," are determined from the following relations for any i = 1, m and /, j = 1, n:
«T = H -- '?%) : fl? = (4 + H ) . p,~ P7
p + == ((pj ^ + Aor}), Aa'), = J/ffj) , p/ = (pj (p$ --- w)), +
6" = = (sr(6:-r = (£+*<*). (£+*<*). ? a t ),6, - Wi) i 6+= 6r Suppose T? and A are strictly greater than zero, S~ ^ 0, while the absolute minimum in problem (7.17)—(7.18) does not belong to S~, and considered for S(ui) are all u> for which there exists at least one feasible solution. Theorem 7.7. The probability that the optimal value of the objective function f(X°(u>)) e [f'xJxJ and the optimal solution of problem (7.17)-(7.18), X°(w), belongs to the region AV]\ is equal at least to Pv>x = P{OJ\H~
< H.(LO) < H+;
p~ < R ( W ) < p+;
K < 6 i H <6, + , Vi = 0 , l , . . . , m } . Proof. It is sufficient to show that the requirement of the theorem holds for the event {u\Hi < Hi(u) < H+; p- < pi(w) < pf; b~ < bt{uj) < bf; Mi = 0 , 1 , . . . ,m}(*). Proceeding in the same manner as in Theorem 7.4, we have S+ D S(ui) D S~ for all us from the event (*), while the set of optimal solutions for these u> is contained in A*,,. According to Corollary of Theorem 7.4 we have then that for all u> from the event (*) the required relations are satisfied for the objective functions: x w fZx = xemir s$ l Ui ) < x™P, jgff /(^> ) -
min
A
xes- -W
') = ft,v
130
CHAPTER 7. STABILITY
OF SOLUTIONS
IN SPP
Now it is easy to show, e.g., by shifting a nonbasic constraint, that depending on the form of objective function and constraint functions there can be realizations such that the optimal solution X°(u>) 6 AVi\ and, respectively, f(X°(uj)j € [/~A; f\iTI\\ but the strict requirement of event (*) is not satisfied for such realizations of u. From this it follows that p,,^ is the lower boundary for the probability that the optimal solution X°(u>) € An,.\, while f(X°(u)) e [f^\, f\
7.4
Feasible Solution and Function Stability with respect to the i-th Constraint
Consider the problem min^ fv(X)
subject to AX > 6(w)
(7.19)
X >0, where fu(X) is a convex function of the n-dimensional vector X = [xi,x2, ■ ■ ■ ,xn), A is an m x n matrix, and b(ui) is a random m-dimensional vector whose coordinates have a finite continuous distribution and the set of feasible solutions S~ is nonempty. The issue of general solution stability amounts to finding a region which bounds the set of optimal solutions inside the n-dimensional parallelepiped. This region is generally difficult to find, because it depends on the form of objective function, the length of interval and distribution of a random variable bt{u>) therein, and the extent to which the random variables bi(u), i = l,ro are dependent. Therefore it seems natural to discuss the issue of solution stability with respect to one random constraint in terms of the optimal solution in the mean. Consider the i-th problem of parametric nonlinear programming derived from problem (7.19):
mmfu(X) rbjectto
= f(X°>(u>))
£ » = 1 airxr = (Ax)j > b^u),
j = l,m;
£?=i a„xT = (Ax)i > 6i(cj),
A' > 0,
j
/
(7.20)
where X°'(LO) is the optimal solution of problem (7.20) for realization b,(uj), and bj(u) is the mean of a random variable 6j(w). Consider an n-dimensional parallelepiped built around the feasible solution X°(LO) of problem (7.19) that is optimal in the mean. So we have [X°(LJ)-8,X°(ID)
+ S};
(7.21)
7.4. FEASIBLE SOLUTION AND FUNCTION STABILITY
131
where 5 is the vector whose components are strictly greater than zero, but can have different values. The problem is to find the probability that in the set of optimal so lutions there is the optimal solution of problem (7.20) contained in the parallelepiped only if 6<(CJ) is a random variable and the other random variables b^oj) are equal to their mean values bj(ZJ),j — l,m.,j^i. In other words, we need to find Pi {X°'(LU)
G [XV)
" S, X°(u)
+
6]}.
DEFINITION. Problem (7.19) is said to be solution stable in the sense of the i-th constraint with characteristics e,, 5 if the relation
P, {.Y°'H e [x°(a) - S,x°(u) + s}} > l -ei is valid. To simplify the issue of solution stability in the sense of the i-th constraint, the solution of problem (7.20) is assumed to be unique for each i and any w. Solving this problem may help to reveal which of the constraints for stability of the optimal solution are unessential, i.e. do not produce perceptible changes in the optimal solution when 6;(u>) is changed, and which of the constraints are essential, i.e., carry the optimal solution far away from its mean value when the random variable bt(uj) is changed. The parallelepiped (7.21) can be constructed in a variety of ways. Consider a special case, where the parallelepiped intersects only those constraints of problem (4.1) which become an exact equality on the optimal solution in the mean X°(u>). The vector 6 can be obtained as follows. Let us take the minimal distance d from the point X°(UJ) to the hyperplanes for which (AX°(UJ))j > bj(UJ). Construct the parallelepiped with its center at the point X° and its sides 2 ■ 8j long, in which case ,/^™_1 5j < d. If for X°(p) there is no constraint which becomes an equality, i.e., (AX°(u>)) ■ > b;(u>),j = l,rn, then parallelepiped is constructed in a similar manner by choosing d as the minimal distance from the point X°(UJ) to each of the hyperplanes that are constraints with the mean bj(uj),j = l,m. Note that if the parallelepiped intersects the constraints which become a strict inequality for X (w), then there is a high, but hard to define probability that the optimal solution for all realizations 6,(w) from some interval [bj,6"] drops out of the parallelepiped. This case requires special investigation. In the first place, the distance d may prove to be too small to make meaningful any examination of the feasible solution for stability; secondly, the vector S is often specified on the basis of practical reasons, and the stability parallelepiped may fail to satisfy the requisite conditions. Let /; denote the set of all indices i = l , m for which the constraints with 6;(CJ) in problem (7.19) become an exact equality on X°(uJ), that is, £ " = 1 at]x°(uj) = 6;(w), while I2 is the set of other indices for which
£oy*J(ZD) >&,(*)• 3=1
132
CHAPTER
7. STABILITY
OF SOLUTIONS
IN SPP
Consider t h e i-th problem of parametric nonlinear p r o g r a m m i n g (7.20) for i 6 / j . We solve t h e problem for A,, 0 < A, < 1 in terms of A; for which
bi: = bt(u) = and t h e condition (AX)i
(i-TAb-+T,bt
= bi(u>) becomes (AX)t
= (1 - \t)b7
+ Xtf
= bK.
We obtain a continuous optimal vector function which depends on A^. Transferring &,-(w) = 6,\, from h(u) into b~, we get t h e lower b o u n d a r y for vari ations in bi(u>), in which case the optimal solution is necessarily an element of the parallelepiped (7.21). Two cases are possible here. 1. Some coordinate xf(uj) leaves the parallelepiped, i.e., t h e r e exists \\ such that (Axoi{wj)
. > t,-(E7);
(Axoi{us)).
j = l,2,...,m;
= b,(io) = (A - Xl)b- + AJ6+, x°'(u)
x°jp)
-5J<
j ^ i.
> 0
s V ( w ) < x°j(VJ) + Sj]
j = 1, 2 , . .. , n.
and there is at least one x°' such t h a t either X°'(UJ) = X°AUJ)-\-SJ or x^(yj) = X°(LJ) — SJ. 2. Moving from the point 6,(uJ) to t h e point b~ we find t h e b o u n d a r y 6" for which AJ = 0, and the optimal solution x°~ is still contained in t h e parallelepiped (7.21). To find the upper boundary, we transfer bi(uj) from bi(p) to 6; - Two cases are possible here. 1. Some coordinate x°l leaves the parallelepiped (7.21), i.e., t h e r e exists AJ such t h a t similar relations are satisfied in much the same way as for the lower boundary. 2. Transferring 6,-(w) brings us to realization bf, i.e., AJ and t h e optimal solution is an element of the parallelepiped (7.21). This means t h a t either t h e hyperplane (Ax,+ )i = bf passes through the optimal point x'+ from t h e parallelepiped (7.21) or the strict inequality (/lz'°); > 6* is satisfied, i.e., this constraint has already become unnecessary and so is replaced by another constraint. So, by changing 6,(w), we found the lower and the upper b o u n d a r y for variations in the random variable 6,-(w) where the optimal solution of problem (7.20) is necessarily an element of the parallelepiped (7.21). Consider those i g / j , i , e , , constraints, which become a strict inequality for X°(UJ). Then {Ax°(ujj). > b,-(E7), (Ax°(ujj).
> br,
When transferring 6,-(w) from 6;(w) to b~, the i-th hyperplane abandons t h e level line f(x) = f(x°(u))) bounding the convex region and including all the level lines which
7.4. FEASIBLE
SOLUTION AND FUNCTION
133
STABILITY
provide the lower value of function f(x) than f(x°(ui)). Therefore the i-th hyperplane does not intersect these level lines and x°(uj) is also the optimal solution for all A; among the-values b~ and 6,(S7). If the optimal solution of problem (7.20) x'+ = x°(lo) for bf, then the optimal solution x°(to) in the mean is the optimal solution for any realization bi(u>) € [b~,bf]. This is proved for the lower boundary. Since transferring 6,-(w) from bi(uj) to bf may only increase the objective function minimum f(x°(ui)), and f(x0(tJ)) = f(x'+), we have f(x0t(u>)) = f(x°(u>)) for all realizations of &;(OJ) and, by the uniqueness of the solution, x'° = x°(uf) for any 6,-(w). If the optimal solution x'+ =£ x°(uJ) for bf is not an element of the parallelepiped (7.21), then from the solution of the parametric nonlinear programming problem, and based on bt(uJ), we get A2 for which s'-f lies on the boundary of the parallelepiped. For each constraint i we have thus obtained the range of random variables 6;(OJ) in which the optimal solution of problem (7.20) is necessarily an element of the stability parallelepiped. We may now find the lower boundaries for probabilities £, which allow the problem to be solution-stable with respect to suitable constraints in terms of random variables bi(uj), i. e., we have Pi = Pi {x°'(u) e [X-°(U>) - 6, ftA2]|&»
> Pf { 6 , H 6
X°(UJ) + S}}>
= f>j(u)j = 1,2,... ,m, i + i} = 1 - £,-,
2
where b\ and 6 are determined from the following conditions: 1. For i € /: hl - I bl b ' ~ \ (i - \})b~ + \}bf
' lf ,if
X
/ # l - A2)&~ + \]bf
' 'f , if
X+ G
tf
'~ G x'- i
(7 221 {l ll ->
' i'+ e
(7 2V
2. For i e h
fc
. -
fc.+
4 /l
\2U.2 , \2L+
( 1 - - A?)&? + AJ6+
,if ';£
,if
x,+ e , + i„.+ $
rf
(7-24)
Proceeding in this manner makes it possible to examine functional stability in the sense of the i-th constraint. Consider the values 6,(OJ) from problem (7.20) for which |/(i 0 (O)) - /(**{w)| < 6. DEFINITION. Problem (7.19) is said to be functionally stable in the sense of the i-th constraint with characteristics &;, 8 (5 is a number) if P, = P, {|/ (X°(5T)) - / (xDi(u))
\<6}>l-et
To find the probability p,, we solve the parametric nonlinear programming problem (7.20) for the random variable bi(ui). We again find A] and A,2, which determine the
134
CHAPTER 7. STABILITY
OF SOLUTIONS
IN SPP
range of random variable 6,-(u>), to make problem (7.19) functionally stable. But we seek the interval [Af,A?] proceeding not from the solution in the mean, but from optimal solutions X'~ and X'+ such that, in problem (7.20), bt(u) are replaced by b~ and bf, respectively. This is due to the fact that, by the nonuniqueness of the solution in the mean X°(UJ) for problem (7.20), we could obtain a feasible solution X°'(LO), for which
| / ( A - V ) ) - / ( A - 0 , H ) | = <5(*),
and terminate the computation even though there may still be realizations 6,-(u) for which the equality (*) is also valid.
7.5
Investigation of Absolute Solution Stability
In some stochastic programming problems a feasible solution X° is optimal under changes of random parameters. For example, this may occur when the absolute min imum of objective function belongs to a set of feasibility regions S(ui), but does not belong to a set of permanent solutions S~ Then it is possible to formulate the prob abilistic problem, in which the feasible solution remains optimal for all realizations w 6 O. Here we may investigate without loss of generality only the case, in which the feasible solution belongs to the region S(tJ) that is feasible in the mean. If the abso lute minimum belongs to some other feasibility region S(u>), X° £ S(tJ), X° £ S~, then investigation can be carried out in a similar manner. We assume that the range of random variables is always finite. This formulation of the problem makes sense only where the convex program ming problem is stochastically constrained or where an extremum of stochastic linear function is to be found on the deterministic polyhedron. We will discuss these two cases. DEFINITION. The stochastic programming problem is called absolutely solutionstable with probability a if the solution in the mean is optimal for all realizations w with probability a. We will consider the direct and the dual problem of stochastic programming by employing usual notations where the conditions of problem form bounded polyhedral regions for each w 6 S7: max CX; min Yb(u);
S{u>) = {X : AX < b(u>),X > 0} , Q = {Y : YA > C, Y > 0} .
(7.25) (7.26)
Since the extremum of linear function of the bounded polyhedron is reached at the extreme point, the question of solution stability in the sense of the above definition can be raised only where not all bi are random and the optimal solution in the mean is generated by deterministic hyperplanes or where we have the trivial case A'0 = 0. The probability a that the feasible solution X° is still optimal is then determined by the set of those realizations w for which A'0 is a feasible point.
7.5. INVESTIGATION
OF ABSOLUTE
SOLUTION
STABILITY
135
The case is different with the dual problem (7.26). Here the polyhedron of the feasible solution region is deterministic, the optimal point X° in the mean is no longer the optimal solution, and for some realization the objective vector leaves the positive cone stretched along the normal of those hyperplanes which generate the optimal solution in the mean. [2] shows that if under some assumptions (uniqueness of solution, etc.) the direct problem (7.25) is basically stable mode, then its dual problem (7.26) is also basically stable with probability 1 — e. Since the solution polyhedron in problem (7.26) is deterministic and the basic stability implies the constancy of basic hyperplanes for some subset of the set of all realizations to, the basic stability of problem (7.26) is equivalent to the absolute solution stability of problem (7.26). So, determining the basic stability region for the direct problem (7.25) we simultaneously determine the probability that the problem (7.26) is absolutely solution-stable [2] provides several ways in which such a probability can be computed. We will present another method of evalution which allows one to determine the probability a that the problem (7.26) is absolutely solution-stable, and hence basically stable. Let us fix all the basic constraints in the mean which become an exact equality. Construct the basic As matrix m x m. Then the probability that the problem is absolutely solution-stable is determined by those realizations for which b = b(uj) £ K/3, where Kp is the cone stretched along the normals with nonnegative coefficients fii,/3 = (/?i, /?2, ■ ■ ■ ,/3m) for which the respective constraints form the basis, and for the solution in the mean we have b = 0A&, where /3i > 0, i = l , m . Hence the probability that b(u) € Kp is the probability P = bT{u)Aj1
=bT{u)A>0,
where A = (a^j) = AJ1 is the inverse to matrix A$, i. e. we have a = P | 0 {aijbi(uj) + a2jb2{uj) + . . . -f a mj 6 m (w) > 0} | In practical problems, however, it is generally possible to determine the boundaries for variations in random variables b»(w), and these boundaries are finite. Then we have: a = P{ n {a!j6i(u>) + a2jb2{uj) + ... + amjbm(tv) > 0} , K < bi(u) < bf,
Vi = I ~ ^ } = P {&M € D} ,
where D is the intersection of the parallelepiped [b~ ,6+] and the feasibility region generated by half-spaces al]b1+a2]b2
+ ... + am]bm>0,
Vj = T^m,
b G Em
Discussion of the absolute solution stability leads to determination of region D. If the joint distribution function of vector components b(u>) is known, then by the linearity of all constraints the probability P{{LO) £ D} can be found by means of
136
CHAPTER 7. STABILITY OF SOLUTIONS IN SPP
complex integration over the polyhedral region of the joint distribution function of random variables bt(u>), i = l , m . Suppose the random variables 6,(w), i = l , m are mutually independent and are all uniformly distributed. Then it follows that / (n) fdbi.. a-
P(b(uj) e D)
.dbm
mAbt-b-)
where the region D is determined from the relation D = {b{uj) :b(u) € {[K,bf],bi
< 6+,V/ = I ^ } n
D{Siib1(uj) + a2]b2{uj) + ... + amjbm{u) > 0,
v» = T7^}} We will now consider the problem for each realization of LO: ' minx F(X) A{LO)X > b(u)
(7.27)
X >0 where F(X) is the deterministic, convex down function, and the conditions of problem form a random polyhedron 5(w). We need to find the probability that some point continues to be optimal for all realizations of w. In the feasible region, the minimum of convex function is reached either at the point of the absolute minimum of this function, if the function exists and belongs to some feasible regions S(u>), or at the boundary of the feasible region. For this reason the issue of absolute solution stability can be discussed in the sense of the above definition only as far as the absolute minimum point of convex function F(X) is concerned, Suppose the absolute minimum is reached by F(x) on x° > 0. Consider the case where x° belongs to the region S(UJ) that is feasible in the mean. Let the parameters of various constraints be all mutually independent. Then we need to make computations under the assumption that the parameters are all continuously distributed:
a = PJf>y(w)*J>&,(w),i = l,2,...,: P | f\lgi(ati,aa,...,ain,bi)
= H*) - X » K
<°
}=
m
= I I f{Si(«ii.««>•■ -,ain,b,) < 0} = ;=i m
= 11
f J
f (" + 1 ) / Pi(Ui,ai2,
...,ain,bi)daiu...,daindbi,
7.5. INVESTIGATION
OF ABSOLUTE
SOLUTION
137
STABILITY
where P ^ a n . a , ^ , . . . , a m , 6,) is the joint distribution density of random vector (an,o.i2, ■ ■ ■ ,a,„b,). Similar relations can be derived when problem (7.27) has constraints of the form 5i( :r > w )) * = 1,2,.. . ,m, where gi(x,u>) are functions that are concave in x, and when random variables are discretely distributed. We will consider a special case and find an explicit expression for the probability a when only the components of vector b = b(u>) in problem (7.27) are random. Suppose that for any i = 1,2,... , m we have n
Y^aijx°j > bi(CJ), x > 0. 3=1
1. Suppose bi(ui), i = l,m assume discrete values y from the set of realizations [fe~, 6^" ]. If we introduce the notation 6° = min \bf J2 aijx°j\, then we need to com puter a = P | &,H < X > t J x ° , z = T~^
=
£
...
■■•IZ-y™ e [6~,6+]P {b,(u>) = yi,...,6 m (w) = y m } , which, in the case of independence of vector components bi(u>), is as follows
« = E-we[br,6f]-..
£
f[ P {b,H = y>) = \
£
Pa).
where p,-2 is the probability that bt(ui) = j / l 2 . 2. Let b,(u>) denote independent random variables that are continuously distributed over the intervals [b~,bf]. If p(bi, 6 2 , . . . , bm) is the joint distribution density of a random vector b(u>), then for the same notation of b° we have
r
_
a = P < &j(w) < £ a , j 2 : ° , i = 1,) I ;=i = /
.
Jb,(uj)
/
( m ) /?(&!,& 2 ,...,& m )d& 1 ,dc> 2 ...d& m = J
(m) fp(bub2,...,bm)db,db2,...,dbm
= | ( m ) | Pi(6i)P2(6 2 )... P m (6 m )d6i, d&2, ...,dbm
=
=
138
CHAPTER 7. STABILITY
OF SOLUTIONS
IN SPP
b°
m m
• <
(bt)dbt. = n / Pi(bi)dbi.
K EXAMPLE. Suppose all 6,(CJ) are independently distributed with the mean values b, and variances a,, i = l , m and with probability distributions similar to normal distributions: i
/
C P
a,y2^ " \ Pi(bi) = •
1
..0, o,
-\m
'-U'f)>
6+
{ 67
b,e[b7,bt]
ki[br,ht}, b,{[b-Xl
r a n d o m variable variable t) is normalwhere the random each i there there is aa rrelevant e levant random normalwhere ran dom variable variable t (for (for each ly distributed with characteristics &,■(&,*-
n bf, £ aijX°A, a x° }, < bt) and <Ji. Let 6° = min | bf, E tJ 3
then have then we have
f
n
)
m h°
I
1=1
J
*=i»r
=\ f[fp OL - P k{uj) I bifo) << J2 J2 *j*% &ii2%i i==T >l,m\ a=P{ = IIi(b/ i)db #( W i = = 1
m
a^J
[
=nn — h ,=1
-\(^)h ——■
/ P
K
1
:-my
Replacing variables in the numerator and denominator, we obtain 1i
- = a = 1
1
(2
"
f~^~ e -•-. d* , L-^ Ib~-b% e~~dt J -1 r~ = b -b ,2 L /•-^r- e'^dt _£ +
*(«
?)- * ( ^ ) = n -y.(sa)_,(si).(*7 )- * ( ^ ') m
:
s
where $(s) = - 4 - / e'~dt ^ -oo
is the normal distribution function with characteristics
0 and 1; this function can be computed for any x from the tables.
7.6. STABILITY
IN PROBABILITY
139
MEASURE
If the stochastic programming problem is absolutely solution-stable with proba bility a, then, since the optimal solution has the same objective function value for all relevant realizations of CJ, we may say that the problem is absolutely and functionally stable with at least probability a.
7.6
Stability in Probability Measure
The solution stability in probability measure for stochastic problem is discussed in [11]. The optimal solution is considered as the probability distribution function. Their Gateaux derivatives are used to find asymptotic distribution of statistic estimates for the optimal solution. Let SI, B, p be a probability space. Consider the problem associated with functions /,: RK X SI -» R mmtp0(x) ipi(x) = 0,
i £ I,
(fi(x) < 0,
i e J,
where ipt(x) = Ep{fi(x,uj)}
= J ft(x,u>)P(doj). I,J are finite sets of indices. It is n assumed that for any x f{(x,u>) is measurable and integrable with respect to the probability measure under consideration. A large class of stochastic problems can be formulated in terms of (P). The behavior of solutions to problem (P) x(P) is treated depending on the probability measure P Suppose that P is disturbed by some measure Q P := P + Q, where Q is in the same space [SI, B]. Consider Pt = (1 - f)P + t(Q -P),Q0,
i € I,
tfi, < 0,
i € J}
where f,{x,t) = fQ f,{x,u>)Pt(duj) = ip,(x) + t{rjt(x) -)Q(duj). Let x(Pt) denote the optimal solution of Pt. Consider the Gateaux derivatives in the optimal solution direction dx(P, Q — P) = hm v t-n-o t A s s u m p t i o n 1. Let y?i(x), -qt{x) i € {0} U / U J be continuously differentiable in the neighborhood of point x0. To satisfy Assumption 1, we may require, e.g., that: 1) for all u : f,{x,u>) be continuously differentiable in the neighborhood U(x0) of the point x0; 2) there be the functions gt{u>) : ||V x /,-(a:,w)|| < g,{u), Vx g U(x0), w g 0 that are integrable with respect to P and Q. Then it follows that
140
CHAPTER 7. STABILITY
OF SOLUTIONS
IN SPP
1) gradient vectors V
with (Pj). Denote L0(x,X) = Z/(z, A,0) and L0 as the corresponding set of Lagrange multipliers satisfying the necessary (Kuhn-Tucker) first-order conditions. A £ L0 if V x L 0 (zoA) = 0, A, > OVi 6 J* and A, = OVz € J/J* When M F conditions are satisfied, LQ is nonempty and bounded. Assumption 3. Let ip,(x), Vi £ {0} U III J* be twice continuously differentiable in the neighborhood of x0. Consider the set LUQ) = arg max { £ A,(/3,(xo) : A G L0 \ and the so called crit1 ieiuj ' T ical cone c = ctU : u Vipt(x0) = 0, i G /, uTV5;(a:o) < 0, i G {0} U J* corresponding to the solution x0 of problem PQ L*0(Q) depends on Q via n,. L*Q(Q) can be given as the set of points of maximum —*-&—'■—<- with respect to L0. A s s u m p t i o n 4. For any u G C, u / 0 max uTVl L0(x0, \)u > 0. When Assumptions MF and 4 are satisfied, x0 becomes an isolated local solution of Po- The conditions of Assumption 4 are generally stronger than standard sufficient conditions (where LQ(Q) is replaced by L0). Assumption 4 is the weakest one and guarantees the existence of the directional derivative. Denote: for A G LQ consider V^ ( L(x 0 , A, 0) = Vn 0 (x 0 ) + J2 ^ V J ; , ( I 0 ) and a pointwise ieiuj maximum /3(u,Q) = max{uTV£,L(:roA0) 4- 1 l2uTVxxLQ(x0\)u A G L*0Q\ quadratic functions corresponding to the quadratic term of Taylor expansion for L(x, A, t). Y1{Q) is the optimal solution set of the problem minu
\7ip0(x0)
uTV^>i(x0) + ^(^o) = 0 i £ I uTVVt(x0)
+ rn{x0) < 0
i € J*
It is apparent that the critical cone C is a recession cone for -={Q) and —(Q) = C ifnt{x0) = 0 , VzG / U J' Theorem 7.8. Let Assumptions 1-4 be satisfied. Then: 1) 3 is a constant K > 0 : ||x(P ( ) - x0\\ < Kt, 2) if, additionally, fl(u,Q) has a unique point of minimum u{Q) with respect to ={Q), then the directional derivative dox(P, Q - P) exists and dox(P, Q — P) = ou(Q). Note that the set of points of minimum (3(u, Q) with respect to —(Q) is nonempty and closed, but is not necessarily a point. To satisfy condition 2), we may require that various forms of sufficient 2nd order conditions (stronger than Assumption 4) be satisfied.
7.6. STABILITY
IN PROBABILITY
141
MEASURE
Consider asymptotic distribution of statistical estimates for the optimal solution of (P). Let Lo-i . . ., u)n be some collection of independent, uniformly distributed obser vations with a common probability measure P. Problem Pn\ (P„)min>on(x)
S.tipin[x) = 0
iel
ipin(x) < 0
i £ J,
n
where V'm(x) = n _ 1 E / ; ( s . w,). Let xn be the optimal solution of (Pn). This solution ;=i
provides a statisfical estimate for the solution xQ of (P). Here the directional derivative is used as a heuristic means for determining asymptotic distribution of the solution of(P). ^ Consider xn as a function of the empirical measure associated with the collection under study. Consider the measures Qn = n~l E A(ujj), where A(w) is the probability measure 3=1
of mass one at the point u>. Then ipln(x) = F Q „ { / ; ( X , w)} = / fi(x,ui)Qn(dui)
and
xn = x(Qn) can be viewed as the function of Qn. Let x(P) = x0 and dx(P, Qn — P) exist; then W(Qn) = x0 + dx(P, Qn — P)-\-rern(Qn — P) (1). Suppose the residual term on the right-hand side of (1) is asymptotically negligible, which implies (Qn — P) = 0 p ( n - 1 ' 2 ) (2), i.e., n ' ^ r e m tends to 0 in the sense of probability. The asymptotic distribution n 1//2 (x n — x 0 ) is then completely determined by nll2dx(P,Qn — P) So the asymptotics n1^2dox(-, •) can be readily described by employing Theorem 1. (2) is more complicated to derive. A s s u m p t i o n 5. A0 = {A0} is a point. Then fi{u,Q) is the quadratic function (3(u,Q) = u \Vrjo(x0,Q) + E .S/uJ
AoV77,-(z 0) Q)l+i« T ViX(xoA 0 ) J
uTV
i £ / U J;(A 0 ) ieJ;(X0)
where J('(A) = {i £ J* : A, > 0}; J0*(A) = {i € J* : A, = 0}. The following form of sufficient 2nd order conditions is provided to ensure the uniqueness of the point of minimum u{Q). Consider linear space: M = {« : uTV0. Since M contains the critical cone G, Assumption 6 is stronger than Assumption 4. A s s u m p t i o n 7. Expectation Ep{fi(x0,uj)2} i g /UJ* is Unite. Since rn(x0,Qn) = - E fi(xa,Wj), by the central limit theoremn1/2{rii(xo,Qn)~<{>i{xo)} witi respect to distribution. Consider the Lagrangian functional, A, w) = / 0 (x,w)+ E
i6/uJ
the expectation 1(X,\,LO)
in measure P is L 0 (x, A).
tends to normal
Kf,(x,uj).
Evidently,
142
CHAPTER 7. STABILITY Assumption
OF SOLUTIONS
IN SPP
8.
For almost any ui l(-,X0,u>) are differentiable in XQ: are finite and VL0(x0X0) = Ep{VJ(x0)\0,uj)y Consider random variables, V; i £ /U J* and Z = (zi,. . . ,Zk)T, that are normally distributed with zero expectations and covariances: cov(YtYj) = Ep{f,(x0uj)fj(x0uj)}
cov(Z,Zj) = Ep{...cov(YlZJ)
= uTZ + \uTVlxL0{x0]\0)u;
Define p(u,Z)
= Epj
£ ( V ) = {y ■ uTV
T
i £ / U J*(A0); u \7ipi(x0) + Y, < 0, i £ Jo(A 0 )j and let u(Y, Z) be the set of points of minimum /?(£/, Z) with respect to ^Z(Y). T h e o r e m 7.9. Let Assumptions 1-3, 5-8 and regularity conditions (2) =$■ nll2(xn — x0) —>■ u(Y, Z) be satisfied.
Chapter 8 Methods for Solving Infinite and Semi-infinite Programming Problems Application of mathematical modeling and analysis techniques to economy, techno logy, design, and management results in infinite-dimensional systems of inequalities with a finite number of variables [67]. This list may include engineering, design, variational inequalities, theory of moments, continuous linear programming, geometric programming, and theory of fuzzy sets. Formation and development of semi-infinite programming, as a division of mathe matical programming, were governed by statement and necessity to solve mathemat ical programming problems in finite-dimensional spaces with an infinite number of conditions or with a finite number of conditions but in infinite-dimensional (function) abstract spaces [46, 72, 123). Semi-infinite programming problems can be examined by employing a variety of reductions to a finite case such as finite systems of infinite subsystems or finite probability measures. A natural extension of semi-infinite programming is infinite programming by which are meant statements and methods of solving the programming problems with an infinite number of conditions in function spaces [61, 63, 37, 108, 3, 34, 81]. What are called as large-dimensional problems [153, 132, 54, 56, 67, 68, 71, 72, 115] and decision problems under uncertainty, i.e. stochastic programming problems [146-148] have made a contribution to development of the theory and methods for solving infinite and semi-infinite programming problems. Applications of mathematical programming techniques to planning and control in complex stochastic systems provide an infinite variety of manifestations of random data (where the number of states of nature is infinite), which is reflected in the models with an infinite number of conditions and/or an uncountable set of desired variables. This generates a need for examination of infinite and semi-infinite programming problems. 143
144
8.1
CHAPTER 8. INFINITE AND SEMI-INFINITE
PROGRAMMING
Statement of Semi-infinite Programming Problems
The term "semi-infinite programming" was defined in [14, 29, 103]. Here, we are dealing with constraints on a compact set. Formally, these are finite-dimensional problems with an infinite number of constraints. In the general case, the semi-infinite programming problem (SIPP) can be stated as follows: mf{iP(x):ft(x)
<0- i = T~^}.
(8.2)
x G X is the Hilbert space. In special cases, the problems (8.1), (8.2) can be stated more specifically. Thus, the statement is: 1. The form and properties of the objective functional are: linearity, quadratic form, convexity, and degree of smoothness. 2. Construction of the set for problem (8.1) is: (If T is finite, we have a classical problem of mathematical programming) T — {1, . . . , oo} is countable, T = [0,1]; T = [a,P]; T=fl
.=i
T,; T, = fa,ft], V* = T ^ .
3. We may select, along with constraints of the form ft(x) < 0, a finite number of "ordinary" constraints fi(x) < 0, i = l,m, or x £ X, where X is a compact set in Rn 4. The form and properties of constraint functions are: linearity, quadratic form, convexity, and smoothness. Referring to [109], we have the general statement of problem (8.1) on the as sumption that tp(x), ft{x) are convex in Rn For the general statement of (8.2), see [41]. The union of two types of "semi-infinite dimensionality" is exemplified by setting a separably-infinite programming problem [31, 67]: v(A) = inf ex - £ (A)
v
m+i{r)i]{r)\
u(t)x>un+i(t), t e S; Ax+ £ u(r)??(r) = 6;
. [
. '
>-eQ
where S C /?", Q C R',u{-): S - » Rn, un+1(-): S ^ R, c e Rn, b € Rm,v(-):Q^ Rn, : mxn w m +i(') Q ^f R, A e R The first constraints in (8.3) contain an infinite number
8.1. STATEMENT
OF SEMI-INFINITE
PROGRAMMING
145
PROBLEMS
of inequalities with a finite number of variables, whereas the second constrains contain an infinite number of variables, but their number is finite. Referring to [32], we have the problem: (5)
ram{/(i):V,(i,()>0;
Vf € TU
i€l^},
(8.4)
m
which only differs from (8.1) by its special construction, T = Y\ If i-i
The solution of a wide class of the analysis and synthesis problems for discrete dynamical systems in a half-closed interval can be reduced to the solution of the nonlinear programming problem [25-27]: inf{J(u):u 6 Rn, G]t(u) > 0,
Vf 6 l^oo,Vj 6 V m } .
(8.5)
Here J plays the role of ip, G plays the role of / , while T = {1,. . . oo} x {1, . . . m} is the product of the countable set and the finite set. Referring to [110], we have a convex quadratic semi-infinite problem: (Q)
m'm{-xTCx-pTx:g{x,t)<0,
teT,x€X},
(8.6)
where X = {x £ Rn: fj{x) < 0 are convex, j € J = {1, .. ., }}. Here i/i is a quadratic function, while the "ordinary" constraints fj(x) < 0 can be written in a semi-infinite programming form: fj{x) = gj(x, t), Vi G T, though it is more advantageous to isolate them as a special group. Finally, two examples of a semi-infinite analog may be provided for the linear programming problem. These examples only differ in the condition for membership in A' [28-30, 53, 55-56, 62, 65, 87]: l r j
in [62]:
I
f ,(F) = inf(c^); \ atx>bt, teT, _ v(Z) = mine x; u(t)x>X(t), VieT;
{
'
(8.8)
Vi 6 I C ff X:{x:a,iX>b, i = l,m}. Here 4>(x):cTx, ft{x):bt - at(x). In the general case (where linear programming problems are extended to cover multiple specification of constraint matrices), such problems are of the form [153]: ma,x{cTx;x > 0; Ax < b VA € G}, (8.9) where x = ( i i , ■■•, xn)T is the column vector of program components, cT = (ci, . .., c„) is the row vector of objective coefficients, A = {an, i 6 l , m , j G l , n } is a constraint matrix, rn < n, b = (bu .. . , bm)T is the column vector on the right-hand side, and G is the matrix set.
146
CHAPTER 8. INFINITE AND SEMI-INFINITE
8.2
PROGRAMMING
Duality of Semi-infinite Problems
Referring to [24], we have that the dual of a semi-infinite problem also is semi-infinite. Thus, the dual of (8.7) becomes [28, 53, 55]:
I
v(D) = sup J2 \{bi = supt/>(A);
EA,* = &><>,
teT.
M
Theorem [28] is given to define the duality of problems (P) and (D). Theorem 8.1. 1) Ifv(P) = — oo, then problem (D) is inconsistent; 2) ifv(D) = +oo, then problem (P) is inconsistent; 3) if problems (P) and (D) are consistent, then v(P) = v(D). The admissible set for A is a convex set in the space of generalized finite se quences and retains the properties of convex polyhedral sets in finite-dimensional spaces. These properties provide a framework for interpreting the simplex method geometrically. Referring to [53], we have the duality theorems using Slaiter condition (existence of interior points of an admissible set). Theorem 8.2. [53] Suppose that: 1) the vector set dt = (aJ,bt)T 6 Rn+l is canonically closed; 2) there exists x* 6 Rn such that atx* > bt, Vi 6 T. Now, ifv(P) is finite, then v(P) = v(D) and the problem is solvable. We have general properties of the optimal value function v(P)(c) and v(D)(c). The statement is: v(P) = cl v(D). In particular, consideration is being given to a duality gap and its relation to discrete optimization of semi-infinite problems. Conditions are provided to eliminate a duality gap by introducing a perturbation into the objective function. In terms of the Lagrange function [23, 30, 38, 39], for general convex programming problems MP = inf{f(x):gi(x) < 0, i S / } their Lagrangian dual is: MD = {supinfi(x,A),A > 0, A
x
A € A C {A,: i e I, A,: / 0 is a finite number }};
(8.11)
L{x,\) = f{x) + '£\igi{x). is/
We have that MP — MD if and only if MP = optimal value
lim MP$, where A'fPs is an
J->+o
inf :r (PA I /( )i { s> 1 »(x) < S, i el.
Suppose C is a feasible domain for problem P, Cs is a feasible domain for problem Ps, and 0 + C is a recession cone C. L = {y. f0+(y) = 0 = / 0 + ( - y ) } ;
M = 0 + C A (-0+C) A L.
8.2. DUALITY
OF SEMI-INFINITE
Proposition.
147
PROBLEMS
Let f, gt, Mi be a closed convex C ^ 0, then MPS -¥ MD as
T h e o r e m 8.3. Let C ^ 0 and /0+(j/) > 0, Vy e 0+C n M1, y ^ 0 =» MP = MD and there exists an optimal solution P. Let I0 be a countable subset of / such that {x\g, < 0, i £ I0} = {x\g,(x) < 0, i e I}; then problem P(I0): MP(I0) = {inf f(x)\gt{x) < 0, i £ / 0 } is the equivalent of problem P in terms of the optimal value and feasible solution. Let us construct problems Pn and their dual Dn: (P„) {Dn)
MPn = {inf f(x)\9i(x)
< 0,: = M } ;
MD.n = < sup inf Ln(x, A), A > 0, A £ /T I >■ x
L„(x,A) = f{x) + J2^9i-
Construct C„: C„+i Q C„, Vn, C = HCnI
T h e o r e m 8.4. Let C / 0 and /0 + (y) > 0, Vy £ O + C f l M 1 , y ^ 0, then MP„ = MDn for any sufficiently great n and MPn ->■ M P The dual of problem (Z) (8.8) for a particular specification of the set X be comes [62]: v(F) (F)
suP( £ A(%(*) + £ &.•«<); £u(%(<)+ £ .iiV, i>j > 0;
= c;
(8.12)
! £ l,m;
tp(t) > 0 for Vi £ T and (^(t) = 0 for any only finite number t £ T. Referring to [153, 132, 133], we have linear programming problems with the con straint matrix (8.9) specified multiply. Duality theorem follows from the assumption of convexity, closure and boundedness of the set G out of which the matrices A are selected. Theorem 8.5. If one of the problems — primal or dual — has a limited solution, then the other problem also has a limited solution, and ex = ipb. If the target functional in one of the problems is not bounded on a suitable set of feasible solutions, then the feasible solution set of the other problem is empty. Here x = argminjez: x £ X} and y = argmax{j/6: y £ Y} are optimal values for the primal and the dual problem on the respective feasible solution sets X and Y generated by constraints on the problems. This theorem immediately follows from the analog of the Minkovski-Farkas theo rem for infinite systems of linear inequalities. The dual of problem (8.3) is also a separably-infinite programming problem [31, 67] v(B) = s u p ( £ un+i(t)X(i) - by); tes £ u(t)X[t) - A*y = c; .13) [B) v(r)y > u m + i(r), Vr £ Q; A(-) > 0; AM G fl<s>; y G BP
148
CHAPTER 8. INFINITE AND SEMI-INFINITE
PROGRAMMING
The duality theorem is provided under the following assumptions: Assumption 1. The set Kg = {x £ Rn:u(t)x > un+l(t); Vt £ S} is nonempty and bounded or u„+i(-) = 0. Assumption 2. The set KQ = {y £ Rm: v(r)y > vm+i(r); Vr £ Q} is nonempty and bounded. Assumption 3.. The convex cone Cs C Rn+1 generated by
..A)H u {(: is closed. Theorem 8.6. [31] If problem (A) (8.3) is solvable and has a Unite value, then problem (B) (8.13) is solvable and v(A) = v(B). Furthermore, problem (B) (8.13) admits of its supremum as a maximum. An assertion similar to this theorem can be made under assumptions such as 1-3. Assumption 1*. The set KQ is nonempty and bounded or D m + i() = 0. Assumption 2*. The set K$ is nonempty and bounded. Assumption 3*. The convex cone CQ C Rm+1 generated by
..-^ M H«M(! is closed. Referring to [39, 87, 92], we have the infinite systems defined by convex functions (possibly by adjoining constraints on membership in some set). These functions and sets are generally assumed to be closed, but the Slaiter points are not required to exist. In particular, if there are no constraints on membership in the set, then the constraints are only required to be consistent. Necessary and sufficient conditions are given for the lack of Lagrangian duality gap (8.10) which contains merely a finite number of nonzero multipliers [95]. In particular, the conventional sufficient condition has been derived for a finite number of constraints. For the infinite number of constraints, a compactness condition plays the role of "finiteness'' in the function space. Semiinfinite programming problems can be treated by employing a variety of reductions to a finite case such as finite systems of infinite subsystems or finite probability measures (see, e.g., [88]).
8.3
Optimality Conditions for Semi-infinite Programming Problems
Referring to [140], we have the Farkas theorem for a finite case and, as a result, necessary minimum conditions for problem (8.2). Theorem 8.7. Let F(x), ipi(x), i £ l , n be continuously Frechet differentiable. Suppose the point x0 is a solution to the problem, and weak first-order regularity
8.3. OPTIMALITY
CONDITIONS
FOR SIP
PROBLEMS
11!)
conditions are satisfied at this point. Then n
F'{x0) = ^ c w ' O o ) The constraints {<^} satisfy the weak first-order regularity condition at the point So if any vector z ^ 0: {0;
i € I(x0) = {i e Y^n:
is slightly tangent to the set A at the point x0. The vector z ^ 0 is slightly tangent to the set A at the point x if there exist the sequences {xn}: xn € A, n 6 1,oo and {An}: A„ £ /? are real numbers such that An(zn — x)T —>• z as n —)• co (weak convergence). Necessary and sufficient minimum conditions [109] are given for problem (8.1) in terms of Lagrangian function. Theorem 8.8. x 6 R is an optimal solution to problem (8.1) if and only if x € 5 s u c n t ia and there exists (X(t))ter(x) £ ^i ^ '
On €
ItSfi(S).
teT(x)
Here the Lagrangian function becomes
*(*, A) =^(A)+ £ > / , ( * ) , /?
A = (A t ), eT e R\ ;
T(x) = {teT:
iGfl"; Vi}; ft(x) = 0}.
The necessary local minimum condition [23, 32] is given for the problems (S) (8.4). Theorem 8.9. [23, 32] Let x" be a local minimum in problem (S) and I, = {t*k g T,.- ipi(x",t*k) = 0; k S 1,A',}; then there exist finite Lagrange multipliers Xlk > 0 m
such that VJ(x')
K,
= E E A ifc V l¥ J,(i*,i^).
A s s u m p t i o n 1. The gradients V # ( i * , t ) for any t S TJ; t £ l , m are linearly independent. Accordingly, the sets /, are necessarily finite and we may write
S = {t»eZi:p.-(*V,\) = 0;
feei.Xi}-
A s s u m p t i o n 2. For any a; there is a finite (possibly empty) set of subsets £lij(x) such that: 1. Hi, GT„- 1 <j <St = S,{x) < o o ; 5,
2. v3,-(x,t) < 0; Vt € n t J and p,-(x,i) > 0; Vi S T, \ U % ( * ) ;
150
CHAPTER 8. INFINITE AND SEMI-INFINITE
PROGRAMMING
3. %(x)nfl;*(z) = { 0 } ; J 7 ^ ; 4. flij(x) is connected and nontrivial, i.e.:
/
dt > 0.
Note that almost all functions will satisfy this assumption. Assumption 3. For any x and any index i there is a nonopen set {/,• that is strictly bounded in T,- and suci thai 73,(3:, t) = 0; Vt € £/;. Denoting A, ; (x) = / dt; $ t J = / tpi(x,t)dt, we may write the penalty n„(x) «,,(*) function as p(x lM) = = /^/(x) -
((*)),
-E(E*. !=1
J=l
3=1
where p is a positive scalar. Assumption 4. Trie functions f(x) are convex and ipi(x,t), t £ l , m are concave in x. Assumption 5. For Vi £ 1, m; j € 1, Si there exists a constant ft > 0 suci that s, 5,
EM*)
s,
< P^{x.
3=1 3=1
,(*) 7=1 i=i
s, for any t g (J n s j ( i ) and any x £ Rn 3=1
Theorem 8.10. [32] Under the assumptions 1-5 the point x* is a global minimum of the function p(x,fi) for any /J. such that 0 < fi < p* for some p' > 0. A series of assumptions and assertions similar to Theorem 8.10 are made for the general case [32]. In [30, 31], for problem (8.7), some generalizations of the Kuhn-Tucker saddle point theorems are examined for arbitrary convex functions, and necessary and suffi cient optimality conditions are obtained under some additional conditions. Sufficient optimality conditions for solution in terms of various notions related to the objective function (generalized Lagrangian function, a direction cone, a special duality problem, etc.) are guaranteed by the boundedness of a feasible set [58]. Semi-infinite programs are investigated using penalty functions and Lagrangian methods Necessary optimality conditions are constructed in the form of a system of nonlinear equations. Also, generalization is given to cover a nonconvex case. Referring to [89], we have two sufficient conditions for strict local optimality that do not require strict complementability for the feasible set to be reduced. It is stressed that these sufficient conditions are the natural extension of typical sufficient secondorder conditions to a finite case. Referring to [31], we have necessary optimality conditions for the primal (A) (8.3) and the dual separably-infinite programming problem (B) (8.13). Theorem 8.11. Let x", rf{-) be optimal for problem (A) and let y*, A*() be optimal for problem (B). Then ( u ( 0 s , - u „ + i ( t ) ) A * ( 0 = 0,
V i e S;
8.4. EXISTENCE
AND UNIQUENESS OF SIP
PROBLEMS
(v(r)y*-vn+l{r))ri'(t)=0,
Vr £ Q.
151
Referring to [36], we have a wide class of problems (semi-infinite programming problems, in which the minimized function and the functions defining the inequalitytype constraints are nonsmooth and represented as a maximum of the set of smooth functions on a compact set) for which there is a set of quadratic necessary optimality conditions, where these conditions are interrelated as closely as in classical analysis. Second-order necessary conditions are also given in [121]. As noted in [96], the wellknown second-order necessary conditions for semi-infinite programming problems are valid only where, along with the system of inequalities, the problem incorporates some constraints in the form of inequalities. If there are no equality constraints, then an envelope-like effect occurs. Allowing for this effect results in second-order necessary conditions for the additional summand. Referring to [69], we have new generalizations of the Farkas lemma. These results allow some familiar infinite analogs of the Farkas lemma to be combined within a unified theory. Applications are given resulting in necessary conditions. Referring to [22], we have the marquee method which is generalized in a finite case [19] to an infinite-dimensional situation. The method is also given for the infinitedimensional case. As an application, necessary optimality conditions are provided for the general programming problem. Basically, they include all of the familiar necessary conditions (in theorems such as the Kuhn-Tucker theorem, etc.).
8.4
Existence and Uniqueness of Semi-infinite Programming Problems
Let us consider a linear semi-infinite problem of the form (8.7) [75, 112, 134-136], 1
'
f mines; \ atx
Suppose that T is a compact metric space, B(t), A, T is the set of all continuous mappings B:T -> R and A:T -* Rn for the given parameters a = {c,a,b} G 9, consideration is being given to a suitable (P1). We define the set of feasible points Z = {x G Rn: a,x < bt ,Vi € T} and the solution set P = {x G Z: ex = inf(cy: y e Z)}. For every feasible point x G Z, we define the set of active points N = {t G T: atx = bt}. The vector x0 is called a strictly unique solution of (P1) if there exists a constant K > 0: Vx g Z: cx0 > ex + K\\x - x0\\. Problem (P1) satisfies Slaiter's condition if there exists a vector y g Rn such that aty < bt, Vi g T. Suppose that Q = {a: (P1) satisfies Slaiter's condition }; L = {a: (P1) has a solution }; S = {a: (P') has no more than one solution };
152
CHAPTER 8. INFINITE AND SEMI-INFINITE
PROGRAMMING
conv (A) is a convex hull A C -ft"; cone (A) is the convex cone generated by A. Strict uniqueness is investigated in [75], T h e o r e m 8.12. If a, fixed pair (c,a) satisfies Haar's condition (i.e. there are no points t, g T, i g l , n — 1 such that the vectors c,atl, ..., atn_1 are linearly independent), then for any b G B(t) the optimization problem has a strictly unique solution. Let us select a point to, but not in T, and let ato = c. Theorem 8.13. [75] Let a g Q and x0 G Z. If there exist points ti € T, i G l,n such that - c G cone {{au:i G l,n}) and a(o, a t l , . . ., a(,_,, a ii+1 , . . . , atn are linearly independent, i g Q,n, then x0 is a strictly unique solution of(P'). In practice, the parameter which defines problem (ft') is often known approximate ly or it may vary. Complete characterization is provided in [111] for linear problems which have a strongly unique solution and are stable under small variations in the parameter. Consideration is given to the relationship between the unique and the strongly unique solution. Consideration is given to global unicity [112]. Also, characterization is given to the problems which have a unique and a strongly (strictly) unique solution for all constraint functions. Theorem 8.14. [112] For a fixed vector c g ft71; c ^ 0 and a fixed map a g A(T, ft"), o ^ O , the following assertions are equivalent: 1) for any b g B(T) with a = (c, a, b) g L, problem (P1) has a unique solution; 2) for any b g B(T) with a g L problem (ft') has a strictly unique solution; 3) there are no points t, G T, i — l , n — 1 such that —c g cone ({o (l , i £ l,n — 1)}; 4) for any points ti g T, i g l,rz — 1 ' a vector y g ft" such that cy > 0 and a (i y > 0, i g l , n — 1. , Effective characterization of strict un of solutions is obtained for the minimax problem {y{x) = max{a ( i — bt] —> m on the assumption that a = mi{g(x): x € ft"} is finite. Here a wide class of best approximation problems [16] and other problems is covered. Referring to [135], we have generic existence and unicity of the solution to the problem (ft'). For a = (c,a,b), we denote Lz = {a, Z / 0}; K = {x g ft", atx < 0, Vt g T}; K* = {c g ft", ex > 0, Vz g K}\ KT = {c g ft", c = E a,bu, t, g T, a, < 0, i=i
z g l,n}. Theorem 8.15. Let Z be nonempty for some a and cy > 0, and Vy g K, y / 0. Problem (P1) then has a solution. In [98], it is established that, for a compact Hausdorff space there is a dense Gj-set in 6 = {
8.5. METHODS AND ALGORITHMS
FOR SOLVING SIP PROBLEMS
153
T h e o r e m 8.16. The set, S is dense and Gs is a subset of 8. T h e o r e m 8.17. The set L* = {a: (Pr) has only one point } , is dense, and Gs is a subset of L. T h e o r e m 8.18. The set L" contains an open dense subset of L. In [136], generic uniqueness of the solution to problem (/-") is investigated as generic existence and uniqueness of saddle points of Lagrange function. L(x, A) = ex + X(atx — bt), where A £ C*(T), A > 0. Suppose M = {a: L(x,X) has only one saddle point }. T h e o r e m 8.19. If for some point x0 g Rn there is a point A0 € C*(T), A0 > 0 such that (x0,X0) is a saddle point L(x,\) in the region (x G Rn, A G C*(T), A > 0),then x0 is a solution to (P1). Conversely, let x0 G Rn be a solution to (P1) and suppose Slaiter's conditions hold for Z. Then there exists A0 6 C*{T), A0 > 0 such that (x 0 , A0) is a saddle point of the function L(x, A). T h e o r e m 8.20. Let T be a compact space, \T\ > n, then M is dense and Gs is a subset of L.
8.5
Methods and Algorithms for Solving Semi-infinite Programming Problems
Consideration is being given to linear problems of the form (P) (8.7) and their du al problems. A review of algorithms for solving linear semi-infinite programming problems [54, 76] includes locally convergent methods [46, 64, 66, 73], discretization methods [56, 57, 74, 75, 77, 94,
.;
the cone of the generalized positive finite sequences of real functions of R The set S C T is called basic if {a,, t G S} is basic in Rn and there exists A G A such that supp A C S. For the given basic set S with the associated A and x, two cases are possible:
(A): irdlatx - bt) = 0; teT
(B): 3u 6 T is such that aux — bu < 0. If S = {ti, ■ ■ ■, tn}, then there are uniquely determined real numbers at, . . . , an such that au = aiai + • • • + anan. For (B), there are two possibilities: (Bl): cti < 0, i € T~n; (B2): fi = min{^-; a, > 0} = -£-. Then we consider a new set
5' = (5\.fe})UWcT.
154
CHAPTER
8.
INFINITE
AND SEMI-INFINITE
PROGRAMMING
T h e o r e m 8 . 2 1 . [54] Suppose the following assertions are true: J. In case (A), x and A are respectively optimal solutions of (P) and (D), and ex = v(P) = v(D) = ip{\); 2. In case (Bl), (P) is unsolvable and lim ip(X + p,p) = + o o , V/i > 0 (here X+p,p is a feasible direction at the point X, p € R^). Furthermore, if au G {at, t g supp A} (e.g., if X is nonsingular), then {X + pp, p > 0} is the unbounded edge of A. 3. In case (B2), S' is the basic set and its respective point X' = X + pp £ A. In addition, if au £ {at, t 6 supp A}, then i/»(A) < ip(X') and [A, A'] is the bounded edge of A. At the r - t h step the simplex procedure either t e r m i n a t e s (A) or (Bl) or we have a new basic set SV+1 ^ Sr. If, however, Ar is singular, t h e n apparently A r + 1 = Ar, and hence we have what is called cycling. The simplex procedure generates in terms of the initial basic set S i , a finite or an infinite sequence of basic sets {Sr}. For the last basic set Sp in this sequence, we either have optimal solutions to (P) and (D), xp and XF, or state the inconsis tency of (P) and the unboundness of (D). For a finite sequence, lim ex' = infer and lim ip(Xr) = ip(X). Although the admissible set A is a convex set in the space r—>oo
of generic finite sequences, it retains m a n y properties of convex polyhedral sets in finite-dimensional spaces. These properties provide a framework for interpreting the leading transformation geometrically. Based on t h e dual approach [55], geometric interpretation is given to such notions as boundary points and feasible directions, edges and conjugate directions, etc. This allows the main iterations of the simplex algorithm to be shown geometrically. In [55], the ways of finding the basic solution are examined. T h e optimal solution of linear SIP problems can be computed in a va riety of ways. T h e most important of these are exchange algorithms (Dantzig simplex method) and descent algorithms (gradient methods). T h e fundamental exchange algorithm for linear semi-infinite programming prob lems is the simplex method. T h e exchange algorithm is a generalization of the simplex procedure which allows introduction into the basis of one or more vectors at an iter ation. Execution of exchange algorithms for linear SIP problems is m a d e possible by the solution of linear equations. The solution methods will be numerically stable. T h e pair of dual problems, (P) and (£>), is examined [118] under the following assumptions: 1. (P) and (D) have no duality gap, i.e. the optimal values of (P) and (D) are equal. 2. (D) is solvable or bounded. 3. There are n linearly independent vectors among {at, ( £ T } such that at = (ai(t),...,an(t)T). We call {T„, A} the basic solution (here Tn = {tx, . . . , t n } ) if the vectors a{tx), . . . , a(tn) are linearly independent and A is a solution to t h e dual problem (D), i.e. A{tx,.
..,tn)X
= c,
(8.14)
8.5. METHODS AND ALGORITHMS
FOR SOLVING SIP PROBLEMS
155
where A(tu . .., tn) = (a(ti), . .., a(tn)) is the basic matrix, T'„ is the basic set, and the vectors a(t{), i £ l , n form the basis. Accordingly, the solution of the primal problem (P) is A(tu...,tnfx=(b(tl),...,b(tn))T. (8.15) We now implement each step of the simplex method , i.e. we pass to the basic solution (T„, A), which corresponds to a greater value of the objective function, by changing only one component of the basis (here only one column A(tu ..., tn) varies). Suppose t* £ T enters into the basic set. The linear equations are then solved to obtain representation A(tu...,tn)d=a(t*) (8.16) d = {du ..., dnf We assume that A(6>) = (xi - 0du ..., xn - 9dn, 9)T, \(6) is a solution of equality constraints in problem (D) such that A,-^) > 0 , i £ l,n.
e=^ =
mm{^-!d,>0}.
If there is no d{ > 0, then (D) is unbounded and (P) is unsolvable. If 9 = ^, then the element tr is eliminated from the basic set. The new objective function value for X(8): c(9) = c(0) + 9(b(t*) -
JTar(t')xr). r=l
In this manner three systems of linear equations (8.14), (8.15), (8.16) are solved. Since only one column or row is changed at one step of the simplex method modifying the initial method, the decomposition methods are successful in reducing the number of operations by an order of magnitude as compared to the customary simplex method. The Jordan-Gauss methods of elimination and inverse matrix which are commonly employed to solve linear equations are unsuitable. To ensure stability, use is made of the modification: the Gauss method of elimination with search for the main element in the column, or decomposition methods. Applying to the matrix (AJ.E) the conventional Gauss process of elimination with search for the main element in the column, we obtain (R\F), where R is the upper triangular matrix and F is such that FA = R. On the other hand, given F and R, the linear equations (8.14), (8.15), (8.16) can be readily solved, e.g., for AX = c <£> FAX = Fc or RX = Fc. In this case, laboriousness is proportional to n2 operations. If the basic matrix is fully partitioned at each step, the number of operations is proportional to n3 The method modification: if A = (o(
FA =
0
*
0 0
156
CHAPTER 8. INFINITE AND SEMI-INFINITE
PROGRAMMING
In fact, comparing only two rows yields a new partitioning A: (R\F). Now the number of operations is proportional to n2 To introduce into the basis [118] two or more vectors at once, we need to consider the solution of system (8.16)
A{tu...,tn)d'
= a(t*),
i = Y7k,keN; k
d1 = ( 4 , • • •, < ) T It can easily be shown that X(9) = X(9U . . ., 9n) = (At - £ 9,d\, T
;)
•, A„ — JZ &id%n, di
is the solution of problem (D). The corresponding
value c{9) = c(0) + £ 9,(b(t*) - £ xraT(t])) .= 1
= c(0) + £ A(i*). We require that
r= l
t= l It
all A(£*) > Q, i £ l,n, an increase in the objective function c(0) — c(0) = £ A(t*) i=l
be a maximum and all components X{9) > 0. Then we get an auxiliary problem in standard dual formulation max^>A(t*) subject to - d\
. .
U •
d\-
"1 . . 0 "
' h'
z
l
" Ai '
n
A„
(8.17)
+
4.
_0 . . 1
Jk. >0,iGl,fc,
z, > 0 , i € l , n .
Using the above simplex method, and considering that we have an optimal basic * \T k' < k (any other variables are solution of the form ($i, ^fc'+i, equal to zero), the constraints of the auxiliary problem become:
d\9l + ... + dfOy = Aj <* Ai(0) = A! - d\Qx - ... - df9k, = 0;
di,&l + ... + d£,ek, = \k,^\k,(e)
= \k,-di,el-...-d£,ek,
= 0;
d)0l + ... + dk/9k, + z3 = Xj & X}{9) = zh k' < j < n,
i.e. the first k! components of 9 are located in the basis and correspond to a(i*), . . . , a(t'k,), while the first k1 components of A(#) are zeroed. In this manner we replace ti, ..., tki by t\, ..., t*k,. It can be shown that this method is successful in seeking a new basic solution of (D). We have the primal algorithm for the SIP analog of a linear programming problem of the (P) type, (8.7). a(t) x > b(t), x G Rn,
t G T is a polyhedron in R: a.b are continuous.
(8.19)
8.5. METHODS AND ALGORITHMS
FOR SOLVING SIP PROBLEMS
157
Some results have been obtained for the generalized LP problem. Their applica tions can be demonstrated by setting problem (8.19) in equivalent form, where a free variable is introduced cTx —> min; T a{i) x - z(t) = b(t); (8.20)
I
z(t)>0, zgC°°[0,l). Let (£, 0 be a feasible point for (8.20). It is placed in correspondence with the set of active points {t g T: ((t) = 0}. At the active points the free variable (greater than or equal to 0) has a strict local minimum equal to 0. To keep track of zero degrees, we define d(i). If {ii, . . . , tk} is the set of active points, then d(i) is the least integer j which has the property C"+1H*«) 7^ 0. Suppose d(i) is defined on Vi g l,k. We consider the properties of extreme points in the generalized LP problem which are k
transported to problem (8.20). Let m = k + "£, d(i). We construct an (m x n) matrix A-A = (a(f,), a'(ii), . . . , a (d(1 »(«i), ■ • •, a{ik), • ■ • > a^k\{tk))T The point (£, () is extreme for problem (8.20) if and only if the columns of A are linearly independent, which is equivalent to span {a'J>(i;): j = 0, . . . , d(i), i = 1, , . . , k} — Rn. The purification algorithm A is used to obtain from any starting feasible point an extreme point for problem (1.20). The process proceeds at any step so as to retain all previous zeros of the free variable until the new one is obtained. A s s u m p t i o n , {x: cTx < 0 and a(t)Tx > 0, T g [0,1]} = {0} (it is valid, in particular, when the feasible region is bounded). Algorithm A. Step 0: (f\Ci) is a feasible point for (8.20); r = 1. r-th iteration: Step 1. Let (ti, ..., tk) be active points for (£ r , (r); define d{i), i = \,k. Step 2. Define a subspace Tr C R: Tr = Rn if k = 0; otherwise Tr = {x: a^{t,)Tx = 0, j = l,d(i), Vi = 171}. If Tr = {0} — stop; ((r,(T) 's an extreme point. Step 3. g'~ = —Prr(c) (PTT is the orthogonal projection onto Tr); ifg' = 0 select gr ^ 0 arbitrary in Tr. Step 4. 0r = s u p { - [ a ( t ) V / C r W ] , t g [0, 1], t ± t%, . .., h}.
Step 5. C+i = f + U / / W ,
CH-I(')
= a(-) T f +1 " &(■)■
Step 6. r := r + 1, go to Step 1. It has been proved that, under the above assumptions, algorithm A yields an extreme point in at least n iterations, and the objective function value at this point does not exceed the one at the starting point. Algorithm B has been developed for problem (8.20), using a feasible extreme point at each step. The resulting points are taken to be nonsingular, i.e. the corresponding matrix A is invertible. Define n scalars \t]: cTA'1 = (Ai |0 , . . . , ^i,d{i), • • •, )>k,o, ■ ■ ■, ^k,d{k)) = ^ Prove the following optimality property: the nonsingular extreme point ({,() is optimal if and only if for Vz = l,k, A,-,o > 0, Ay = 0 , j = l,d(i), which is utilized in algorithm B.
158
CHAPTER 8. INFINITE AND SEMI-INFINITE
PROGRAMMING
Algorithm B. 1. Select (f°,(o) to be a feasible point. 2. Applying algorithm A to this point, we obtain an extreme point (f 1 ,(\); r := 1. 3. T = (tj, . . ., tk)T are active points for (£ r , ( r ). Find A, A. Here A is taken to be of full rank. 4. If Ajfi < 0 for some j , then g = A~1e2j-i (e2]-\ is the vector with 1 at 2j - 1 and zeros elsewhere), x = £r +ag where a is determined from If a = 0 = sup{ — ayJ, t £ [0,1], t / h, ..., tk}. Let z(t) = a(t)Tx-b(t). Apply algorithm A to (x,z). Obtain (f + 1 , (r+1); k = r + 1. Transition to 3 ; 5. If Ai(] ^ 0 is some j , then h = e,, A(t) = (a(ii)a'(t 1 ), . . . , a(tk)a'(tk))T, l T b(t) = (6(<1)6'( min is feasible subject to x(r + o/i). £ r + 1 ; = X(T + oft); r = r + 1, go to 3. Consideration is being given to the case, where the solution at the extreme point is degenerate, and the general case, where the index sets T are in RF. Even application of the algorithms (A,B) may encounter computational problems which still remain unsolved. Theoretically, however, this method may be of interest in infinite extension of the simplex algorithm. One of the methods of descent is proposed in [100]. The algorithm has been developed for a generalized linear optimization problem in Hilbert space. Let S be any set. W is a function space: S —¥ Rq and U, Z are vector spaces. V = U ®W Here V is taken to be Hilbert space with the scalar product (•, ■). It is necessary that the elements of W have the norm associated with the scalar product or the supremum norm ||W^||oo = sup{||u)(s)||0O, \s 6 S}. Define the convex cone W+ = {w 6 W\ Wi{s) > 0}; W° = {w £ W; Wl{s) > 0, i = IT?}; V+ = U ® W + ; V° = U ® VF°. Introduce the relation ">" into V: f > v <=>({ — v) G V+; 0 is zero in V. Let W° ^ 0, and select some e 6 W°. Any element of the form (u, e) £ V° is taken to be preferable in V General statement of linear problems: ( min((c,d);(u,w));T((u1w)) = b;(u,w)>0; 1 («,ui) G V;
, '
c £ U, d £ W, b £ Z. T: V -> Z is a continuous linear operator. At each step we define V? to be an invariant non-identity endomorphism in V which transfers the current point (x,z) £ V° to the preferable (x, e) 6 Vt that is "far removed" (in the sense of the supremum norm) from the convex cone boundary. In the transformed space, a step is then performed in the direction of steepest descent (in the sense of the minimized functional) and the inverse endomorphism is used to obtain a new current point in the initial space. Endomorphisms Fz, F^ are obtained explicitly for reciprocal linear mappings of V into itself, FZ(V°) C V^, F^(V°) C V°. The generalized algorithm may work well under the following assumptions: 1. V = U®W is Hilbert space with the scalar product (■, •), U is the vector space, W is the functional space S —^ Rq; T: V -¥ Z is the linear mapping;
8.5. METHODS AND ALGORITHMS
FOR SOLVING SIP PROBLEMS
159
2. 3£ G V: ( > 0; 3. V(u,w) G V, {(c,d),(u,w)) = {F2((e,d)),Fz((u,w))); 4. Vz g W the kernel of linear operator T o Fz is Hilbert space in the norm induced by the scalar product (o indicates composition). We have the problem equivalent to (8.21): (SIP)
mm(Fz((c,d)),(u,y)),T
o Fz((u,y))
= b,(u,y) >0,(u,y)
e V,
(8.22)
where y = z~l Q w, and 0 : W x W -> W: (w Q v)i(s) = wt(s)v,(s), \/s G S, i = \,q. For (LP) we have a feasible point (x,z). Iteration: transform V using the endomorphism Fz'. (x,z) —I (x,e). Step is accomplished from (x,e) to obtain in Fz(V) the least SLP functional point. Applying Fz to V yields a new current point in V°. In [101], the step direction is taken to the orthogonal projection of — Fz((c, d)) onto kerTo Fz. Algorithm. Step 0. k := 0. Step 1. Transform the feasible point (x,zy^> > 0 into (x,e) using Fzw, i.e. (x,e) = FzW((x,z)W). Step 2. Project orthogonally — Fz((c,d)) onto kerT o FzW. Accordingly, let Step Step Step Step 1. This
3. Find a: e + azp > 0, z' = e + azp. 4. Invert (x,z)( t + 1 ) = F*m((x + aCp,z')). 5. Check the stopping criterion. If so, then stop; if not so, then k := k + 1, algorithm is applied to LSIP problems (LSIP)
{mmcTx,A(s)x
> 6(s),Vse 5'};
T
(8.23)
n
A{s) = {a,j(s)}; 6(5) = (bi(s), ..., bq{.s)) ; c, x € R ] S = [lt,v,]. The problem can be rewritten as (8.21). Let us introduce free variables Zi(s) G C°°[l„Vi]. Obtain {min cTx, \A(s)x - z(s) = b(s),Vs 6 S},
(8.24)
where z(s) > 0, x G Rn, z € f[ C°°[/;,i>,-]}. Here U = Rn, W = Z = f[ C=°[U,vx\, !=1
T(x,z)
= A(-)x-z(:);
{(c,d)(x,z))
>=1
= cTx+
£ / di(s)zi(s)ds.
Select e, G (7°°ft,Oi]:
•=if,
e;(.s) = 1, Vs 6 [/,, Uj]. It can be readily seen that the direction of steepest descent is
-M). Step 2 requires projection onto kerT 0 F*m, and it is a finite subspace of V Perform the following steps: i. Find the basis {ut; 1 = ~L/n} oiToFzW, since for any (u,w) G k e r T o F | ( M , w = AuQ z~l =>■ Ui = [ki,u), i = V " , h = (0, . . . , 1, 0, 0, . . . , 0), n(s) = (a « . ( s ) M i ( a ) ) -
ii. Orthonormalize this finite basis using the Gram-Schmidt process.
160
CHAPTER
8. INFINITE
AND SEMI-INFINITE
PROGRAMMING
Step 3 requires a > 0: e + azp > 0. This is accomplished by selecting m i n o r s ) in [0,1]. Let t h e minimum be equal to -z< ml ' n >; then a - ' " i ? ^ . a ( m u , ) is a constant z
p
multiplier in (0,1). Assume z£ m "" < 0. In Step 5, use should be m a d e of t h e criterion of a sufficiently close distance between two adjaced solutions. Suppose t h a t in problem (8.8) c G Rn, c / 0; T is t h e compact and X is the convex compact set in Rm and Rn, respectively; u(-) and A(-) are continuous in T; there exists x such t h a t u(t)x > X(t) for any t € T, and it is not optimal for (8.8). One of t h e first applications of t h e cutting plane m e t h o d (based on t h e Kelley cutting plane method for solving convex programs) is provided by t h e alternative algorithm [86]: SO: Problem (Zo): min{cz:Vx € X C Rn} is examined with k = 1. S i : Let Xk G Rn be a solution to problem (Zk-\). S2: Seek rmn{u(t)xk — X(t)}, t G T, and let tk G T be an element in which a minimum is attained. If u(tk)Xk — X(T — k) > 0, then Xk is optimal lor (8.8) and the algorithm terminates. S3: Eliminate the constraint u(tk)x > \(tk) from problem (Zk) and go to step SI. Referring to [65], we have the limit points of t h e sequence {x^} t h a t are optimal for (1.8) if t h e method is not bounded. If X is a polyhedron, each problem (Zk) is linear and execution of steps SI and S2 is significantly facilitated. In the general case, however, this presents a. complicated mathematical programming problem. As suggested in [62], this alternative algorithm can be modified in such a way that a "sufficiently'' violated constraint is sought instead of t h e constraint most suspectible to violation. Accordingly, the steps 5 0 and 5 1 are replaced by t h e following steps: 5 0 " : For (Z0), specify from step SO a sequence {6k} of nonnegative numbers such that lim Sk = 0. k—>-oo
52*: Let S(xk) = mm(u(t)xk — X(t)) for anyt G T. IfS(xk) > 0, then Xk is optimal for problem (1.8) and the algorithm terminates; otherwise we seek tk G T satisfying u(t)xk
- X(t) < 0;
u(t)xk
- X(t) < S(xk) - Sk.
Referring to [62], we have t h e algorithm based on t h e cutting-plane method [40] for convex programming problems: 5 0 : Let B > v(Z). Select a constant f) G (0,1) and consider problem (SZ0): max
! Select y G Y and let k = 1. 5 1 : Let (xk,<Jk) G Rn x R be a solution terminates.
(8.25)
x G A'. to (SZk-i),
If Ck = 0, then the
algorithm
8.5. METHODS AND ALGORITHMS
FOR SOLVING SIP PROBLEMS
161
52: In accordance with two cutting rules, (Rl) and (R2), we may or may not cut conditions from (SZk-i). S3: (I) If xk is feasible for (Z), i.e. u(t)xk ^ ^(0> V^ G T, then add the condition ex + \\c\\a < cxk to (SZk-i). Let yk = xk. 53: (II) Otherwise we seek tk 6 T such that u(tk)xk - \(tk) < 0. Add the condition u(tk)x - \\u(tk)\\a > \(tk) to (SZk-i). Let yk = yk-iDenote the resulting problem by (SZk). Let k = k + 1 and return to 51. Cutting rule (Rl): Cut off the condition ex + \\c\\ff < B or any other condi tion obtained in Step 53(7) on the previous iteration if xk is feasible for (Z), i.e. min(u(t)xkX(t)) > 0. Cutting rule (R2): Cut the condition from (SZk-i) if: (a) this condition was generated in Step S3(II) on the j-th iteration, where j < k; (b)ak<pa3; (c) the condition was not an equality in (SZk-\) in (xk, ak), i.e. u(tj)xk—||u(ij)||cr/t > \(t3). The convergence of algorithm for a convex problem prevails. Theorem 8.22. [62] The failure of the algorithm to terminate generates a se quence of feasible points {yk}, and the limit points of this sequence are optimal for (Z). Referring to [84], we have problem (SZk-i) with a polyhedral set X of the form X = {x:a{X > bi] i £ l , m } , i.e. problem max a; ex + |[c||<7 < cyjt_i; u(t3)x - ||ti(ij)||A ( ^ _ j 6 Jk; d{X > £,-; i £ l , m ,
(8.26)
where Jk is the index set of constraints generated in Step 53(11), but not cut off by Rule (R2) prior to the fc-th iteration. Thus, we have to consider only the last constraint generated in Step 53(1). The dual of problem (8.26) is min(cyt_,Wo -
E Kt,)W3 - £ fcV5); '=1
j£jt rn
(SPk-i)
cW0-
£ u{tj)WjieJk
T,aiV,=0; '=i
(8.27)
||c||Wo+ £ H*f)||^- = 1; W3>Q,je{0}Ujk; Vi>0,ie l,m. Let (xk,ak) be optimal for (SZk-i) and let Wj = Wf, j € {0}(JJk; V, = Vf; i g l , m be optimal for (SPk-i). Then we have the following theorem. Theorem 8.23. [62] If the algorithm fails to end, then among feasible points the algorithm linearly converges in the objective function value.
162
CHAPTER
8.
INFINITE
AND SEMI-INFINITE
PROGRAMMING
In [65], the use of dual relations results in nonlinear equations for primal and dual variables. This system is determined by the set of necessary and sufficient optimality conditions for the primal (Z) (8.8) and the dual (F) (8.12) problem. It is also shown that more than n variables assume nonzero values. We have the s ystem of equations: (
n
m
J2 u{ti)tp{U) + J2 a,vt = c; (N)
\ v{t'~){u(ti)x - A(ii)7= 0 , ! G U v,(a,x — b,) = 0,J 6 l , m .
(8.28)
is system are taken (u(t)x — A(<)) has a local minimum tt if0. Variables in th to be
R: g{x,t) is continuous in X x T; g{-,t) is convex for any t and differentiable in X, and Vxg(x,t) is continuous in X x T; 4. 3x e int X: g[x,t) < 0; Vt 6 T (there exists x =£ x*, with ;* a being optimal); 5. f: Rn -> R; f{x) = \xTCx - pTx; C is positive definite. Let the set X be described by a finite number of constraints: X = {x e Rn: fj(x) < 0;
fj are convex ; j £ J = {1 , . . . , 9 } } .
For clarity, we consider only affine linear constraints, i.e. we have a quadratic problem (Q) (8.6). In order to implement the Kelley cutting-plane method, it is necessary to determine the most violated constraint at each iteration by solving th< s auxiliary problem (Hk)
max{g(x\t):teT},
8.5. METHODS AND ALGORITHMS
FOR SOLVING SIP PROBLEMS
163
which is nonlinear, generally nonconvex, nondifferentiable. The following discretiza tion is used, with {Hk) replaced by a sequence of approximating problems: max{g(xk,t):t€Tm},
(fl?)
where {T m } is the set sequence Tm —> T in the sense of Hausdorff metric. These assumptions guarantee the convergence of optimal values and solutions of the sequence {(H™)} to the optimal value and solution of (Hk). Algorithm A. [131]: 50: k = 1; T° ^ 0; z° = X and select Tm for m = 1,2,. . . such that Tm C T; Tm -* T; 51: Compute the solution of the subproblem min{/(;r):zeZ':-1};
(fit-,)
52: Compute the exact solution tm of the auxiliary problem (HI71); 53a: If g(xk,tm) < max{g(xk,t) : t € T}, then go to Step 52 with m = m + 1; otherwise tk = tm; 536: If g(xk,tk) > 0, then go to Step 54; otherwise the algorithm terminates and k x is the approximate solution of(Q); 54: Adjoin the intercept Sk(xk, tk, x) = g(xk, tk) + Vxg(xk,tk)T(x
- xk) < 0
and define Fh = {V g Tk~l: SJ(x3,V,xk) = 0}, where Tk~l is the index set of(Pk-i) k k k k and the sets T = F {J{f }; Z = f) {x: Sj(x>,V,x) < 0}f]X. Go to Step 51 veTk
with k = k + 1.
The auxiliary problem (Hk) in algorithm A is solved by the sequence of associated problems (H™). If in Step 53a the iteration terminates at m = m 0 , then the ap proximation th = tmo of the global maximum t of the problem (Hk) is achieved. Let dk be a deviation of the value of the approximately computed and most violated constraint g(x,tk') in (H]*°) at the point Xk(Hk)
g(xk,tk')
=
mzx{g(xk,t):teT};
g(xk,th')-g(xk,tk).
dk =
Theorem 8.24. [110, 131] Let lim 4 = 0 and suppose the sequence {xk} obtained by algorithm A converges to the point x* Then x' is a solution to problem (Q). Theorem 8.25. [110, 131] Let lim dk = 0. Then the sequence {f(xk)}
obtained
fc-+co
by algorithm A increases monotonically. Theorem 8.26. [110, 131] Under the assumptions of Theorem 8.24, the sequence {xk} converges to the unique solution of problem (Q). Let constants M > 0, N > 0 be such that: ||V/(s)]| < M; Vx £ X and ||V.j(a!,*)|| < N\ Vs G X, "it £ T.
164
CHAPTER 8. INFINITE AND SEMI-INFINITE
PROGRAMMING
T h e o r e m 8.27. [131] Under the assumptions of the Theorems 8.24 and 8.25, {xk} and {f(xk)} have the arithmetic convergence rate, i.e. there exist constants ko, ai and a2 such that: 11/* - f{xk)\\ < di/k,Vk where a, = l /
7
( < ^ p ^
> fc0 and ||a* - xk\\ < a2/Vk~,\/k > k0,
a2 =
lMflg^).
For application of the algorithm, it is of interest to consider the case where T = [a, b] C R or T = [a, b] x[c,d]c R2 In the discretization of the problem (/?*), T is replaced by 7\ m C T with the density max min \\i — t II < hm, t€T t'eThm "
i.e. we have the problem
(#£"*)
m&x{g(xk,t):t
£Thm}.
Modifying Steps S2 and 53 of algorithm A, we have Algorithm B . [110, 131]: Select hi; h > 0; e\ > 0; e G (0,1) and replace in algorithm A Steps 52 and 53 by the following ones: 52*; Compute thm which exactly solves problem (Hkm). S3*a: If g(xk,thm) < e 1; then go to Step 53*6, otherwise set tk = thm and go to 54. 53*6: If hm < h, then the algorithm terminates, and xk is an approximate solution of (1.6), otherwise return to Step 52* with hm = hme. We have the discretization network method for the nonlinear problem: min/(x);
X = {x £ Rn\c{x,t)
< 0,
t € T};
(8.30)
c: R71 x R —> R; T is a compact set, T C Rr, f is continuous, A' is bounded. This method does not require optimization over the entire T, while the solution of the original problem is replaced by a subproblem sequence: mmf(x),
XkCX,Xh
= {xeRn\c(x,t)<-ph,teT},
(8.31)
x£Xk
where Tk is a finite subset of T. It is assumed that: c(x,t) is Lipschitzian: |c(x,<) c(x,s)\ < L\\t — s\\; L > 0, Vf, s, c is convex in x, Vf. Slaiter's conditions are satisfied: 3x* e Rn: c{x*,t) < 0 , Vf e T. Special significance is attached to the function a c which yields a neigbourhood, where the constraint inequality is still retained: ac(y,t,f3):c{y,t)<-0,
when \\t-s\\
s£T.
(8.32)
For the generalized algorithm to be operative, we need to specify ac and the way of specifying an arbitrary network in T For two cases the ac is obtained explicitly:
8.5. METHODS AND ALGORITHMS
FOR SOLVING SIP PROBLEMS
165
1. Let c(y,s) be differentiate in 5 on T, and suppose g(y,s) is a gradient c and there exists G(y,t) > 0: \\g(y,t) - g(y, s)\\ < G(y,t)\\t - s\\, s g T, then ac{y,t,0) = l(\\9(y,t)\\2 + 4G(y,t)\c(y,t) + t3\y/i-\\g(y,t)\\}/2G(y,t). 2. Let y g Rn,teT,P> 0, c(y, t) < - / ? , select h > 0. Suppose there exists Q(s) which is a quadratic function with the following properties: | Q[a)>c(v,*)+Pfat\\a-t\\ < h; \ Q(s) = c(y,t) +/3 otherwise. Let Shea, distance from t to 5 = {s g T\Q(S) - 0}, 6 = +oo, if S = 0, then define <*c{y,t,P) = min{h,S}. An algorithm is proposed for generating an arbitrary network for prime sets T. Let T = I\ x • • ■ x Ik, I, = [p,; g,], /i, = (q{ - Pi)/rn.i; the network M3 = {px + fc1/i1/2J,...,pr + fcr/v/2J},fc€ {0,...,2-'m,}, j = ]~F. Algorithm. Select Po > 0: 0 < 0O < mm{-c(x',t), t € T), x' g X. Given L > 0 — Lipschitz constant, function ctc, the rule for generating "/-network in T. Step 0. 70 = Po/L. M0 — 7o-networJc in T. TQ = MQ, k = 0. Step 1. Solve the problem {min/(x); c(x,t) < —Pk, t S Tk} which is a. finite convex programming problem. Let y be its solution; j = 0, EQ = To. Step 2. If j = k, go to 4, otherwise E'j+1 is a point set: {t £ E3\ac(y,t,Pk) < 7y},' £ J + i is a point set in 7 J+ i networic M}+i, such that the distance to E'J+l does not exceed 3~fj/2; j is replaced by j + 1. Step 3. If c{y,t) < —Pk for t g E}, go to Step 2, otherwise c = {t g £j|c(y,<) >
-/?i},ri = rj,uc,gDtostepi
Step 4. xk = y, Tk+l = Tt, Pk+i = Pk/%, lk+\ = 7*/2, fe = k+ 1, go to Step 1. This algorithm generates the sequences {Pk}, {jk}, {Mk}, {xk}. A theorem is given to prove that with Vfc, xk g Xk C A' is a feasible solution to problem (8.30), limit = z — optimal solution (8.30). If / is convex, then there exists k > 0:
f(xk) - f(z) < k1~k The following discretization algorithm was proposed in [18] and developed for quadratic SIP problems. Consideration is being given to QSIP problem (QSIP)
{min F(x) = cTx + -xTCx,
xTa{t) < b(t)},
C is real, positive definite, t g T, T is a compact set, a, b are continuous. The original problem is replaced by the sequence of problems QSIP{Gt)
{min F(x), xTa(t) < b(t),t g G,-},
where Go C G\ C • ■ ■ C T, Gk are finite sets. The following method is designed for constructing Gk: the network T[h] is an intersection of the set T and the set p
{£0_|_ ^2 jih,ei\ji g Z}, hi is a step, t0 is selected in an arbitrary way, e< is the i-th basic i=i
166
CHAPTER 8. INFINITE AND SEMI-INFINITE
PROGRAMMING
vector Rn We may specify {ht} -* 0: T[h°] C ■ ■ ■ C T[hn] ■ ■ ■ C T. Let T[fefc] = Gk. The number of points in the network Gt rapidly increases with increasing i. This generates a need for subtler methods of constructing the sets involved in solution of iteration problems. Referring to [19], we have a new approach. It is essentially aimed at finding a solution to QSIP(G,) not on the entire network T[/t,], but on its specially chosen subset T[h%\ of points, about which the problem constraints are violated, and in terms of the solution of QSIP(G;_i) which has been treated at the preceding stage. Define R(xk) = {t: t G T[hk}\ xja(t) = &(/)}; Xk is a solution to QSlP{T[hk}). Algorithm for selecting f[hk] is proposed in [22]: Step 0. Find a solution x0 to QSIP{[h°]), R = R(x0), f[h°) = T[h0}. Step k. Given: R, f[hk-1}- R C f[hk-1} C T[/ifc->]. xk is a solution to QSIP
(f[h^}).
a) if'Xk-i satisfies the constraint of QSIP (T[hk 1 ]), go to c), otherwise go to b); b) select a new T[hk~1] from the points of R and those points in T[hk~l] ,where the constraints are violated; solve the new problem; the new solution is Xk-i; if R(xk-i)add to R an arbitrary point from f[hk-1} \ R, go to Step k; c) select f[hk}: R C f[hk], find a solution to QSIP(f[hk]); R := R{xk). Transition to Step k-f-1. Referring to [153], for large-scale block problems we have the algorithm for the sequential truncation method. The analog given for the Kornai-Liptack method is based on dual estimations of the linear programming problems specifying the con straint matrices multiply. Referring to [80, 144, 145], for the solution of a nonlinear problem of the form (8.29) we have globally convergent methods (a generalization of the Wilson method solving the proposed problem by a sequence of quadratic programming problems). Let X0 C Rn be open, Y C Rm compact , and F: X0 -»• R; f: X0 x Y -> R; g1: m R —> R; j £ J; \J\ < oo three times continuously differentiable functions. We now solve the following problem: mmF(x); x G X = {x 6 A'0 : f(x, y)<0,ye
V};
(8.33)
! Y = {y.g>(y)<0,]£ J}. To ensure convergence, the sufficient condition must hold for local quadratic con vergence to the optimal solution So, including the assumption of strict local unicity of x0. We assume that x 0 is a strictly local unique solution to problem (8.29) if for any l e A ' f l U(x0), (U(x0) is some neighborhood) F(x) - F(x0) > Q||.T - x 0 ||; a > 0; ||x|| is some vector norm. Referring to [144], we have the algorithm extending the Wilson method to SIP problem: Algorithm WSI. 50: Given zk = {xk, u'k, yik); i G T~f; let k = 0;
8.5. METHODS AND ALGORITHMS
167
FOR SOLVING SIP PROBLEMS
Sk: For k > 0, we calculate the optimal solution zk+1 for the following approxi mation problem: minF(x,zk) = F(xk) + (x xk)Fx(xk)-r k T k + ^(x-x ) Lxx(z )(x~xk); k k { f(x,y,z ) = f(x ,y) + (x- xk)Tfx(xk,y) <0,y (
(POzk)
where L is the Lagrangian L(z) = F(x) + J2 u'f(x,y'),
Je
(8.34) eY,
- problem {PO) is approx-
;=i
imated by the SIP sequence (POzk) whose structure is simpler than that of (PO). T h e o r e m 8.28. [144] Let some second-order sufficient optimality condition hold for the optimal solution x0 of the problem (PO). Then: 1) for any k, the Zj, satisfies this condition for (POzk); 2) the sequence (z<.) is quadratically convergent. To solve SIP problem (POzk), we need to consider the finite optimization problem:
(FPOk
mmF(x,zk) = F(xk) + (x-xk)Fx(xk)+ k T k + \(x - x ) Lxx(z )(x - xk); x £ A'(-Y is a neighborhood of xk+l); / ( * , y'(x), zk) = f(xk,y'(x)) + (x- xk)Tfx(xk, y'(x)) < 0; k I f(x, y', z) = f(x , y') + (x- xk)Tfx(xk,y>) < 0, y> e Y,
(8.35)
where Y is a subset of Y, and L is the previously introduced Lagrangian. T h e o r e m 8.29. [144] xk+l is strictly optimal for (POzk) if and only if xk+1 is strict-locally optimal for (FPOk). The optimal solution xh+1 for the problems (POzk) is computed as follows [80] (Step Sk in algorithm WS1 is replaced by Steps Sk0-Sk3). 510: Start with xs = xk, u" = ulk, i € l , r : s = 0 (for k = 0, x', u" can be chosen arbitrarily). Ski: Maximize the function f(xs,y,zk): y £ Y, specifying y", i G l,r. Sk2: Test for optimality. Go to Step Sk + 1 with zk+1 = z" or set s = s + 1. Sk3: To compute x3, u", solve the finite optimization problem:
(FPO
s-l^ k ,
( min F(x,zs-\zk) = F(xs-1,zk) + (x-xs-1)Fx(xs-\zk) + \(x - x^)TLxx(z3-\zk)(x x^); f(x,y<'°-\zk) = f(xk,yi^1)+_ + (x-xk)Tfxixk,y^-1)<0,iel,r; k k { f(x, y>,z ) = f(x ,y<) + (x- xk)Tfx(xk,y>) <0,y'e
where L is the Lagrangian L(z,zk)
= F(x,zk)
+ £ u'f(x,y'(x),zk); i=i
that Lxx is of the form Lxxiz, zk) = £ ufxyixk,
y')ylix)
+
Lxxizk)
+ (8.36) Y, it can be shown
168
CHAPTER 8. INFINITE AND SEMI-INFINITE
PROGRAMMING
For each i, an (m x n) matrix y'x is computed from the solution of linear equations. Return to Step Ski. is solved in such a way that 'k Fx(x,z°-\zk)
+ YJutfx(x,y''s-\zk)
= Q,
■=i
then for zk, zs is sufficiently close to z0, Step Sk3 is equivalent to an iteration of the Newton method for the system of nonlinear equations: F(x,zk)+£u'fx(x,y'-°-\zk) k
f(x,y'(x),z )
= 0,
= Q; z € V .
Thus the sequence {z3} converges at quadratic rate. For separably-infinite problems (A) (8.3) and (B) (8.13), in addition to Assump tions 1-3 (section 8.2) we make the following assumption [31]: Assumption 4. Denote by «,(•) thej-th component ofu(-) and byvi(-) the i-th component ofv(-). Let Uj(-), j € l,n — 1 be continuously differentiate on the open set containing S, and let «»(•), i £ 1, m — 1 be continuously differentiate on the open set containing Q. Theorem 8.30. [31] Let x* 6 Rn; >?*(•) € R[Q) be feasible for problem (A), and suppose that {rt,... ,rq} is contained in »]*(■) and rj* = n*(rt). Also, let y* £ Rm; A*() £ i?' s ) be feasible for problem (B), and let {t 1 ; . . ., tp} be contained in A*(-) and A* = A*(ts). Ifri, i 6 1,9 are interior to Q, and t{, i G l,p are interior to S, then £ Vu^tAx*
- Vu„ +1 (t.) = 0,i e l,p;
771
E Vi)j(r,-)y* - Vvm+1(n)
(8.38)
= 0,i £ l,q.
Given p and q, the unknowns in system (8.38) are x" £ R*, y* £ Rm; t, £ /?*; A* £ R; i £ l,p; r, £ /?' and ry* £ /?; i £ l,g. System (8.38) is composed of p + g + m + n-l-fcp-r'g/ nonlinear equations with the same number of unknowns. Referring to [31], we have one of the most successful methods for solving separablyinfinite programming problems (A) and (B), which is as follows: (1) region 5* in problem (A) is replaced by a subset containing finite points; and definition is given to the solution of the finite linear problem and its dual; (2) using the results of Step (1), the initial point is obtained for iterations of the Newton-Rafson method in the form of the nonlinear equations (8.38) derived from the dual relations between problems (A) and (B). Most of the methods proposed in [46, 54, 72] for solving SIP problems are locally convergent. The global algorithms which guarantee convergence to a stationary point in the problem are given in [33, 52, 144, 145], Referring to [74], we have a comparison of three numerical search methods for solution of SIP problems: discretization, exchange algorithms (e.g., simplex method
8.5. METHODS AND ALGORITHMS
FOR SOLVING SIP PROBLEMS
169
and its modifications), and continuous methods (such as the Newton method) which do not use discretization in the ordinary sense. Theoretical results and examples show that the first of the methods is generally useful only in obtaining some approximations to be further improved by the third type method. The idea of the first discretization method is that the infinite set should be replaced by a finite subset sequence, where upon the primal SIP problem should be replaced by a finite subproblem sequence. Computational time substantially increases as the discretized problem increases in dimension. Modifications of the simplex method permit a reduction in the amount of com putations by an order of magnitude, but these methods retain the convergence rate that is linear or a little better. The methods based on the Newton method [64, 74, 80, 145] have superlinear quadratic convergence. The penalty function methods which ensure adequate convergence of iteration procedures to a stationary point are important in the construction of global algorithms for nonlinear programming. Referring to [32], we have an exact penalty function for semi-infinite programming which is a generalization of the ^-exact penalty function in nonlinear programming. Such function is constructed for a convex problem, and then necessary existence condi tions are provided for a local minimum of the penalty function (Theorem 8.9) subject to the existence assumptions of the global minimum of this function (Theorem 8.10). In the general case, additional assumptions are made and a proposition similar to Theorem 8.10 is formulated. Proofs for the stated theorems are constructive, and the quadratically convergent algorithm is based on these ideas. The results compare favourably with [33]. Referring to [33], we have a projected Lagrangian algorithm for semi-infinite pro gramming (SIP). Efforts are made to develop the locally convergent methods on the basis of an exact penalty function. The method of the last type is discussed in [33]. The drawbacks of this method are: each of its iterations calls for solution of a quadrat ic programming problem with inequalities, while the high rate of convergence calls for conditions that are not necessarily satisfied. The modification of this method proposed in [33] permits obviation of such drawbacks. We have the algorithm for SIP problems using the quadratic sequential program ming methods together with an exact penalty function of class L^. Its global con vergence is proved. The proof does not require that the implicit function theorem be applied to SIP constraints; it suffices to impose a weaker condition on the finiteness of points transferring a global maximum to any SIP constraint. We have the minimization method for locally Lipschitzian functions using approx imate values of the functions. This method yields the SIP minimization algorithm for exact penalty functions which is similar to the method of confidence regions employed in nondifferentiable optimization. A finite method is constructed for SIP approxima tion of an exact penalty function. In [91] proposes alternative penalty methods. As suggested in [913], the proce-
170
CHAPTER 8. INFINITE AND SEMI-INFINITE
PROGRAMMING
dures of discretization, regularization and those of the penalty method are harmonized in such a way that for each discretized problem there is a fixed unconditional mini mization problem which is to be solved before the results of two adjacent iterations are brought into a certain proximity. Provision is made for the adaptive method of constructing the next discretized problem using the approximate solution of the two preceding problems. Consideration is given to the problem mm{/(u) : u G K0}; K0 = {u g Rn : g(u,t) < 0,Vi <= M};
(8.39)
f,gt are convex and differentiable in Rn M is a compact subset in a normed space Y Denote u* = arg min /(«); / m ; n = min f(u); Sr = {u g Rn: \\u\\ < r } , || • |j is the Euclidean norm. We assume that u* ^ 0 and the constraints satisfy Slaiter's condition for arbitrary r > 0, u £ Sr, t,t' €. M with some L = L(r) < oo: \g(u,t)-g(u,t')\
(8.40)
Let u be Slaiter's fixed point with Mi as a finite /i;-network on M. K, = {u g Rn;g(u,t) c
< 0;V< G M t }:
i M = min{/(u) + -||u|| 2 },a = - m a x g ( « , f ) ; C (r)
= a- 1 [/(S) + ( P | | + r ) 2 - C l ( r ) ] .
Suppose that {/*,}, {£;}, {£ t }, {T,} are the given sequences of positive numbers ap proaching zero. Let us fix r so that u'r]Sr/8^0;
u x ' 0 e A'ifl 5 r/4i
K M ^ E b K ' l + ^ t J + nft !£M,
^,5(u) = /(u) + v;(u) + ||u-u^- 1 || 2 , where c(r). To find a solution to this problem, we construct a sequence {ti 1,s }; s = 0,. .. ,s(i); i — 1,2,... ,s(i) are determined during iteration. Algorithm, i — 1, s = 0, s(i) > 0. Step 1. If s = s(i), then u'+h° = u*'sW; i := j + 1; s := 1; if, however, s < s(i), then s := s + 1. Step 2. Determine the point u ,,s from the condition ||VV'i,su''s|| < e\. Step 3. For ||u;'3 - u'^-l\\ > e, we set s{i) = s + 1; go to Step 1. Step 4. If ||u ; ' s - u^-'H < e, then s(t) = s; go to Step 1. Denote q(r) =2rLa~',£, = ^- + (j(i|Mi|)5r, 4 , where |M,| is the number of elements in the set Mt. is(r) = sup ||V/(u)||. u6S r
8.5. METHODS AND ALGORITHMS
FOR SOLVING SIP PROBLEMS
171
T h e o r e m 8.31. Let hi+1 < ht, Vz, h, < aL'1; let condition (1.40) hold, and -[2u{r)q{r)h, - (e,- - e,)2} + e, < 0 for any i; oo
EiiMr)q(r)h^
+ E, + 2q(r)ht} < r/2.
!=1
For the algorithm we then have s(i) < oo with any i, \\u^'\\ < r, Vi, s and the sequence {u*,s} converges to some u* 6 U". Moreover, /(u*'*W) - inf f(u) < 4re, + (4r + v(r))e,; p(u'^\
K2) < £,;
p(v,Q)=
inf | | t ; - H I -
An estimate is made of the method convergence rate. Denoting IM,O = /(«''°) /mm + t/(r)(ff('")^« + £,-)) a n < i adjusting the zero tendency of the sequences {hi}, {e^}, {e,-}, {r,} in a special way, we guarantee the estimates: ||«" s -«*||<3[ / i ! ,o/6] 1 / V / 3 ' + 3 ) / 2 , t-I
6 is a constant > 0, q is a constant: 0 < q < 1; /?; = 2 s(£). In the case s(«) = 1 for Jt=i
any i we get:
||u" s - U *||<3[ W ,o/&] 1/ V i - 1)/2 Referring to [26], we have the iterative algorithms for nonlinear semi-infinite prob lems with recurrent constraints (systems of associated optimization problems with or without recurrent constraints). For the convex problem (8.5), [27], however, there is an iterative algorithm of the gradient form using the penalty function method. m
ut+1 = ut-
S, fin1)
+ Y, ^fl,«(u')GJ,{u'); J=l
t
H]t(u ) J
=
a]t-b]lG]t(ut)\\G'}t(ut)\\-2;
{ 0, if G}t(uf) < 0,j e I7m,i e I~5o~; ~ \ 1, if G J( (u') > 0 , j 6 l,m,f € l,oo;
Proof is also provided for the theorem which determines convergence conditions for iterative algorithms of this form. In [68], semi-infinite problems are investigated using penalty functions and Lagrangian methods. A review is provided for relationship between SIP and optimiza tion problems with a finite number of variables and constraints, and two classes of convex SIP problems are introduced. One of these classes is based on the fact that the convex set can be represented as closed subspaces; the other class is defined by rep resenting the elements of the convex set as a combination of "points and directions"
172
CHAPTER 8. INFINITE AND SEMI-INFINITE
PROGRAMMING
A generalization is given to the nonconvex case. To solve SIP problems, necessary optimality conditions are generally taken to be a system of nonlinear equations. In the proposed method, such a system is constructed from a discretized version of the SIP problem under discussion, i.e. the problem with a finite number of variables and constraints. It is standard practice to solve this system by the linearization method. Finiteness is also the subject matter of [67]. SIP problems can be investigated by employing a variety of reductions to the finite case such as finite systems of infinite subsystems or finite probability measures. In [60], one of the general abstract methods for solving extremal problems (mar quee method) is generalized to the semi-infinite case.
8.6
Statement of the Infinite Programming Problem
The first statements of linear infinite programming problems probably refer to [100]. In accordance with [6, 7, 8, 10, 11], we define Eoo = {x = (xi,... ,Xk, ■ ■.)} to be a numerical sequence space with componentwise addition and multiplication by scalar. For x = (xj,. . ., zj.,. . .,) £ £«, we define the set J(x) = {k: x ^ 0}. Let E'^ = {x £ Ex: J(x) < oo} be a subspace of finitely defined elements. For a = ( o i , . . . ,ak, ■ ■ ■) G £«>, on the subspace E'^ we define the functional (a, x) = J2 a<x' i'SJ(i)
which is obviously linear in E'^. Let a0 = (a0\,. . .) £ E^,, b = (blt. . .) 6 Em. The infinite matrix is On . . . a\j .. . an
. . . a,ij
Denote a,. = (o,-j,...) € E^, ( a;, is the i-th row of the matrix A); a.3 = ( a ^ , . . .) 6 Ex (a.j is the j - t h column of the matrix A). The linear infinite programming problem then becomes: L oooo :inf{(a 0 ,x) : (a,.,x) < bhi - 1,2,..., x £ E'x} =
Vx>00.
(8.41)
Let 3 be the vector space of all vectors a;;, i 6 J , where J is an arbitrary index set of any power which is determined over the ordered basis Y with many, but only finite nonzero values, and let T = Yn be a finite-dimensional vector space over Y, while A0, {A,: i G ,/} is a collection of vectors therein. State the problem [100, 28] mint/'Ao; j/'A; > a,, i £ J,y where the vector af i g J, a, G Y.
eYn
(8.42)
8.6. STATEMENT
OF THE INFINITE PROGRAMMING
PROBLEM
173
Referring to [1], we have the linear problem min(a:, c); subject to Ax = b;
( 8 - 43 )
x e P,
where b £ Z, c £ Y; Z, Y, X are topological vector spaces. Problem (8.43) can now be represented as min(x, c); Ax-beQ; (8.44)
|
x £ P, where Q, P are positive definite cones in spaces Z and X, respectively. Referring to [108], we have an infinite-dimensional analog of the programming problem f{x) -» min; subject to (8.45)
I
Ax > j/o, where X, Y are locally convex spaces, A: X —> Y is a continuous linear operator , the functional is linear and continuous in X, and inequalities in Y are interpreted in terms of a partial ordering which is generated by a closed convex cone Ky. Consideration is being given to problem (8.45) with the additional condition x > 0 which is interpreted in terms of the partial ordering generated by a closed convex cone K~x in X [63], viz.:
I
maxci; x > 0 in X;
(8.46)
Ax < b in Z, where c £ X* is the X-adjacent space, b £ Z and A is the linear transformation from X into Z: A: X -> Z. Referring to [43], we have a linear problem with a finite number of inequality constraints and an infinite number of variables
I
c(x) —¥ max; A(x) =_d1_
(8.47)
Pj(x) >bhj £l,n,x£ V. Here V, U are vector spaces over the field R; A: V -> U is a linear operator from V into U; 0j, j € l,n, C are linear forms on V with values in R; &_,, j £ l,n are real numbers, d £ U. According to [27], (X, X') and (Y, Y') are two dual linear space pairs, A is a liner, weakly continuous operator mapping X into Y; A* is its associated operator; P(Q) is a weakly closed nonnegative cone in X(Y); P+(Q+) is the P{Q)-adjacent cone of
174
CHAPTER 8. INFINITE AND SEMI-INFINITE
PROGRAMMING
nonnegative elements in X'(Y'); b g Y, c 6 X' The partial ordering is induced by the respective cones in the spaces X, Y, X', Y' The problem is stated as follows: (TT) :
> b,x > 0} = F
mf{cx:Ax
(8.48)
or [122, 124, 125] x —> inf;
Ax > b.
(8.49)
The convex infinite programming problem is investigated in [100] minp(y); subject to G{y) > 0,
.
l
'
where G(y) = (gi{y),... ,gr{y),...) is the vector of concave functions given in the region which is defined by a convex set S of conditions on y, and ip(y) is convex. A more general problem is studied in [63]
I
max/(a:); x > 0 in X;
(8.51)
g(x) > 0 in Z, where X is also the set of all n-dimensional functions defined on T = { 1 , 2 , . . . } , while Z is the set of all Z: g(x) = b — z(x). Referring to [61], we have the general programming problem in function spaces [
(
/(a:)-+sup; $ ( x ) > 0 (G,);
(8.52)
xeG.
If G is convex, the functional f(x) is convex up on G, the operator $(x) is convex up on G (with respect to a positive cone Gi), then we have a convex programming problem in function spaces. Referring to [61] and [37], we have a functional analog of the LP problems. Suppose f(x) = ex is a linear functional, $(z) = b — A{x) is a linear operator of (6 6 E\), and G is a convex cone. Then we have
I
ex —> sup; Ax
(8.53)
z > 0 (G), i.e. the problem which is similar to (8.46). Referring to [141], we have the problem in Banach spaces (_ g(x) < 0,
x € G,
v
where / : X —> R is Frechet differentiable, the operator g: X —¥ Z is continuously Frechet differentiable; C S X, C is convex and closed.
8.7. DUALITY
OF INFINITE PROGRAMMING
175
PROBLEMS
Referring to [138], we have a quasiconcave problem with affine constraints
I
F(x) -> sup; G(x) = 0;
(8.55)
x € D, where F(x) = F2(F1(x)); F: D -4 W and G: X -4 Y are affine operators; F 2 : B —>■ y is a quasiconcave operator, i.e. the set {at 6 S: F 2 (x) > b} is convex V6 G F, and D C X, Fl(D) C B C VK; X, y , Z, and W are linear topological spaces.
8.7
Duality of Infinite Programming Problems
Let us define a series of problems that are the duals of the infinite programming problems stated in section 8.6. The dual of (8.41) becomes [6, 7, 8, 10]: (£««,)
sup{(-6,y) : (a.^y) = -a0j,j
= 1,2,. . . , y > 0, y £ &„} = v*^,
(8.56)
the dual of (8.42) [100, 28] is max Yl aixi] E K'x, = Ao; x € H,
(8.57)
x > 0.
In [1], the dual of (8.43) is max(6, w); -A"w + ce P*]
(8.58)
! w e W or that of (8.44) is max(fc, w); -A*w + ce P*;
(8.59)
|
weQ' Here the P- and Q-conjugate cones are P' = {yeY:{x,y)>0yx£ P};Q' = {w € W: (z,w) >0,VzG<3} and A* is the /1-conjugate operator. Referring to [127], we have the dual of problem (8.48): (O :
sup{y*fe : A*y = c, y* > 0} = G*
(8.60)
or the dual of (8.49): [122, 124, 125]
I
y'b —>• sup; A*y' = c;
y*>o,
(8.61)
176
CHAPTER 8. INFINITE AND SEMI-INFINITE
PROGRAMMING
the dual of (8.52): [61] \f(A) -» sup f o r A > 0 (G*),
(8 62)
-
where f (A) = sup F(x, A) = sup[/(as) + A$(r)] as well as the dual of (8.53): [61, 37] xeG xeG ( Xb -> inf; | c
I
(8.63)
A > 0 (G?)
and the dual of (8.46): [63] [ min(6,z*); I z*>0,z*EZ*; ( A*z* >c,c£ X".
(8.64)
The early investigations in finite linear programming made it apparent that the primal and the dual problem must be treated jointly. The duality theory, generalizing the classical Lagrange method, allows for extremality criteria and provides the basis for many computational algorithms. In the infinite case, the duality theory provides a principal means of studying extremal problems, but the situation here seems to be significantly complicated. In the finite case, dual problems (subject to fmiteness of extreme values) are always in duality. For infinite-dimensional spaces, however, this is generally not the case [100]. For this reason, it is standard practice either to consider a weaker relationship between problems (a week duality) [61, 37] or to prove duality under additional conditions on spaces, cones, etc. (solid properties of a cone, compactness of a feasible solution set, Slaiter's conditions, etc.) which are guaranteed by the validity of the required separability theorem [100, 127, 128, 105]. For linear infinite programming problems, the basic result is stated as a duality theorem [100, 28]. Theorem 8.32. If the system of inequalities yA, > at, i £ J has the MinkovskiFarkas property and is consistent, then we have one of the following results: 1. problem (8.42) is inconsistent, sup^Za.x, = oo or sup J2aixi 's finite; %
i
2. problem (8.57) is inconsistent, inf y'Ao = —oo; 3. problems (8.42) and (8.57) are inconsistent; 4. both problems are consistent and infy'Ao = sup J2 a;£; = JZ cnx* holds for a set of feasible solutions. In [1], the fundamental results of duality theory in linear infinite programming are treated systematically, and much thought is given to the strong duality theorem which ensures equality of the functional optimal values in the primal and dual problems. For the equality-constrained problem (8.43), the strong duality condition is stated as a closure condition for some sets constructed in terms of the problem parameters. For the inequality-constrained problem (8.44), the strong duality is guaranteed if there exists an inner point (Slaiter type regularity).
8.7. DUALITY
OF INFINITE PROGRAMMING
PROBLEMS
177
Referring to [61], we have a generalized duality theorem for convex programming problems in function spaces under assumptions of the form (8.52), (8.62). Theorem 8.33. If the objective functional f(x) and the constraint operator $(x) determined on the convex set G are convex upwards and the closure of the feasible solution set is nonempty, then the upper bound value is in the dual problem (8.62). As a special case, a generalized duality theorem is obtained for linear problems in function spaces (8.53),(8.63), as in [28]. Referring to [61], we have classes of convex and convex-fractional programming problems for which the generalized duality theorem changes to the conventional dual ity theorem of the extreme-value equality in the primal and the dual problem. In par ticular, some theorems establish the duality relationships without any requirements similar to Slaiter's familiar condition. Theorem 8.34. Let f(x) be a functional which is convex upwards and continuous in the convex set G, and let $(x) be an operator which is convex upwards (in a convex cone G\) and continuous in G. If lim fiJf,G)
= -co,
p—»oo
where h,[f,G)=
( { {
sup f(x), if GC\C,±% C\c" - c o , if Gf[Cp = Q\
xeG
(8-65)
C„ = {x : |as| = / ) , ! £ E}, G and Gi are closed sets, each bounded subset of G is weakly compact (which is guaranteed by the reSexivity of E D G), then the dual problems are related by duality relations. The functional0} is weakly closed for the functional
178
CHAPTER
8.
INFINITE
AND SEMI-INFINITE
PROGRAMMING
In [37], duality theorems for linear programming in abstract spaces (infinite pro grams) are stated subject to conditions which are similar to Slaiter's condition. In [127, 128], such theorems are formulated on the assumption of solid properties of the cone U = {(Ax - y,cx - v) x > 0,1/ > 0 } . T h e o r e m 8 . 3 7 . Let n and v' be dual LP problems (8.48) and (8.60), with 7r solvable. Moreover, let 1. cone U be solid in Mackey topology; 2. b E int(AP - Q) in Mackey topology. Suppose then that TT and IT' are in duality and problem n* has a solution. T h e assumptions of the theorem can be readily verified if cone Q is solid, specifi cally if Y is t h e function space with a homogeneous norm. A similar theorem can be proved if the closure of V is taken instead of solid property. Bourbaki-Alaoglu theorem relates the solid property of a convex set to the weak (local) compactness of its polar. T h e duality theorem [128] is also provided. T h e o r e m 8 . 3 8 . Suppose that U, = {A*y*+x", y*b*-v*) : y* > 0, x* > 0, v* > 0}, its conjugate cone is 11+ = {(x,v)
:Ax + vb>Q,x>Q,v>
0},
t i e pair of dual problems ix and IT' (8.48), (8.60) are in duality, and TX has a solution if: 1. U* is T-solid, and subject to one of the following conditions: 2. TX is feasible, c g int(A*Q+ + P + ) ; 3. TX is feasible, (c,v0) g intU* for some v0; 4.a) ix is feasible; 4.b) x > 0, Ax > 0, ex > F implies x = 0; 5. the e-optimal set Dc is nonempty and weakly bounded for some e. Under assumption (1), the conditions (2)-(5) are equivalent to each other. Solv ability of 7r* (8.60) can be assumed instead of admissibility of TT (8.48). We have a uniform continuous-time programming problem. Using the convex infinite analysis constructs and the corresponding Kuhn-Tucker variational theorem, a dual problem is constructed and t h e strong and weak duality relations are derived. Here the uniformity assumption allows a special convenient representation of the dual problem for analysis. Linear and quadratic programming problems are treated as special cases. In [41], duality is studied for convex programming problems with operator con straints. In [137, 139], an abstract duality theory is developed for general vector optimization problems. This theory is applied to quasiconcave programming prob lems with linear convex constraints in the ordered topological spaces [138]. Duality relations are obtained for marginal functional values in t h e primal and dual problems. Strong and weak duality relations are provided for problems in which constraints are specified by convex operators.
8.8. OPTIMALITY
CONDITIONS
FOR IP
PROBLEMS
179
For nondifferentiable programming problems in [86], a duality theory is construct ed from application of the alternative theorem. Similar issues receive sufficient at tention in vector problems with a convex structure [143]. Issues of duality theory in vector optimization are treated in [34, 85] for nonconvex vector optimization problems [50, 51].
8.8
Optimality Conditions for Infinite Programming Problems
In [43], for problem (8.47) there are necessary and sufficient optimality conditions (the previously obtained results can also be used for problems with infinite inequality constraints). Theorem 8.39. Let x0 g W. In order that the functional C attain a maximum on W at the point x0 ■& the form C must admit of representation such as
c=
(8-67)
Y, W& + «, jeJ(x0)
where W = {x : Ax = d, 0j(x) >bj,j£l,...,n,x£ V} is a constraint polyhedron; J(x0) = {j g l , . . . , n : Pj(z0) = bj}, fij > 0, j g J(x0), a g (ker A)T (if x g Y, then x g {a g V* : a(v) = 0 Va; g X}; V is the set of linear forms on V). Similar necessary and sufficient optimality conditions are provided for convex and, specifically, quadratic problems [42-44], A special feature of the optimality criterion (8.67) is that there are no topological constraints or regularity conditions which are common for the infinite case. The fulfilment of condition (8.67) means that in problem (8.47) the Lagrangian has a saddle point. In [63], the results of the Kuhn-Tucker theorems are extended to cover the infinite case. Under a certain constraint (the weak closure of a special cone) on the form of the constraint operator, the conjugate cone and the cone of feasible directions, the necessary and sufficient conditions for extremum are provided as the existence of a nonnegative saddle point for the Lagrange function. In connection with prob lem (8.51) the Lagrange function is defined as ip(x, z') = f(x) + (z', g(x)), x eX,z*
g Z'.
Then the nonnegative saddle point (x,z*) of the function ip(x,z*) is defined by the following relations: x > 0 in X, z* > 0 in Z*] tp(x,z*) < tp(x,z*) <0 from X and z* > 0 from Z* The Kuhn-Tucker theorem is proved for convex programming in linear spaces, but this requires no differentiability. Theorem 8.40. Let X and Z be linear spaces, and suppose that Px and Pz are closed convex cones in X and Z, respectively; f is the convex functional specified in Px; g is the convex function in Px with values in Z. And let 3 a point x° such that
180
CHAPTER 8. INFINITE AND SEMI-INFINITE
PROGRAMMING
x° > 0 in X, with g(x°) lying in Pz- Then the point x 6 X is an optimal solution of problem (1.66) <=> if there is another point z* 6 Z* such that (x,z*) is a nonnegative saddle point of the Lagrange function ip(x,z*). In addition to Theorem 8.40 the following theorem may be formulated for prob lem (8.46). Theorem 8.41. Let X and Z be local convex linear spaces and let some closure conditions be fulfilled. Then x is an optimal solution of problem (8.46) <=> 3 z* 6 Z* such that (x,z') is a nonnegative saddle point of the Lagrange function ip(x,z*). The Kuhn-Tucker Theorem is generalized to linear spaces. The way to do this is through consideration of programming in Banach and other spaces, where differenti ation is determined. The assumption of Theorem 8.40 that the cone Pz must contain a region substantially narrows the field of application of this theorem. The theorem can be formulated under substantially weakened assumptions. Theorem 8.42. Theorem 8.40 continues to hold for Z if the requirement that g{x°) be inside Pz for some x° > 0 is replaced by the following conditions: (A') for any nonidentically zero nonnegative continuous linear functional z* there exists xz- 6 X such that z*(g(xz>)) > 0; (A") the set A = {(t/, z) : y < f(x), z < g(x) for x > 0} has a nonempty kernel; (B) any element from Z can be replaced as the difference of two elements, each belonging to PzHere the kernel of cone Pz is composed of all elements such that for any direction in space Z there is a segment containing the point z and lying entirely in P%. The saddle points of the Lagrange function are considered where there is no dif ferentiability. Theorem 8.43. Let S be a, linear system and let Y and Z be linear topological spaces. Suppose that Py, Pz are convex cones in Y and Z with nonempty interiors; X is a (fixed) convex set in H; / is a concave function mapping X into Y; g is a concave function mapping X into Z. Let 3 the point x* £ X : g(x*) > 0 (i.e. g(xt) belongs to the interior of Pz). If x0 maximizes f(x) subject to g(x) > 0 then there exist linear continuous functionals y* > 0 and z* > 0 such that for the Lagrange function
*(*.*') = y'l/(*)] + **[s(*)] the saddle point inequalities ®(x,4)
<
${x0,z*)
hold for all x 6 X and z* > 0. Note that, in applications, X is a convex cone, namely a nonnegative orthant of system S. The following theorems were obtained for the case of differentiability (differen tiability is taken to be Frechet differentiability, spaces are Banach spaces, cones are closed). Theorem 8.44. (A) Let f be differentiable in X and let g be a differentiable function mapping X into Z. The cones Px = {x : x > 0} and Pz = {z : z > 0} are
8.8. OPTIMALITY
CONDITIONS
FOR IP
PROBLEMS
181
closed. Let x0 maximize f(x) subject to x > 0, g(x) > 0, and let g(x) be regular in x0. (B) From the relations x > 0 and 5g(x0, £)—g(x0) > 0 (( = x- x0) it then follows that -6f(x0;£) > 0. Theorem 8.45. (A) Let all of the assumptions (A) in Theorem 8.44 be fulfilled. Further, suppose the set W is regular and convex, while U and a are such that U(x) = 6g(x0; x),\/x £ X; a = Sg(xQ;x) - g(x0); x*(x)
e X;
= -Sf(x0,x)yx 13 =
-8f(x0;x0).
(B) Then there exists zj > 0 such that the Lagrange function $(x,z*) = f(x) + 3*[5(x)] has a nonnegative quasisaddle point in (XO,ZOJ!/O)> VO = ^ ,',e- satisfies the conditions Sx$((x0;z'0);() < 0 , V i >Q,x = x0 + (; &x${(xo;zo);x0) 5*9{{x0]4);Q
= 0;
= ([g(x0)] > 0 , W > 0,C = z* - z*Q;
Sz.^({xo;z;);z"})
= z'a[g(xo)] = 0.
Theorem 8.46. (Generalized necessary condition). (A) Let all the assumptions in (A) of Theorem 8.45 be fulfilled, and let the functions f and g be concave. Then the function $(x, z*) has a nonnegative saddle point in (xo, ZQ), where x0 is the maximal element. Theorem 8.47. Let X, Z, Px, Pz and g have the same meaning as in The orem 8.45 (including the regular convexity assumption for Wf and the regularity assumption for g). Let Y be such that, in the normed linear space Y there exists the bounded linear functional which is strictly positive on a closed convex cone K. Let us further assume that f is a differentiable function mapping X into Z so that x0 is the maximal element proper. (B) Then for some yj > 0 the function 4i(*,**) = !£[/(*)]+ *o[0(*)] has a nonnegative quasisaddle point in (x0,Zg,i/g). Here Wj = {w* 6 W*: w* = T*(v*), v* > 0, v' e V*}. For problem (8.45) there are regularity conditions [34] under which the minimum point is a saddle point of the Lagrange function. In a sense these regularity conditions were demonstrated to be necessary. In [122, 124, 125], necessary optimality conditions are given for the linear prob lem (8.48).
182
CHAPTER 8. INFINITE AND SEMI-INFINITE
PROGRAMMING
T h e o r e m 1.47. Problem (8.49) has an optimal solution such that 1. cx0 = F = G; 2. y'0(Ax0 - b) = 0; 3. y'Q + Ax - b = 0.
Theorem 8.47 corresponds to the complementary slackness theorem for finite LP problems. The formulated problem (8.48) fails to meet the nonnegativity require ments for variables. For the quasiconvex programming problem (8.55) with finitely many convex in equality constraints [138], necessary and sufficient optimality conditions are obtained in subdifferentiable form. The generalized Kuhn-Tucker conditions are provided for abstract problems of concave vector optimization in topological vector spaces[139]. In [126], optimality conditions are given in the form of Kuhn-Tucker theorem for problems in complex spaces. In [86], the generalized Farkas theorem is provided and necessary optimality condi tions are obtained as applications for general problems of nondifferentiable optimiza tion in infinite-dimensional spaces. Moreover, the alternative theorem is provided, the optimality and Kuhn-Tucker conditions are obtained as applications for mathematical programming problems. In [141], the necessary optimality conditions of the Lagrange multiplier rule type are provided under some assumptions of regularity. In [116], the problem of minimization on a subset of the topological vector space is studied from the point of view of the computational algorithms inducing specified minimizing sequences. Namely, the problem is defined to be an optimization problem on a sequence set in the generalized normed space. First- and second-order necessary optimality conditions are provided for the smooth and nonsmooth cases.
8.9
Existence and Uniqueness of a Solution to Infinite Programming Problems
Consideration is being given to a pair of dual problems such as (8.48) and (8.60) [38]. Some of the duality theorems, in particular, Theorems 8.37 and 8.38 are presented in [128]. The first assertion of Theorem 8.38 is provided in [105] on the assumption of solvability of w" (which can be assumed instead of admissibility of IT) and existence of the element (co,t><>) € Tint{{A*y* + x*,y*b) : y* > 0,.r* > 0} and fits in the general scheme of proving duality [120]. The use of the BourbakiAlaoglu theorem makes possible the proof of not only the equivalence of condition 3) in Theorem 8.38 and more convenient conditions 2), 4) and 5), but also the existence of a solution to the primal problem and the weak compactness of an £-optimal solution. Referring to [128], we have the theorem which provides a framework for link ing the duality, the compactness of a feasible set and the existence of a solution to
S.JO. METHODS AND ALGORITHMS
FOR SOLVING IP PROBLEMS
183
problem IT. Theorem 8.48. [128] Let: 1. n and IT* be a pair of dual problems, with n feasible, and let 2. p+ be solid; 3. x > 0, Ax = 0 => x = 0. Then: 1. w and n~ are in duality; 2. the feasible set n is weakly compact; 3. TV has a solution. The linkage between duality and existence of a solution for linear and convex infinite programming problems is also observed in other literatures, e.g. in [100]. Existence of solutions is proved for the class of parametric linear programming problems [100] where linear problems are studied in metric spaces. Just as the existence and uniqueness theorems for SIP problems (section 8.4), compact metric spaces are examined [98] and the uniqueness theorem is formulated for compact Hausdorff spaces X. Theorem 1.49. [98] There exists a compact Gs-set in 6 = 2X x cx such that any problem {f(x) ->• min, x e A, (A, / ) 6 0} has a unique solution. As suggested in [98], Theorem 8.49 is also valid for some classes of nonmetrizable compact spaces, specifically for weak compact subsets of Banach spaces or spaces which are homeomorphic with respect to weakly compact subsets of the dual Banach space possessing the Radon-Nikodim property.
8.10
Methods and Algorithms for Solving Infinite Programming Problems
For problems like (8.47), the procedure of simplex method [44] is generalized and proves to be finite-step even in the finite-dimensional case. The point xQ <E W is called an .E-point of W if (kerA)nt
fl
kerft) = ff; H = (ker A) f)(f)
ker ft).
The E-point of the polyhedron W is called nonsingular if \J(x)\ = diam{keiA \ H); the polyhedron W is nonsingular if any one of its E-points is nonsingular. Theorem 8.50. (On improvement of a basic solution). Let W ^ 0 and be nonsingular, H C kerc, and let x0 be an E-point ofW. Then, in ker A, there exists a collection of vectors e3, j £ J{x0) such that Pj(ek) = 5]k, j,k 6 J{x0), with Sjk as Kronecker delta. (I) If c(e1) < 0, Vj 6 J(x0), then x0 is the point of maximum of the functional c on W
184
CHAPTER
8.
INFINITE
AND SEMI-INFINITE
PROGRAMMING
(II) Ifc(eM) > 0 for some j 0 £ J(x0) and f3j(ek) > 0 Vj £ 1 , . . . , n \ J ( x 0 ) , then the functional c is not bounded above on W (III) If c ( e , J > 0 a n d t h e set I(x0) = {j £ l , . . . , n \ J ( s 0 ) : & ( e i o ) < 0} is nonempty, then: 1) there exists a unique j \ £ /(xo) such t h a t e = \bn - Pn{x0)\l\Pn(eK\
= m m 16,- - 0 i ( x o ) | / | A - ( e * ) | ;
2) the point y = x0 + eeJ0 belongs to W and is an 3) c(y) > c(x); 4)J(y) = (J(x)[j{j1})\iJ0}; 5) the vector collection gj, j £ J(y), where
E-point;
(8-68)
9h = e3o/0jAeio); g, = e, - {fih{ej)l0h{ek)^,j
£ J(y) \ {ji}
(8.69)
satisfies the conditions g} G k e r A , j € >/(y), 0}{gk) = Sjk, k G J(j/)- If H <(_ kerc, then c is not bounded on W below or above. T h e o r e m 1 . 5 1 . [43] (simplex method). Let W satisfy the conditions of Theo rem 8.50. Starting with any E-point of the polyhedron W, in a finite number of steps (one step is described in Theorem 8.50) it is possible either to reach the E-point in which a maximum of the form c is attained or to establish that c is not bounded above on W. In particular, any linear form either attains a maximum in one of the E-points or is not bounded above on W Note t h a t t h e number of steps in t h e simplex procedure does n o t exceed c™ = n\/m\(n — m)\, where n is t h e number of inequality constraints in problem (8.47), rn =0, 3 e 0 > 0, h, i £ 1 , . . . , n in R such that \b{ — bf\ < £, j £ 1 , . . . ,n and for 6,-, i € 1 , . . . ,n such that \b, — B,| < e, i € 1 , . . . , n, the polyhedron W defined by the conditions A(x) = d, (33(x) > b3, j £ \,...,n, is nonempty and nonsingular. In [107], some numerical properties of the simplex are examined for t h e infinite case, specifically a new definition is given to the infinite simplex with a countable set of extreme points. Consideration is being given t o t h e question as t o which of the simplexes from this class allow approximation by their finite subsimplexes. In [44], t h e dual simplex method is extended t o cover convex infinite problems with linear constraints like (8.47). T h e theorems given here a r e largely similar to Theorems 8.50 and 8.51. For quadratic infinite problems [42] there is the Wolf algorithm, which is as follows: firstly, one has t o solve the problem in which some of t h e inequality constraints are
8.10. METHODS AND ALGORITHMS
FOR SOLVING IP PROBLEMS
185
omitted, while the other are replaced by equalities; secondly, the resulting solution is "improved" by a finite-step procedure over a finite-dimensional space. Namely, imple mentation of the algorithm calls for solution of the corresponding infinite-dimensional problem, which, in a sense, governs the choice of a function space. If, however, one is successful in solving the infinite-dimensional problem, then implementation of the second part of this algorithm is automatic. The infinite-dimensional part of the algo rithm is implemented by solving a differentiable matrix equation of Riccati's form. Referring to [11], we have an extremal problem with infinitely many criteria and constraints for the smooth and nonsmooth cases. Lagrange multiplier rule is stated for deriving necessary extremum conditions. Referring to [18-22], we have conditions for application of the marquee method developed in the finite case to extremal problems in infinite-dimensional spaces. The separability theorem for the system of convex cones with the common vertex at zero is generalized to the infinite case. T h e o r e m 8.53. Let E be a locally convex topological space and let K\,..., Ki be a system of I convex cones in E with vertices at zero such that 1) ViK, ^ 0, i = 1 , . . . , / ; 2) if L\, Li C E are subspaces of E each of which is represented by an intersection of some subspaces from the system aff A' 1 ,...,aff Ki, then L\ and Li are in the common position. This cone system then has the separability property if and only if there exists o,i G A'j*,... ,aj G K*, a, = 0 for any i, i = 1 , . . . , / such that ai + .. - + ai = 0. As in the case of finite-dimensional spaces, no assumption is made about sol id properties of the sets and their approximating cones, wherein lies a distinction between the known results and those obtained from the marquee theory.
Chapter 9 Optimization on Fuzzy Sets In large and complex systems, the optimization problem is difficult to state formally. Furthermore, in a number of cases the main system variables are fuzzy and given approximely, and their stochastic characteristics are sometimes impossible to define. When functioning, the model experiences uncontrollatle perturbations such that its structure mat substantially change, i.e., the model initially regarded as a linear one can be transformed into a nonlinear problem, etc. Then it seems wise to ensure stability of the model for the entire period of its functioning. Note that this is crucial for long-term planning models. In the general case, the model stability will be interpreted to mean the property that the optimal solution obtained from utilization of the optimization model is invariant under model modifications such as changes in parameters or structure. If the model stability conditions are not satisfied, e. g., when perturbations produce changes in the model parameters, and the previously obtained optimal solution is not preserved, then one has to define the limit solution set (a solution cone) for this level of perturbations. This section deals with a theoretic background and applied aspects of the solu tion of nonuniquely formulated optimization problems which formed the basis for development of the estimation algorithms for stability of optimization models under uncontrollable perturbations. Consideration is given to the main features of inad equately fomalized optimization models, the techniques designed to synthesize the solution algorithms for such a class of optimization problems, the methods for esti mating the stability of optimization models under uncontrollable perturbations, and the issues of development of applied program packages for solving inadequately for malized optimization problems.
9.1
Optimization problems in Large and Complex Systems
The modern framework of solution for optimization problems suggests that the stat ed problem is unconditionally formalized, i.e., there are a uniquely defined criterion, 187
188
CHAPTER 9. OPTIMIZATION
ON FUZZY SETS
an objective function or functional, or a set of criteria, a uniquely defined system of constraints. If the proposed problem is not formalized, i.e., when the system of constraints and criteria is not uniquely defined, then it is considered to be mathemat ically incorrect or its solution is inappropriate, or additional information is required to make this problem correct and solve it by employing traditional methods. Note that the absence of uniqueness is the general case of problem statement and a special case of incorrect problems. This class of problems is mentioned here because very efficient algorithms are provided for solving some types of incorrect problems, and if general conditions of uncertainty allow a problem to be reduced to this type of problems, then it can be solved in an exact or approximate way. In the general case, optimization problems can be classified as the deterministic, the stochastic and the problems to be solved under uncertainty. The deterministic problems employ completely determined information. The me thods for solving these problems are the most diversified ones. In the stochastic optimization problems, provision is made to express the conditions of model perfor mance and its variables in the form of probability relations, i. e., the distribution laws for random variables or random functions. The difficulties in solving this class of problems are as follows: firstly, the distribution laws derived for a restricted amount of statistical data are not necessarily reliable or stable, because a slight increase or decrease in the amount of initial a priori data may completely alter the design law of distribution; secondly, multidimensional distributions are very difficult to compute; thirdly, the distribution law is not necessarily derived. A special feature of the optimization problems to be solved under uncertainty (or under incomplete information) is that the stochastic characteristics of the sys tem (object) under study cannot be obtained. Although these problems are the least developed and there are no universal algorithms for their solution, certain recommen dations can be given here. Basically, such problems must be solved by employing a comprehensive approach, i.e., applying various methods or their combinations. The comprehensive approach includes the following phases [3]: — definition of the set of conditions to characterize the model and generation on their basis of the set of optimization problems to provide an adequate (to some extent) description of the object; — solution of the formulated optimization problems and identification of the un certainty zone for each of the obtained solutions; — adaptation of every solution to various combinations of initial data; — finding solutions in the uncertainty zone. Although this approach seems to be comprehensive, it does not allow one to syn thesize constructive solution algorithms for optimization problems under uncertainty. This is due to the following facts. If, e.g., the linear form is given as intervals Cf'n < d < Cf^, Vi, then in order to form the set of problems which reflect the set of conditions characterizing the object (alternatives), one has to take into account 2" linear forms, where n is the dimension of the vector of unknowns, i.e., one has to
9.1. OPTIMIZATION
PROBLEMS
IN LARGE AND COMPLEX SYSTEMS
189
solve 2 n linear programming problems. It is easy to see that even for comparatively small n the number of problems to be solved is sufficiently large. So it is essential to form the minimal set of problems (alternatives) which can adequately reflect the actual performance conditions of the object. The existing methods of solving optimization problems under uncertainty, as a rule, take into account only sufficiently small changes in the objective function co efficients and model constraint system. Actually, such methods do not allow for variations in the model structure. In the general case, the solution of the optimization problem under uncertainty must be taken to be some set of vectors. The principles of selecting "coordinated" decisions have not been investigated and are largely dependendent on the intuitive prerequisites of a decision maker. Investigations show that under actual performance conditions of various control systems one has no opportunity to rely on initial chance-constrained information, and the optimization problem has to be formulated with due regard for this characteristic property. In these conditions, one has to ascertain whether the model is adequate and whether it can be successfully utilized for specific (nonuniquely defined) performance condi tions of the system (object). Note that an attempt to solve a nonuniquely formulated problem is necessarily associated with an attempt to formalize this problem. Then care should be taken to prevent the substance of the problem from being corrupt ed by formalization. One may often observe situations, where formalization of the problem amounts to "fitting" an actual model to a programming model such that its solution becomes accessible to a specific executive. Such a "universal" model is often represented by a linear programming problem. In view of the diversity of nonuniquely formulated problems and because of the necessity to solve such problems, some methods were proposed for formal description of qualitative relations. These methods are based on fuzzy sets as in L. Zade [22]. To solve approximate decision problems in the nonuniquely defined setting, one may utilize fuzzy F-sets which can be viewed as models for problems of some classes whose parameters have no clear-cut boundaries. According to [22], the fuzzy F-set is defined as follows. Let X be a universal (ordinary) set of elements. Suppose the F-set A C A' is defined as a collection of pairs (HA{X),X;X e X), where \iA is the belonging function of x -> [0,1], while HA{X) is interpreted as a degree to which the element X belongs to the F-set A. Among operations on F-sets (see [22, 16]) we select only those — inclusions, intersection and union which are necessary for further discussion: a) A C B & pA(x)
< HB{X), (Va); x G A';
6) C = A U B •£> pc[x) = min {nA{x)} ; x £ X; X
B) C = A n f i H / i c ( i ) = max {/J.A(X), ^B( )}
(9.1)
;x e X.
Of special interest are optimization problems whose solutions are based on fuzzy sets. Let m a x / ( i ) ; x 6 X be a uniquely defined traditional optimization problems,
190
CHAPTER 9. OPTIMIZATION
ON FUZZY SETS
and X" a solution vector. In the informal setting, the optimization problem can be represented as follows. There are a set of objectives (which contains only one element for the uniquely defined problem) and a set of choices whose intersection allows one to define an optimal solution of the original problem (as is evident from the foregoing, the intersection of these sets may generally contain more than element, in which case the problem does not have a unique solution, or it may be empty, in which case the problem has no solution). The problem of achievement of the objective described by the fuzzy set with the belonging function \ic '■ B! —^ [0,1],
f Mc =
lj{x)
(9.2)
\ 0,/( S ) > / ( ! ' )
can be described by incorporating the method of execution of nonuniquely defined instructions into the formulation of the problem. We obtain the optimization problem fi(x) = min {/ip(x), HN(X)}
—> Extr(max),x g X,
(9.3)
where pip(x) is the belonging function which characterizes the pre-image of the F-set B C X for / : p.p{x) : x —*• [0,1]; JJ,N(X) is the belonging function which characterizes the F-set of N feasible choices for HN(X) : x —> [0, 1], . .
f 0,/(x) < /(**); >/(x* v
W(z) = |
^W
:*), , = ]f 01,31,,i1 6g /AT;VN;;
r Thee solution so lution of this problem is the fuzzy set wwhich hich is the intersection of the fuzzy sets of objectives and choices [16]. The optimization problem [9.3] does not allow one to obtain a unique solution, although such a solution may actually exist. This is a serious handicap to the use of F-sets for solving optimization problems where no adequate description is provided. Despite its apparent universality, the proposed approach to optimization problems as a way of realizing nonuniquely defined instructions for achievement of a nonunique ly defined objective does not allow one to synthesize constructive applied algorithms for solving actual nonuniquely formulated optimization problems. This is mostly due to the fact that the problem of constructing a feasible choice set is essentially the prob lem of determining feasible solutions for the original optimization problem, which is highly conjectural. So, without rejecting Zade's approach in the general case, one has to accept its nonconstructivity, which in turn necessitates classification of nonunique ly formulated optimization problems and development of applied algorithms for their solution. Investigations show that in the actual solutions of optimization problems for large and complex systems the system parameters are often represented by nonuniquely defined variables. The nonuniquely defined variable can be interpreted as follows. It is well known that the fuzzy set A C V is characterized by the belonging function
9.1.
OPTIMIZATION
PROBLEMS IN LARGE AND COMPLEX SYSTEMS
191
H'A{X) which assigns some number x 6 A (the belonging function) to each element HA ~~^ [0,1]. In a first approximation, the element of the fuzzy set is regarded as a nonuniquely defined variable. The definition of the nonuniquely defined variable which is based on the concept of assignment equations [22, 16] is more rigorous, and hence more extensively applicable. To describe the variable as an element of algorithmic language, one has to indicate the name of the variable (identifier) and the range of values assumed to be this variable. Formally, these requirements can be satisfied if the variable is described as a group (triple) of parameters (A', U, R(X; u)) each of which can be interpreted as follows: X is the name of the variable (identifier); U is a universal set to which the variable may belong. The range of values A' — R(X;u) or, for simplicity, R(X) can be viewed as a constraint which is imposed on u £ U and specified by the variable X; if x — u, u £ -R(A'), this is equivalent to the existence of the assignment equation for the variable x = u : R(X). In other words, the assignment equation reflects the fact that the variable cannot assume other values than those which belong to the range of its values. If the variable is nonuniquely defined, this means that the assignment equation for the nonuniquely defined variable x = u : R{X) is nonuniquely defined and reflects the fact that the element x is assigned the value u under constraint R(X). The degree to which this equality (assignment of the value x to the u) is satisfied is called the consistency of the value u with R(X). This reasoning can be written as c(u)
= HR{X){U);
U£
U,
where HRIX){U) >S the degree to which the u belongs to the constraint R(X). Note that for the uniquely defined variables (J.R(X)(U) — 1 if the assignment equation is satisfied and HR(X){U) = 0 otherwise. The belonging function and the belonging degree are actually equivalent concepts. Application of fuzzy sets will be illustrated for formalization of the nonuniquely formulated optimization problem. Technical propositions often call for minimization of the functions referred to as posinominal:
max/(i) = 2J
n x.
,;
x £ X
(9.4)
Here ambiguity of mathematical description manifests itself in: — the indeterminate functional structure (because the parameter — the exponent k — can assume various values); — the nonuniqueness of the parameter (the components of vector c = {c,}; j = l,m can assume different values belonging to fuzzy sets); — that the constraint set X determining the system performance conditions may belong to fuzzy sets.
192
CHAPTER 9. OPTIMIZATION
ON FUZZY SETS
When problem [5] is inadequately described at the level of nonuniquely defined variables, it can be represented as max (U(c)), xu^)
;
x 6 U(X),
where parameters c, k "approximately" belong to a fuzzy set and can be represented as (V?)(ci = c,4 approximately c, = 1/cJ + fi]/c} ± Ac] + /i 2 /c 2 ± Ac,2 + . . . = (9.5) k = /cA
approximately
=U/ii/fe ± AJ.
k = 1/fc, + ^i/fci ± A^ + id2/k2 ± Af + . . . =
(9 6)
'
i 3
where fi t, fit are the functions of belonging of the elements c;, k to the sets U(C), U(/c), respectively. The sign A_ means "by definition is" The notation (9.5)-(9.6) means that the parameters c and k can assume any values from a given set, in which case for every assumed value there is a particular value of the belonging function. Note that, unlike probabilities, the belonging functions form a complete set of events, i.e., Yl^k / 1. k
Problem (9.4) is rather difficult to solve even in the uniquely defined setting. The difficulties encountered in the solution of problem (9.4) with due regard for nonunique definition of model variables are much greater than those in the unique formulation of the problem, because the construction of the feasible choice set is the feasible solution set for the original problem (9.4). As noted before, such solutions are sometimes difficult to obtain. [6] proves the validity of the extension principle which makes it possible to extend the basic theorems and concepts of uniquely defined sets to fuzzy sets. The possibility to solve problems (9.4)-(9.6) allows one to determine the boundaries for applicability of the theorems of necessary extremum conditions in convex optimization problems under nonuniquely defined conditions for model performance which, in turn, allows one to make the decisions more adequate. One cannot claim that the solution methods of traditional (uniquely formulated) optimization problems can be extended to problems with nonuniquely defined vari ables. The resolution of this issue is of fundamental importance. Mathematics is now equipped with numerous algorithms and methods for solving optimization problems. So there is no need to develop new methods for solving special classes of problems without examining the potential of the familiar methods.
9.2
Optimization of Problems with Nonuniquely Defined Parameters
General recommendations on the solution of nonuniquely formulated optimization problems are given in L. Zade [22] and include implementation of procedures for con-
9.2. OPTIMIZATION
OF PROBLEMS
WITH NDP
193
structing fuzzy sets of objectives and choices and the domain (set) of their inter section. Mathematically, these recommendations may often involve insurmountable difficulties, and provide no means for synthesizing constructive algorithms intended to solve actual applied optimization problems in the nonunique setting. One of the special, but important issues is the estimation of optimal solution invariance in nonuniquely formulated programming problems where the model param eters are nonuniquely specified. The general formulation of the algorithm synthesis problem for determining invariant solutions as a prerequisite for synthesis of solution algorithms of the convex programming problem is very complicated when mathemat ical description is inadequate. So it may be wise to distinguish special classes of problems, where such a formulation is not only possible and permits machine real ization, but also provides the openings for solving the general problem. The most suitable for this purpose are linear programming problems whose general formulation for nonuniquely defined model parameters is given below. Application of traditional methods for solving linear programming problems, where the coefficients of a linear form are nonuniquely specified, requires solution of 2™ lin ear programming problems (n is the dimension of the vector of variables). For the cases of moderate dimension, the number of problems is rather large, which is not only wasteful of machine time, but also involves perceptive disturbances in accuracy (because of the iterative nature of procedures). Computational procedures become more complicated where variations in constraint parameters have to be allowed for. This emphasizes the advisability of applying special algorithms which are not required to solve the optimization problems with nonuniquely defined parameters by reducing it to a particular set of correct prob lems. The stability (or invariance) algorithm of the optimal solution is to ascertain whether the functional gradient vector belongs to the cone of allowable variations for all the functionals from a specified set of the optimization problems. When synthesizing the estimation algorithm for the solution stability under vari ations, the main tasks in determining the structure of the algorithm are to find an advantageous (from the point of view of machine realization) way of specifying the allowable variations cone and the functional cone, and the ways of determining the set of solvable problems and selecting the design optimization problem. Additionally, it is very important to examine the efficiency of the developed algorithm from the point of view of its machine implementation and from the point of view of isolation of those classes of problems which can be effectively solved by the developed algorithms. Note that one of the classes of such problems includes the estimation problems of the optimal solution stability under perturbations.
194
9.3
CHAPTER 9. OPTIMIZATION
ON FUZZY SETS
Basic Notions
We will consider the design problem as a linear programming problem such that its linear coefficients are nonuniquely defined variables max ([/(c), x) ;
X = {x\Ax
i £ l ;
x > 0} ,
where U(c) is a fuzzy set of coefficients each of which belongs to a fuzzy subset ''approximately Ci" Ci^
approximately
Z\ = n\\c\ + fJ."\c'( + . . . + p.\\c\\
c2A
approximately
c2 = p.'2\d2 + /4'lc2 + ■ ■ ■ + A^kli
c„4
approximately
cn = fi'n\c'n + ^'|c'„' + . . . + n'n\c'n.
The notation (9.7) implies that the coefficient ci can assume values c\,c'(,... ,c\, and the quantity JJ,\ shows for the coefficient c\ the degree of its belonging to the universal set of coefficients {c[, c",. . ., c\}, that is, C\ = c\ with the belonging function Hi. So the coefficient can assume any value from the universal set {c[,.. ., c\} with a suitable belonging function. In what follows the notation (9.7) will be used in a more compact form: c,A
approximately
ct =U f4 |c; , Vi,
j = 1,2,. . .
It may be wise to solve the linear programming problem with nonuniquely defined coefficients of a linear form in two stages: 1) solve a simpler problem, i.e., the linear programming problem with the coeffi cients given at intervals, and then 2) solve the general problem. Suppose that in the problem max(c, x);
x £ X;
X = {x\Ax
x>0}
the coefficients of a linear form are given at intervals c™ln < c, < c f ^ V t . Define the conditions under which the solution of this problem is unique. If the solution is not unique for a given interval of variations in the coefficients of a linear form, then define the limit solution set which allows one to find a. coordinated solution. The limit solution set is the set which can be obtained when all LP problems are generated on the basis of the condition cfln < c; < cf'^jVi. The coordinated solution is that solution £' for which f = m a x | s - i * w | -> min,;r* w £ RX)RX
= {.T*1''} ,i = 1,2,. . . ,
9.3. BASIC
195
NOTIONS
x* w is the solution of the LP-problem max(c ( 'l,i); x G X. Let us form the matrices C\ and C-i where each row is constructed on the model: one element is maximal, while the other elements are minimal, and vice versa. We have diagC 1 {c; nin ), diagC2 = ax
{c }>
/^min
/"'max
o2
/~>max
/Tnax
o2
Ci fmax
(9.8)
/^raax
fmax
/~<min /^min
^2
/~*min
r>max
/"•mill
(9.9)
C2 /~<min
/^min
/^max
The number of rows in the matrices (9.8), (9.9) is In. Note that the number of vectors whose components assume only two values cf ax ,c™ m , Vi is 2" Here (9.8), (9.9) are the limiting matrices. T h e o r e m 9.1. Let the vector c = {c,}; i = 1,», be given by the upper and the lower values of its components cfm < c, < c| nax ,Vi. Here a) the matrices C\ and C2 (9.8), (9.9) are such that
diagCi = {Cr}
diagC2 = { c r x } ;
,
l,n;
b) the circular solid cones A'+ and A' are such that K+ is circumscribed about Ci, A"~ is inscribed in C2 and the center of cones A'+ and A"~ is defined as follows
£
C;=
!.<> (9.10)
i = 1,2n
2n
Then any vector S $ C = C, U C 2 , for whose components Cfn has the property of indlusion ck 6 A*+ x A'~
< Cf1 <
Cf^.Mi,
Proof. Let us enumerate the vectors c(j> G d U C2 = C; c ' = c « / | | c W | | , | | c ( i ) | | = (c' ',c' J ') 1 / ' 2 ; j = l,2n and calculate the direction cosines a' J ' between the vectors c' J ' S Ci U C2 = C; j = 1, 2n and the normalized vector c: J
o^=(c,c(;) o a
2
o(j)
c,c
oU)
c o(j)
c
_
e Ci; ,-,
€ C3;
l,2n.
196
CHAPTER 9. OPTIMIZATION
ON FUZZY SETS
In the arrays < a\ > and < cq \, we define respectively the maximal and the min imal element, a^ = m a x j o ' | , a'"' = m i n j a j } a n c ' t n e vectors c^K &v' forming the angles area' 1 ' and area'"' with respect to C. Take the generatrices to be the circular cones A'+ and K~, respectively. We will consider the scalar products (c,c), (c,c), where the vectors c, c are such that among their components there are at least two or more components c™ax (or c™111). If we compare ( c , c J with f c, cj, where c'A' is any row vector dk> f C, then we may see that o{k)min
c T •. .T
o( )rnm Cn
0
0 (A:)min
■ cx + c2 „ -a C „ ^ Cj
0(t)max
0
■ c2+...
+ cf
o
o(^) min
o
■ cj + c / + 1
„ , -min ° , i -max ° ,;mii ° Cj -t-C 2 ■ Cj -f- . . . -f- Cy • Cy -|-Cy_|_j • C j +
c/+1 + 1
. . ° -min ■+■ . . . -f- C„ -C„
because the number of maximal components is larger in the product (c, c) (note that all £, > 0, c] are normalized). Hence the angle between the vectors c and c is always smaller than a' z ', i.e., the condition c £ A'+ is necessarily satisfied. For the vectors c, whose components include two (or more) c™m, one may show that there is always {c,c) < (c^,c) which means that a'"' > (cA"',c), i.e., c f K~ Hence the assertion c G K+\K~ is true. The theorem we have now proved makes it possible to state the existence condi tions for the unique solution in the LP problem with the coefficients of a linear form specified at intervals. It has been known that the solution of the LP problem is regu lar, i.e., it is placed at the node of the polyhedral set X\x\Y^o,ijXt < bjX > Of; j = l,m, which forms a constraint system. Let x" be a solution of problem max(c'*',i); x 6 X = \Y,aijXi < bj, x > 0J; j = l,m where c'*1' is the fc-th row of matrices (9.8), (9.9) which is selected in an arbitrary way. We may distinguish a subset of constraints M C X, M = lx\ ^ a . j i ; < bj,x > 0J; j = l,u which form the roof at the pont x'\
(v,-) ( E «<»*• = &*) i
j — i,f.
By the theorem of necessary and sufficient extremum conditions [22], we have the proposition if c[k] 6 D (M(k)), then x* = max (c[fc], a;); Vc^; h)
xeX,
it = 1,2,...,
where D(M ) is the dual cone which is constructed at the point x* and formed by the normals of hyperplanes J2 aijx] = bf, j — 1, v, forming the roof at x~ So, in order to establish the fact that the solution is unique in the LP problem (9.7) with the coefficients of a linear form specified at intervals, it is necessary (and
9.3. BASIC
197
NOTIONS
sufficient) to run 2n tests for inclusions of vectors c(k] 6 C = C\ U C2 in the cone D(MW), or, what amounts to the same thing, for inclusion of the cone A'+ in the cone D(M). To define the limit solution set (cone), it is advisable to utilize the following algo rithm. Assuming the limit matrices (9.8), (9.9) are constructed, the computational algorithm scheme can be represented as follows: Al:l. Formulate the original problem. max (c' 1 ', xj ; X = {x\Ax
x 6 X; .T>0},
1
where C' = {C, }; i = ~L/n is the first row of matrix C\. 2. Solve the original problem in 1, and fix the solution vector A'*(1). 3. In the constraint set X = lx\ J^a^x, < bly x > 0}, find those constraints which form the roof at the point A'*(1). 4. Construct a polyhedral cone D(M'-1') whose edges are the vectors atJ of the roof hyperplane normals j = l,v; i = l,n, 5. For all row vectors of matrices (9.8), (9.9) starting with the second, the se quential test is run for inclusion of the vector c'", k = 2 , 3 , . . . in the cone D(M^). The vector designated with number / , whose direction does not belong to D{M^1'), is taken to be the estimated one. State the problem max f c ( / ) , x) ;
x € X = {x\Ax < 6, x > 0} .
6. Problem in 5 is taken to be the original, and transition is made to para 2. Implementation of the algorithm will yield a solution set x*'1', i*' 2 ',...,, which can be ordered or an acceptable solution can be selected as a vector x determined as r r = max \x — x"^'\ —¥ min;
r*'1' G A',
or a vector x= {x,y. 0
s !. ==i1
•W/5;
i = 1 ,n.
, i + ir»n ct»t .^karl about - k ^l-,o solution The solid cone Rx circumscribed the set a;*'1', x*' 2 ',. . ., has the condition that the solution of any LP problem which can be based on interval speci fication of model coefficients belongs to the cone Rx. The validity of this inclusion follows from the fact that any vector c- > ^ C which can be based on the condition cfm < c, < c™ax, Vj is included in the cone A' + , i.e., c<*> G A'+. It should be noted that, in the general case, replacement of the set 2 n vectors with the matrix C = C\ U C2 containing 2n vectors may give rise to errors in the definition of an acceptable solution, but such errors may not affect the result in the general case.
198
CHAPTER 9. OPTIMIZATION
ON FUZZY SETS
Since the procedure for testing vector inclusions in the polyhedral cone D(M) (the cone of allowable variations) is rather complicated (from the point of view of machine implementation), it is advisable to introduce two types of tests: approximate and accurate. The approximate testing covers the inclusion of vector C' ' in the allowable variations cone which is a circular solid cone Qx inscribed in the allowable variations cone. We well describe the operator scheme for testing the belonging of vector directions c'fc' to the polyhedral convex cone Qx(c^ € Qxj- This scheme is implemented in the "accurate" version of algorithm and includes the following procedures: A 1': {AiAiAzAtAsAeA-rPs} , where A\ — definition of the center of gravity for the vectors of normals a tJ of hyperplanes J2aijxi — bj] j = l,o, which form the bundle for x", i V
Ci= ^ a , j / n ;
i = l,n;
3=1
A2 — formation of a vector collection {ot,3}; i = l,n; j = l,v, whose norms are respectively fj =c /ctf, a3 —the direction cone between vectors c and cM>] A3 — calculation of the direction cosine j3k between vector c' ', c and generation of vector c'" with norm rc =c ff}k; Ai — definition of the "zone", where the vector c'" lies; A$ — formation of an equation for a straight line passing through the points c and o mk
°c,
X
i~
O C
l
_
_ c < * > -■■■-
x
n~
j
O c
n ik)'
A6 — definition of the intersection point a;'" for the straight line m\ and the straight line q(k> closing the zone; A7 — calculation of a distance from c° to c- ' ans x'fc' — r^ck\ r^ respectively; Ps — testing: r<*> < r<*>. Fullfilment of condition P 8 shows that the direction of vector c'fc) belongs to the cone Q. The zone, where the vector c'*' lies, is determined by the algorithm whose operator scheme is represented as the relation AI" : {A,A2P3}
,
where At — definition of the collection of direction cosines j3j, j = T~u between vectors mk and bj (the direction vector Cj coincides with the straight line passing through the points c and a,-); A2 —definition in the collection {ft}; j = l,t> of the maximal element ft C {ft} and the corresponding vector 62;
9.3.
BASIC
199
NOTIONS
P 3 — t e s t i n g : ft+1 > & _ , ? Suppose t h e vectors {a,.,} are ordered, i.e., the sequence of vectors 0,1,0,2,. ■ ■ ,a,v; i = l , n is defined to be such t h a t the straight lines passing through the points a,k -T- a,fc+i; 2 = l . n ; fc = l , v form a closed convex planar polyhedron. T h e "zone" is a triangle whose sides are straight lines passing through t h e points c and gt, (c;a,t), (ajfcjajt+i), respectively. T h e straight line passing through t h e points (ajjort+i) is referred to as closed. If condition P3 is satisfied, the "zone" is formed by t h e straight lines passing through t h e points Cc;c*z), ( C , Q 2 + 1 ) , (az;az+i); otherwise {c;a2), ( c ; a , . i ) , ( g ° ; a , _ i ) . Note t h a t t h e solution set £*(*>, A; = 1 , 2 , . . . is t h e limit one in t h a t t h e solution vector of any LP problem, whose linear coefficients satisfy t h e condition c™m < c, < c™ ax ,V,, has t h e property t h a t its direction belongs to t h e solution cone a;*'", k = 1 , 2 , . . . This fact has escaped t h e attention of investigators. W h e n compared to t h e algorithm of t h e LP problem which has t h e coefficients of a linear form specified at intervals, t h e nonunique definition of coefficients is allowed for in t h e following way. T h e belonging function can b e assigned to each row of ma trix C. Since t h e matrix row is taken as a linear vector and vector components are noninteracting variables, t h e belonging function of t h e row (vector) must be taken, according to (9.6), as t h e maximal value of all belonging functions of separate com ponents. Consequently, t h e matrices C i , C2 can be complemented by the columns of the belonging functions. So t h e belonging value corresponding to t h e row is assigned to t h e solution of each LP problem whose linear form represents t h e row of matrix C. T h e proposed algorithm can be used most efficiently where there is no need to ran domize solutions (indicate t h e belonging functions of each of t h e solutions obtained) or where t h e solution is invariant with respect to nonuniquely defined coefficients. Solutions can be randomized in t h e following way. Formulate t h e LP problem whose linear coefficients are selected at t h e a-level 1 t h a t is equal or close to 1. Suppose t h e problem is max(cw,i);
x 6 X;
X = {x\Ax
x > 0}
(a)
and vector x* is its solution. As in (9.8), (9.9), we formulate t h e limit set of linear vectors by taking t h e lower b o u n d a r y of t h e a-level is a specific value aA < a 2 . After constructing t h e allowable variations cone D(M) on t h e solution x*a and testing t h e vectors directions for belonging to t h e cone D{M), one m a y establish t h e optimal 'The a-level of a fuzzy set A is an ordinary set Aa of all the elements of a universal set U such that the degree to which they belong to the fuzzy set A is greater than or equal to a : Aa = |u|/j/i(") > a \. The fuzzy set A can be decomposed in terms of the sets of levels: A = U1 a-A a=a
CHAPTER 9. OPTIMIZATION
200
ON FUZZY SETS
solution invariance with respect to nonuniquely defined coefficients or determine the value of the a-level ji = aD for which the invariance conditions are preserved. Formulating a new LP problem with the coefficients selected at a new a-level D a a{P) < a( ) yields a new solution x*' ' which can be associated with the belonging p function value fip = a' '. The solution with the maximal belonging function value, i. e., i*' 1 ' can be taken as an acceptable solution of the LP problem with nonuniqiely defined linear coefficients.
9.4
Solution Optimization Algorithms for Linear Programming Problems with Variations in Constraint Coefficients
The principle of duality in mathematical programming problems implies that the correct (uniquely defined) LP problem maxfc, x)\
(P):
x 6 X;
X = {x\Ax
x>0}
can be placed in correspondence with the dual (9.9). We will suppose that in the primal problem the constraint system is given ap proximately, which can be generally represented as interval definition of the vector of right-hand sides b e U(S);
bfn
6;max;
Vj;
j = I~^.
As noted before, with the linear coefficients defined nonuniquely at intervals, cf™ < c, < c™ax, Vi, i = l,n the optimal solution invariance condition can be represented as cU)eQx;
V/;
f = T^H,
where Qx is the cone of allowable variations and c"' is a row vector from the limit matrix C — C\ U C 2 . This provides a way of constructing the estimation algorithm for the optimal solution invariance under variations in constraints: 1. Set up the correct (uniquely defined) LP problem, where the components of column vector b are arbitrarily selected from the interval [6fax;6™m]. We have the problem max(c, ;r); x € A'; A = [x\Ax < b{s\
x>0}.
2. Set up and solve the dual problem min (6 ( i ) ,y) ;
I E A';
ye V;
9.4. SOLUTION
OPTIMIZATION
201
ALGORITHMS
Y = {y\Ay>C,
y>0};
and fix the solution vector. 3. Based on (9.8), (9.9), set up the limit matrices Bu B2\ diagBj = {b™x}, diag£ 2 = {b™}; i = l,n. 4. With the solution y* for the problem in para 2, construct the allowable variations cone Qx. 5. Test the vectors 6<J' £ B = B, U fi2 for bel onging to the allowable variations cone Qz. 6. If the condition in para 5 holds for all the rows (vectors) forming the limit matrices By and B2, this is a necessary and sufficient condition for the optimal solution to remain stable in the given range of constraint variations. 7. If the condition in para 5 is not satisfied, then the first vector 6"', whose direction does not belong to Qx, is used for setting up a new problem min(6(/),y);
y £ Y,
whose solution y*'1' is employed to construct a new cone of allowable variations and to test sequentially the directions of the other vectors for belonging to the cone of allowable variations. Here, as in Algorithm (A 1), one can obtain a limit set of solutions y*' ', k = 1,2 ... Note that G. Dantzig was the first to consider the definition of the range of linear coefficient variations, for which the optimal solution is not violated, as applied to the continuous LP problem [4]. Dantzig shows that in solving the continuous LP problem by the simplex method for determining the maximal and the minimal values of linear coefficients, for which the optimal solution is preserved, it is possible to use the analytic relations [4] max c.'i = c° + min I —
\anl
min Cij = c°, + max
,
— I,
atJ > 0;
atJ < 0,
where i is the index of the z-th basis process, and j is the nonbasic index; Cj — Cj
/
j
&ijCji.
This method has the following limitations: 1. The proposed estimations can be used only if the problem is solved by the simplex method, i.e., there are simplex tables at each step. If the continuous LP problem is solved by such methods as the penalty function method or the gradient method, the proposed analytic relations are basically inapplicable. 2. As noted above, application of Dantzig relations requires computer-storage of all the arrays obtained at each step. The result of such nonrational utilization of
202
CHAPTER
9.
OPTIMIZATION
ON FUZZY
SETS
computer storage is t h a t t h e values [c™ax; c™m] can be determined by Dantzig relations only for the problems of a fairly low dimension. 3. Basically, Dantzig relations cannot be used for estimating t h e optimal solution invariance in t h e integer and nonlinear programming problems. 4. As noted in the formulation of the problem, some cases require not only esti mation of the optimal solution stability for a specified range of variations, but also determination of an acceptable solution for t h e entire range of variations. In such cases, Dantzig relations fail to hold. So development of t h e estimation algorithm and m e t h o d for optimal solution in variance, as based on the separability properties of convex cones, is not only necessary, but also permits a significant increase in t h e efficiency of computational procedures.
9.5
Optimization of Integer LP Problems
Integer and diserete linear programming is one of t h e most actively developing spheres in m a t h e m a t i c a l programming which is due to the demands of numerous engineer ing applications. W i t h o u t touching upon all possible fields of application of integer and discrete programming (see [11]), one has to emphasize high efficiency of integer optimization models in solving the problems of computer system adaptation. At the same time it is to be admitted t h a t the integer and discrete linear pro gramming problems represent one of the underinvestigated classes of mathematical programming problems, and the existing algorithms for their solution cannot be con sidered as efficient, and in some cases they prove to be inadequate. T h e formal meth ods for solving LP problems t h a t have enjoyed t h e widest application are Gimiry's methods and branch-and-bound m e t h o d [10]. W i t h o u t touching upon t h e efficiency of these m e t h o d s , we note t h a t t h e necessity, e.g., in Gomory's method, to effect the n-multiple (n being the dimension of the vector of variables) solution of the LPproblem presents obstacles such t h a t even an approximate solution of /LP-problems cannot be obtained. After careful analysis of the formal solution methods for ILP problems F. Gill and W. Murrey concluded t h a t ". . . no significant progress is currently observed in the integer linear p r o g r a m m i n g . . . one can hope to be successful only where heuristic methods are developed" [5, pp. 269, 273]. T h e approximate methods t h a t have enjoyed the widest application for solving ILP problems are the methods of barrier functions and r a n d o m search [5, p. 171]. In some cases, however, these methods are not very effective in applications. Note that the existing solution algorithms (formal and heuristic) of ILP problems do not allow for t h e fact t h a t the control problems reduced to the integer optimization problems can be solved on multiple machine systems, and in t h e process, the solution may require multiple corrections for adapting the system. Since none of the existing methods of solving ILP problems can be used to solve nonuniquely formulated ILP problems, one has to consider t h e possibility of synthesiz ing a unified algorithm which can be utilized for solving t h e uniquely and nonuniquely
9.5. OPTIMIZATION
OF INTEGER LP
PROBLEMS
203
formulated problems. As before, in the solution of nonuniquely formulated continuous LP problems, we shall observe the following procedure. First we consider the solution algorithm of the uniquely formulated ILP problem, then turn to the problem with interval-defined parameters and finally go to the original problem, that is, the ILP problem, where the nonuniquely defined variables are regarded in terms of a linear form and constraints. The proposed solution algorithm of the uniquely formulated ILP problem is based on the heuristic self-organization principle, It is as effective as the simplex method and admits a very simple machine implementation. The heuristic self-organization principle [7] allows for the following procedures: — selection of elementary algorithms for generation of inputs; — selection of heuristic sampling criterions; — selection of a law for generation of combinations. A special feature of the heuristic self-organization principle, which is responsible for its high efficiency, is that the gradient methods, the method of random search, and other methods for solving optimization problems can be used at the elementary algorithm selection stage. Let us formulate the main algorithm heuristics. Suppose max(c, x), x € X = {x\Ax < b,x > 0} ,
x = integer
(9-11)
is the original integer LP problem. Based on (9.11), we formulate a pair of equivalent LP problems without integer requirements: the primal problem m a x ( i . i ) ; x e X = {x\Ax < 6, x > 0} ,
(9.12)
the inverse problem min(c, x); x 6 X = {x\Ax < 6, x > 0}
(9.13) (1
Suppose the solutions of problems (9.12), (9.13) are the vectors x* ' and x*(2', respectively. Bisect the line segment passing through the points x*(1), x*(2). In the direction of x*'1' we obtain a vector set X = {£'*'} ,fc= 1,2,..., ,(1)
if) = - 3
«(*-l)
~X>
+ xf-x); Vz; i = ! > ; x<°> = x*<2)
in the direction of x*(2) we obtain a vector set X = {x ( -' ) |, j = 1,2,... f»'l v
.(j-ll i ~
x
x > = —■
*(2)
x
i
-(j-l)
x-
w
'
1
- (0)
»(1)
; Vz;z = l , n i ' ' = x * '.
2 Bisection is effected until |x*(1) - xf'| < Ax or |x*(2) - x\k)\ < Ax„ where Ax is a descreteness interval, k = 1,2,. . .
CHAPTER 9. OPTIMIZATION
204
ON FUZZY SETS
From the collections X = {i{k)}, k = 1,2,...; X = { i b ) } ; j = 1,2,... we form the collections of integers X I , X2, X3, X4 in which Xl<*> = Entier(z<*)), X2{3) = Entier(iW), X3(i> = Okrugl(xW), X4<3> = Okrugl(i<J>), where Entier(x) is the procedure of forming an integer by dropping its mantissas; Okrugl(s) is the rounding of the number x to the nearest integer. From the collections XI, X2, X3, X4 we eliminate those integer vectors which do not belong to the original set of constraints, i.e., for which we do not separate the condition AX{LO) < b, where w = 1,2,3,4; xl 6 X I , x2 62 X2, x3 € X3, x4 € X4. Then the integer vectors we have obtained can be viewed as feasible solutions of the original ILP problem. We will suppose that implementation of stage II of the heuristic self-organization o(') o(«+i)
principle resulted in selection of vectors x ,x , . . . , that are used for generating combinations (stage III of the heuristic self-organization principle). Let us form twodimensional arrays: o((+l)
>(')
o (0
X =
x
fo(*+l
l,n,
=U.
where each row represents the vector x or x whose components are changed to an integer kp such that the following inclusion holds °(0 o(0
3(0
o(0
V
p> xp+l<
o(01 ■ .*»
eA;
X = {x\Ax < 6, x > 0} The array X
is of the form: ,
o(0
xx X
+Ai
o(0
°(0
>(0
>(0 o(0
e.Y
»(0
+/c2
SA o(0
o(0
(9.14)
,
Xn
GA' •>(0 The number of arrays X , t = 1,2,... corresponds to the number of vectors 3(0
xK~'eX selected by the criterion r<'> > ( 0 , 7 - 0,8)r* For each row of the arrays (9.14), calculate the values of a linear form of the At) . . , , r . o ( 0 , o(0 original ILP problem (11): L(x{ ),L(i 2 ) i M^n )' x i £ A , i is the number ot a (0, {t) o(() (r/° A i u ■. , row. The maximal element Lm(x m a x | L ( i t ) j and its associated row z m are determined in the array of linear forms we have obtained. The computational algorithm scheme for the maximization problem (9.11) is of the form [15]:
9.5. OPTIMIZATION
OF INTEGER
LP
205
PROBLEMS
A2:l. Form the inverse problem which, in our case (with respect to (9.11)), is the continuous LP problem min(c,a;);
x G A' = {x\Ax < b,x > 0}
2. Solve the primal and the inverse LP problem without integer requirements, and fix the solution vectors a;*'1',a;*'2'. 3. Perform the bisection procedure and form the arrays X I , A'2, A3, A4. If A1UA2UA3U A4 / 0 for |a;*(1) - a-*(2)| > 5, the original /LPproblem has an integer solution which can be obtained by the proposed algorithm; otherwise one has to alter the objective function vector c ± w. Go to para 1. 4. Calculate the vector norm x*(1), r* = (V(1>, a**1') ,
oft)
the vectors x
, select from the array .{'
o(*+l)
,x
,. . ., whose norms exceed (0, 7 — 0, 8)r". oft)
o('+l
,
o (0
o(t+l)
oft)
5. 1 he vectors x , a . . . are used to form the arrays X , X , ■ • • |diag X = {x\ ±k,} e A = {x\Ax 0}, where i is the number of a row, and k, is an integer. o ft)
6. In the arrays A
o (t+1
,X
0(()
„((+l)
, ■ ■ ■, determine the rows xm , xm
, . . . , , where the
values of a linear form of the ILP problem are the largest. The vector x m £ o(t + l )
,xm
lxm
1
, . . . , > , which ensures the largest value of a linear form among the vectors
of) o('+i)
x
m i xm
i ■ ■ ■ ii fakes an exact solution of the original ILP problem to be o(a)
xm — max(c, x);
x € X = {x\Ax < b} , where xt are integers. In the meaningful treatment, the algorithm can be integrally represented as fol lows. The solution of the integer LP problem is taken to be an integer vector whose Euclidean norm (length) is as close to that of the solution of a suitable continuous LP problem as possible. The integer vector, whose norm has the least departure from the optimal solution norm of the continuous LP problem, is sought by exhaustion of some finite set of the feasible solutions of the original ILP problem obtained from bisection of the line segment connecting the points x*''),x*' 2 ' that are the solution vectors of the LP problem generated in terms of the original problem. Investigations and machine experiments lend support to the high efficiency of the algorithm and advisability of its application for solving applied problems in a wide range. Based on the solution algorithm of the ILP problem, the invariance conditions for optimal solution can be formulated as follows. The basic algorithm operations, which make it possible to obtain solutions close to the optimal, are calculations of optimal solutions for the primal and the inverse problem of continuous linear programmingmaximization and minimization of a given linear form. The solution of the continuous linear programming problem is regular, i.e., it is located at the node of a polyhedral
206
CHAPTER
9.
OPTIMIZATION
ON FUZZY
SETS
set formed by a constraint system. For this case (regularity of solutions) t h e necessary conditions for an e x t r e m u m in any of the existing forms (Kuhn-Tucker, DubovitskyMilyutin, Boltiansky) are also sufficient. For t h e regular solution, it is possible to determine a dual cone (the cone of allowable variations) t h a t is a convex polyhedral cone whose components are gradients to hyperplanes the intersection of which forms the point of solution. Since t h e necessary e x t r e m u m conditions are satisfied, the inclusion of the vector of a linear form in t h e dual cone of suitable LP problems is valid. Introduce the definitions. D E F I N I T I O N 3. The sequence of vectors X= XI U X2 U X3 U X4 with X± 0 is called the support sequence. D E F I N I T I O N 4. T h e solution of the ILP problem is called invariant in the small if, irrespective of t h e value u in the ILP problem max(c ± w, x);
x € X = {x\Ax
< 6, x > 0}
the support sequence is kept invariant. D E F I N I T I O N 5. T h e solution of the ILP problem max(c ± ID, x);
x £ X = {x\Ax
< b,x > 0}
is called invariant if, irrespective of the value w, in t h e ILP problem there is a unique optimal solution x * ' s \ T h e following theorem is valid. T h e o r e m 9 . 2 . Let max(c, x);
x € A' = {x|Ax < b, x > 0}
be the original ILP problem whose solution is the vector x*' s '. Suppose the following conditions are satisfied: a) the linear form (c, x), is bounded above on the set X; b) there exist the solutions x*' 1 ' and x*' 2 ' of the continuous LP problems max(c, x);
x g X = {x\Ax
< b, x > 0} ;
min(c, x);
x e X = {x\Ax
< b,x > 0} ,
respectively; c) the dual cones D(M{1)) and D(M< 2 ) ) are defined for the solutions x* (1) and x*' 2 ' respectively. In order for the integer solution x*' s ' to exist in the integer linear programming (ILP) problem max(c±iy,x); x G X = {x\Ax < b, x > 0} it is necessary to satisfy the condition for inclusion of t h e vector c±wm QI =
D(Af"')n(-D(M<2')),
the cone
9.5- OPTIMIZATION i.e., c±w
OF INTEGER
IP
PROBLEMS
207
e Qx.
Proof. The solutions of the LP problems a) max(c,i),
x £ X = {x|i4x < b, x > 0}
6) min(c,x),
i £ A ' = {x|Ax < b,x > 0}
are regular. Based on the necessary and sufficient extremum conditions (9.6), this allows one to claim that the fulfillment of the inclusion condition c ± w G Z)(M (1) ), c±w e (--D(A-f(2))j is necessary and sufficient for the continuous LP problems max(c ± w, c),
x € X = {x\Ax < b, x > 0}
min(c±u;,x),
x G A' = {x|/lx < 6, x > 0}
to have the solutions which identically coincide with the solutions of LP problems in a) and b). In the solution vectors of LP problems a) and b) are respectively x*'1' and x*'2', then the fulfillment of the inclusion c ± w £ D(M^) and c±w e (-D(M)W)) or (2 c±w e D(M^ n (-£>(M >)) ensures that in the ILP problem max(c±ui,x);
x G A' = {x\Ax < b, x > 0}
the support sequence X obtained from bisection of the line passing through the points x*'1' and x*'2' is the same as in the ILP problem max(c, x);
x 6 X = {x|i4x < b, x > 0} .
This completes the proof of the theorem. Based on the theorem, one can claim that the fulfillment of the inclusion condition c ± w 6 Qx is necessary and sufficient to ensure that in the ILP problem max(c ± w, x);
x G X = {x\Ax < b, x > 0} ;
the optimal solution is invariant in the small and necessary for the optimal solution to be invariant. Investigations show that the estimation algorithm for invariance of optimal solu tions in the ILP problem is highly effective where the conclusions of Theorem 9.2 are used. Note that the proposed algorithm does not guarantee that the solutions will be obtained for all cases. On the assumption 1 for \x* — x* ' < Ax, V? the algorithm does not work. o
But the algorithm can be complemented and modified. If the support sequence X is o empty, X^- 0 (which do not preclude the solution of the ILP problem), then taking
208
CHAPTER 9. OPTIMIZATION
ON FUZZY SETS
the design vector to be some vector c' 1 ', whose direction does not belong to Qx, allows one to form a pair of LP problems without integer requirements: max(c (1) ,x);
x G X = {x\Ax < b, x > 0} ;
min(c (1) ,:r);
x € X = {x\Ax < b, x > 0} ,
whose solution makes it possible to determine vectors c 0 ' 1 ' and xoi-2\ for which x°l
-
x° \ > Ax,Vi, and thus obtain a new support sequence X which may contain integer vectors. In what follows the basic algorithm is implemented (para 2), the matrix (b) is constructed, and the maximal element is determined (with respect to the linear form of the original problem). Note that such a case is "diagnosed", i.e., revealed with the use of this algorithm. There are other "abnormal" cases. Firstly, the solution of the continuous LP problem max(c, x), x g X = j:r|/l;r < b,x > 0 | is the vector a:*'1' with integer components, and the bisection procedure is inadvisable here. The solution vector of the ILP problem max(c, x), x 6 X = <x\Ax < b,x > 0 | can be determined by rounding off to the nearest integer and dropping mantissas only for solving a suitable continuous LP problem. Secondly, the dual cone D(M' 2 ') represents only one vector, and hence the cone Qx must represent here this vector alone. Although the solution of the ILP problem max(c±u), c), x £ X = \x\Ax Q\ is also the vector x*'1', the solutions of the continuous LP problems m a x ( c ± i » , i ) ; i £ A ' = | X | A T 0[; min(c ± IO, x); x € X = | x | / l x < 6, x > 0 \ yield an empty support sequence. This form of "abnormality" is revealed by the algorithm, because the symptoms of the "abnormality" x* — x*1 < AxjV^x*' 1 ' being an integer vector) are obvious. The following estimation system is introduced to estimate the efficiency of the pro posed algorithms: 1) time required for solving the problems of invariance estimation in terms of the proposed algorithms; 2) accuracy of problem solution; 3) convergence of algorithms; 4) cycling probability. The estimation algorithms for invariance of optimal solutions in LP problems with variations in functional parameters and constraints are highly efficient and ap propriate for estimation of the invariance stability of optimal solutions with respect to perturbations. In the existing solution methods of ILP problems, provision is made for unique specification of model parameters (coefficients of a linear form and constraints). When the parameters are nonuniquely specified (e. g., at intervals), the existing methods al low one to solve the problem by significantly enlarging its dimension. The proposed algorithm makes it possible to estimate the optimal solution invariance without en larging the problem dimension, hence its efficiency is by an order of magnitude higher than that of the familiar algorithms.
Chapter 10 Optimization of Nonlinear Programming Problems with Nonuniquely Defined Variables As discussed earlier, the problems of optimal control in a large (or complex) system, as a rule, are inadequately formalized and must be viewed as an inadequately described model. In what follows, the term "optimal control" is interpreted to mean the optimal control itself and the finite-dimensional optimization problems. This is due to the fact that the problem of optimal control in a continuous system can be thought of as an infinite-dimensional programming problem in an infinite-dimensional space [20]. The Kuhn-Tucker theorem is central in the mathematical programming theory and can be extended to the problems in infinite-dimensional spaces. The design model of the inadequately described nonlinear optimization problem is taken to be the nonlinear programming problem with nonuniquely defined variables. The existing methods of solving the optimization problems with nonuniquely defined variables are very complicated and allow for construction of the fuzzy sets of objective and feasible choices and their intersection. These features do not preclude uncertainty in selecting a solution even though the solution can only be unique by the physical properties of the model. So it is essential to determine those conditions which, when satisfied, allow the nonlinear programming problem with nonuniquely defined variables to have a unique solution, whence follows the optimal solution invariance with respect to nonuniquely defined model parameters. Note that this approach is not necessarily advisable because estimations of appropriateness are decisive in all cases. Investigations show that solving the optimization problems with nonuniquely defined variables yields a class of conditionally ill-defined problems, i.e., the problems can be ill-defined according to Hadamard-Kolmogorov and well-defined according to Tikhonov.1 To solve the ill-defined (unstable) optimization problems, one may use the 'According to Hadamard-Kolmogorov, the problem is ill-defined if a particular concept of the solution x = R(U) in the initial data U makes it possible to satisfy the following conditions: a) for 209
210
CHAPTER
10.
OPTIMIZATION
OF NONLINEAR
PP
efficcient algorithms t h a t are based on the Tikhonov's principle of regularization and make it possible to obtain a solution for a wide class of m a t h e m a t i c a l programming problems. A special feature of regularization algorithms — determination of the normal and the approximate optimal solution — is t h a t t h e optimization procedures must be ef fected m a n y times (or at least reiterated). T h e approximate n a t u r e of iterative pro cedures, which can hardly guarantee the required accuracy for calculations in the medium-dimensional optimization problems, may preclude estimation of the optimal solution invariance when the parameters of the optimization problem are subject to variations. Investigations show t h a t in the convex programming problem there are invariant solutions (/^-solutions) t h a t are independent (to a certain extent) of variations in the functional or constraints. When inadequately described, the estimation problem of the optimal solution invariance amounts to solving one well-defined optimization problem, which permits a significant increase in the efficiency of the algorithm: ac curacy, speed, convergence, etc. Investigations also show t h a t the theorems of Kuhn-Tucker, Dubovitsky-Miliutin, Boltiansky hold t r u e for the problems with nonuniquely defined variables. Theoretical propositions make it possible to offer the m e t h o d s for synthesizing the solution algorithms of the optimal control problems t h a t are inadequately described.
10.1
Optimization of Problems with Nonuniquely Defined Variables and their Solution Methods
As noted before, the posinominal function optimization problem max/i(u>) = £ cx [rijwf'-'] ; w eW
= {w\g(w)
< 0,
(10.1)
w > 0, c > 0}
may have extensive engineering applications. In actual conditions, the coefficients of c, and index ki3 are nonuniquely defined and can be viewed as the nonuniquely defined variables (coefficients) belonging to the fuzzy sets U(C) and U(K), respectively. So, in the general case, we need to consider the optimization problem with nonuniquely defined variables. F-sets provide a means of obtaining the solution of t h e control problem with nonuniquely variables as some F-set of solutions. To this end, one has to determine any element u € U there is a solution x from the space F; b) the solution is uniquely defined; c) the solution is stable. According to Tikhonov, the problem is well-defined if: a) for some class of data. U G m there is a solution x which belongs to a given set M, a; £ M; b) the solution is unique in the class of functions belonging to M\ c) the solution is stable [21].
10.1. OPTIMIZATION
OF PROBLEMS
211
WITH NDV
the existence conditions for the unique solution when the coefficients of the objective function and constraints are nonuniquely specified in the problem maxh(l/(w)j, wG U{W). The conventional optimization problem can be represented in terms of F-sets as a special optimization problem (9.4). When stated traditionally and in terms of F-sets, the optimization problem allows one to determine the conditions under which the identity of solutions can be estimated for both cases. These conditions are usually associated with the F-set support in the original problem. L. Zade [22] investigated the control issue in large systems to show that the mathematic tools, which are based on the generalization principle and are designed for solving traditional optimization problems, can be applied for solving control problems under uncertainty (with nonuniquely defined variables). The direct application of the generalization principle sometimes involves substan tial difficulties. For this reason, one has to make significant preparations which actu ally amount to developing the solution algorithms and methods for ill-defined control problems with due regard for inadiquate description. We shall show that the optimization problem with nonuniquely defined variables (10.1) can be solved by the traditional methods. Additionally, we shall estimate their efficiency. Turning to the optimization problem in the traditional setting and its analog stated in terms of F-sets [16] Z', :
max/(z);
x € X;
Z'J :
m(x) = min{mp(i),mjv(i)} —> Extr
(max);
x £ X,
we note that the main criterion for adequacy of the problems Z' and Z" is taken to be the coincidence of the F-set support m(x) (the set x 6 X\m(x) > 0) with the solution x" of problem Z' This allows one to consider the solutions of problem Z" as determination of the conditions under which the F-set of solutions to this problem will coincide with the support of the set. In addition to operations on F-sets, as defined in Ch. 9, we define for the solution of the formulated problem the following concepts: 1. The support of the fuzzy set {u,} C U, i = 1,2,..., is such that (yu,(rnA(u)) > 0], A C U. The height of the fuzzy set h = supm^(u). 2. We adopt the following notation of a fuzzy subset of A C U n
A = /Ii\ul + . . . + /in|u„ = X ^ ' l " ' = V WKl i= l
where the plus sign denotes the union operation (U). 3. The set of the a-level of the fuzzy subset A is defined as A = U\fj.A(u);
A=
a- A.
212
CHAPTER 10. OPTIMIZATION
OF NONLINEAR
PP
4. The fuzziness increment operator FH is utilized for transforming the ordinary (nonfuzzy) set to the fuzzy set. Let A C U be a fuzzy subset and suppose the operator F acts to produce a fuzzy subset F(A;K)=UiiA(u)-K(u), where K(u) is a set, i. e., the core of the operator FH:
K(u) = F(l/u;K) 5. The nonuniquely defined belonging function is that belonging function for which HA{U), ^B(U) a r e intervals in [0,1], i.e. for a fixed u HA = {aua2}
-» [0,1];
I* = {biM-*
[0,1].
6. The complement to the fuzzy set
A = A' = j
(l-^A\u).
7. The union of the fuzzy sets A and B A + B = A U B = I (JIA{U) A /ifl(u)) \u.
8. The intersection of the fuzzy sets AC\B=
[ (fiA{u)\J
nB(u))\u.
Here (AAmax), (VAmin). 9. The product of the fuzzy sets A, B A B = j fj,A(u) ■ fiB(u)\u. 10. The convex combination of fuzzy subsets {A;}; i = 1, n is the fuzzy set A with the belonging function: A = Ui^.4, + . . . + UnPA„-
11. The generalization principle. Let / be given as the map A" —> V and let Z be a fuzzy subset: Z = Hx\xx +/*2|a;a + . . . + /j n |z n . Based on the generalization principle,
f(Z) =f(Pl\x}+...+
p„|s n ) = /Kil/C*!) + . . . + ^ n | / ( x n ) .
10.1.
OPTIMIZATION
OF PROBLEMS
WITH NDV
213
We will show that problem (10.1) can be reduced to a finite set of traditional op timization problems. To do this, we utilize the Zade-Bellman generalization principle and mathematical induction principle. We will actually consider the cases of interval (nonunique) definition of the vari ables c = {c,}; i = l , n and k and then unite both cases. Suppose there exists a universal set c, € C of the functional coefficients of U(C) in the problem m a x / ( a ) , x € X where accordingly there is a fuzzy subset CM; C^
= j/jt\Cl; c
i=T~^
and the support of the subset has the power of continuum
C(A) = I fj.C{A)\C, c where /i, is the function of belonging of the component C, to the fuzzy subset C ^ ' f 1, Ci — Ac, < c, < c, + Ac,-:
(v»)(/4e>) = ( 0, c, > c, -+- Ac,, c, < c, — Ac,. Note that the belonging function y.\ be represented as
can be nonuniquely defined, while Hc{A) c a n rs
(c)
pew = r>; In view of the last relations, the belonging function fj,p(x), which characterizes the pre-image of the F-set of objectives C for the map / : X —> Y, becomes the function UP and, according to [16], takes the form: fj.p(x,c) = (J.p{x) PI no[Ay, ,
v_ /
l,f(x)>f(x*),ccU(C);
>"p[X'C)-\0,f(x)
and is nonuniquely defined. In view of the above relations, the optimization problem Z", which is an analog of the uniquely defined problem Z', at the level of fuzzy sets (variables), becomes fl(x) = mm{fip(x),iJ.N(x)}
-4 Extr(max).
xeX We shall define the form and structure of the fuzzy set support, fj.{x), and formu late the auxiliary problem of determining the conditions under which the supports of fuzzy sets, fi(x) and fi(x), coincide (this is equivalent to the fact that the solutions of the problems (f(x), U(C) j —>• max, x g X with nonuniquely defined and ordinary variables coincide).
214
CHAPTER 10. OPTIMIZATION
OF NONLINEAR
PP
We shall consider the cases in which the functional coefficients are normalized < co < £»(n»K)) V J : c° = c,-/||c||, where ||c|| is the functional norm, ||c|| = (c^) / Here the belonging function values p.(x) are fuzzy subsets in the interval [0,1], (Vi).
co(min) 1 2
Let us isolate a-levels in fj.p(x,c), fip(x,c) = / / j p (z,c), where /^p is the value o of the belonging function fJ.p{x) at the a-level; (x,c)^{c\v(c) > a } , where v(c) is the degree to which the vector c = {c,}; i = l , n belongs to the fuzzy subset fip(x,c). If the number of a-levels is equal to w, then it is possible to generate a finite set of special problems of the form: /i(a> = min \jj,p (x,c),fj.iv(x)\ i6l,
—> Extr(max),
a = l,w;
1, xeNJ(x)>f(x),
(10.2)
cCU (&">);
/i(z)
0,xeN,f(x)0\ coincides with the solution of the problem f(x) —> max, x g A'. Consequently, if the support of the F-set p,(a'(x) coincides with the solutions on all o-levels, this is a necessary (and sufficient) condition for existence of the general unique solution in problem (10.1) when k is a uniquely defined variable. The assumption that k belongs to the class of nonuniquely defined variables is of no importance. In this case the belonging function fj,p (x,c) in problem (10.2) must be regarded as jiSp (x, c, k) = fip '(x, c) fl fip(k).
10.2
Generalization in Nonuniquely Defined Functional Optimization Problems
Let m a x / ( C , x); x € .V be the original (well-defined) optimization problem. Suppose the objective function is given explicitly. The set of coefficients C is universal, i. e., all the coefficients are uniquely defined variables. We introduce the following definitions. DEFINITION 1. The functional f{C,x) is nonuniquely defined if at least one of the coefficient sets c, £ C is a nonuniquely defined variable or a nonuniquely defined number which can be characterized by the. triple of parameters C, U, R(C, U), where C is the name of the variable; U is the universal set; u € U is the common name of elements of the set U; R(C,U) is a fuzzy subset of the set {/ that is a nonuniquely defined constraint on the value of the variable U specified by C. The assignment equation for c is of the form c = u : R(C).
10.2.
GENERALIZATION
IN NONUNIQUELY
DEFINED PROBLEMS
215
The degree to which this equality is satisfied is called the consistency of the value u with R(C) and is denoted as fi(u) --= VR(C){u),
where HR(C)(U) is the degree to which it belongs to the constraint R(C). EXAMPLE. The nonuniquely defined coefficient c, can be represented as a nonuniquely defined number APPROXIMATELY c and written as c,4
APPROXIMATELY
c ; = 1 / 4 " + U ^]/c\j),
j = 1,2,...,
3
where fj,[3' is the function of belonging of the value cp' to the universal set C = \C{k}V k = 1,2,... We shall consider the formation of a "fuzzy" functional as a result of the action of uncontrolled perturbations. In this case it may be wise to represent the "fuzzy" functional as a result of the action of the fuzziness increment operator on the uniquely defined functional f(x,C). The fuzziness increment operator can be defined according to [3] as follows. DEFINITION 2. Let U = {uk}; k = 1,2,... be a universal set; A C U—a fuzzy subset for whose elements there are the belonging functions /i/i(u,); K(u) — a fuzzy set that is obtained through the action of some operator on a singleton 1/u : K{u) = F(l/u; A'). Then the action of the fuzziness increment operator on the fuzzy subset A C V may result in a. fuzzy subset F(A; K) that is defined as F(A; K) = HA{U) ■ A"(u), where (J,A(U) • K(u) is the product of the number (J.A{U) and the fuzzy set K(u) for which A = [ a ■ fi^(u)\u holds when \a\a ■ s u p / ^ a ) < 1, Va\. Suppose the result of the action of the fuzziness increment operator is that the set of functional coefficients f(x,C) (or at least one of the coefficients) becomes the set of nonuniquely defined numbers APPROXIMATELY C. We obtain the nonuniquely defined functional optimization problem max/(r,C4
APPROXIMATELY
c) ;
x £ X.
(10.3)
The presence of the functional as in (10.3) necessitates the solution of two prob lems: — definition of the conditions under which the solutions of the original and the perturbed problem (10.3) coincide; invariance of the optimization problem with re spect to perturbation; — development of the general solution method for problems of the form (10.3). We shall examine the possibility of solving problem (10.3) by the traditional me thod. In [3], the generalization principle is formulated to transfer the main properties of the universal (uniquely defined) mappings or relations on some universal region U(C) to the nonuniquely defined mappings or relations whose domain of definition can be nonuniquely defined, U(/J,,\U,). For the mapping problems, the generalization principle can be represented as follows. Let / be the map u -* v, A a fuzzy subset, and A = U(/J.,\ut). According to [3], f(A) = U(ni\f(ui)). The last relation is called the generalization principle.
216
CHAPTER
10.
OPTIMIZATION
OF NONLINEAR
PP
For the optimization problems, t h e generalization principle can be t r e a t e d at the level of mappings as prescribed by the procedure in [16]. But the efficiency of this approach is not very high. This is due to the fact t h a t in some cases the tradition al optimization problem treated in terms of F-sets does not allow one to obtain a unique solution even though it exists. Since the solution m e t h o d s for traditional op timization problems are adequately developed, it is advisable to determine at least those conditions under which the traditional methods can be applied to solving the problems with nonuniquely defined variables. By analogy, t h e possibility of solving the nonuniquely defined functional optimization problem with t h e use of traditional methods will be called the generalization principle. In the solution of the LP problems with the coefficients of a linear form specified at intervals ( C h . 9 ) , it was shown t h a t in order to e s t i m a t e t h e optimal solution invariance or construct the limit solution set (if the solution is not invariant) it suffices to run In tests for inclusions of vectors (rows of matrix C = C\ U C2) in the allowable variations cone Qx. Using t h e concept of near-gradient, 2 one m a y obtain another form (condition) of t h e optimal solution invariance in the LP problem with nonuniquely defined (interval-defined) coefficients of a linear form. In order to define near-gradients in the LP problem with nonuniquely defined coefficients of a linear form, one has to take into account t h e fact t h a t t h e rows of matrix C = C\ U C2 are. gradients of a linear form. We shall determine the limit — maximal and minimal — elements of the vector set generated by the method given above. Let us normalize t h e vectors c: ^=CW/(cW,CW)1/! and define the direction array of cosines between the normalized vector c of the center of gravity of t h e vector set C = C\ U C 2 and t h e vector c' 3 ' S C; j = l , 2 n : 6;=(c,c(i));
i=3T25T
(10.4)
In the array {bj}; j = l,2n, we define the maximal and t h e minimal elements, bz = max{6j}, bv = min(&,), and their associated vectors c C C, c C C. Each of the vectors c' J ' C C; j = l , 2 n has the property CWgA'J,
bv>k\
CWeAp,
k>bf,
(10.5)
where KQ, KQ are circular cones whose generatrices form the respective angles arc6„, arc&z with the vector c. 2 The near-gradient of function f[%) at the point x is the vector that is the limit point of some gradient sequence V / ( n ) , V / ( s j ) , . . . , V/(x ft ), where {xk}f=l is a sequence of points converging to x such that f(x) is differentiable at all points of this sequence [15].
10.2. GENERALIZATION
IN NONUNIQUELY
217
DEFINED PROBLEMS
Since the vectors c? C C are taken to mean the gradients, then by the definition of the near-gradient [15], the vectors c and c are the limit points of the gradient sequence, and hence near-gradients. So the nonunique definition of the functional leads to a nonuniquely defined near-gradient and, accordingly, to the optimization problem m a x { ( c J , i ) , ( c v , i ) } ; x G A"; X = {x\Ax0}
[
>3)
If x* is a solution of the problem max(x' z ', x); x € X, then the condition that any one of the problems max(c' J ', x); x G A'; j = 1,2,. . .; X = {x\Ax < b, x > 0} will have a solution as the vector x* can be written in terms of Kuhn-Tucker theorem as follows: if c[v}eD(M) then max(c[j],i) = i". (10.7) Vj; j = 1,2T where D(M) is the dual cone which is made up by the vectors of hyperplane normals generating at x* a bundle (roof). If this proposition is not satisfied, then the basic algorithm can be utilized for solving the problem of determining the limit solution set. After solving the problem and constructing the cone D(M)1, each of the vectors c?G C; j' = 1, 2n is successively tested for belonging to the cone D(M)1 The first vector c"', whose direction does not belong to the cone D(M), is utilized for formulating the problem max(c>,x); x G A; X = {x\Ax < b, x > 0}
(10.8)
After solving problem (10.8), determining the solution vector x' 1 ' and construct ing the cone D(M)2, one continues testing the vector directions c " + 1 ' , c^t+2\ . . . for belonging to the cone D(M)2 This procedure is carried out until all the vectors c' j) G C\ j = l,2n are surveyed. Suppose the total number of problems arising from the solution of the LP problem with nonuniquely defined coefficients is equal to w, i.e., there exist vectors x*' 1 ',. . ., a:*'"'. The vector x, for which r — \x — x*'!M —> min; ! = l,w can be taken as a solution of the original LP problem. In the general case, one can state that the solution cone Rx = (x* ( ''|; i = l,u includes all possible solutions x**c), x*(c' G Rx of the original LP problem which could not occur under various combinations of components c (j) that did not appear in C = Cx U C2. Note that suitable belonging functions can be defined for the components of the solution cone Rx. We shall exam ine the ways of constructing near-gradients for the nonlinear objective function which contains nonuniquely defined coefficients. On the assumption that the functional / ( x , C) is well-defined and convex, we con struct its linear approximation at the point x (s) : / ( s ) ( x , C ) = / ( x ( s ) , C ) + + V / ( x ' 5 ' , C)(x — x^'); x' 5 ' G X. The fuzziness increment operator is forced to act
218
CHAPTER 10. OPTIMIZATION
upon the functional ps\x,C).
PP
Similarly to (10.3), obtain the optimization problem
s
max/< >(x,[/(C)) = / ( x < s ) , C 4 C +V/(x's>,CA
OF NONLINEAR
APPROXIMATELY
APPROXIMATELY x a
( i )
C)(x-x^);
C) + (10.9)
,
where C is a fuzzy set of coefficients which is distinguished from the set C by the scale coefficient (c, = c, ■ M, M being the scale determined by differentiation, c, G C, c, e C ) . Problem (10.9) is a linear programming problem, where the objective function coefficients are nonuniquely defined. Similarly to (10.4)—(10.6), one may define neargradients V/j 2 '(a:, C), V / | ^ ( s , C). If the numeric of iterations is equal to p, then it is possible to construct the cone of near-gradients:
K» =u {v/((;»(x, o,v/((;»(x,o];
S = YTP.
Assumption 10.1. Suppose there exists a convex optimization problem max f(x); x £ A'. The action of the fuzziness increment operator, under which at least one of the functional coefficients becomes a nonuniquely defined number or a nonuniquely defined variable, transfers the original problem to the class of vector optimization problems of almost differentiable functions. The methods of generalized gradient descent (GGD methods) are widely used to solve the optimization problems of almost differentiable functions. But here they are very difficult to use. This is due to the fact that the GGD methods require unique specification of the generalized gradient; otherwise there is no algorithm which provides a means of constructing a vector V / | ( x , C) [15] such that for any e > 0
min|V£-V/,(*,C)|<e; v/.(*,C)e A'f. If the problem is approximately solved, one may construct a vector V/ X (x, C) for which, with the positive e specified, there exists a point x belonging to the neighbor hood of X, x 6 U(X) such that min|V/;(x,C)-V/.s(x,C)|<e;
V/,(x,C)e A'f Additionally, we require the conditions to be satisfied in such a way that these relations would hold for Vf$(x,C) = V/s(~">, V/ S (.r, C) = V / s \ Then we assume that this is a necessary (and sufficient) condition for x to be a solution of the nonuniquely defined functional optimization problem. If Vf°(x,c) cannot be constructed, then solve a pair of problems by taking V/ S (x, c) to be the functionals whose gradients are respectively V / j 1 ' , V/}">. So the proposed formulation of the optimization problems with nonuniquely de fined variables allows for the solution by traditional methods.
10.3. OPTIMIZATION
10.3
OF NONLINEAR
CONVEX FUNCTIONS
219
Optimization of Nonlinear Convex Functions with Nonuniquely Denned Coefficients
We shall consider the "original" convex optimization problem min/(A',C4C
APPROXIMATELY
x 6 X = {x\gi(x) > 0,
C) ;
x>0},
where the objective function (functional) is explicitly given and the coefficients are nonuniquely defined variables: Cij)A
APPROXIMATELY
CU) = U
(nU)\Cij))
We shall define the conditions under which the solution of the original problem is invariant with respect to the fuzzy set of coefficients t/[/i' J ''|C' J 'J. If the invariance conditions do not hold for the given fuzzy set of coefficients, then one has to define some fuzzy subset [/(/j' J '|Cj for which the invariance conditions hold. The statement of the problem can be modified as follows. We have the convex optimization problem with nonuniquely defined variables: Z_:
min/i(z);
x £ X.
The influence of perturbations leads to the problem: Z:
mmh(x,K{X;W));
x e X,
where K(x\ w) is the fuzziness increment operator. What kind of structure must the operator K(x;w) have to make these solutions coincide? The reverse statement of the problem seems natural, i. e., we have the nonuniquely defined functional optimixation problem. Then one has to estimate the existence of a fuzziness reduction operator such that its action will transfer the otiginal problem to the class of problems with uniquely defined variables, where the solution of the latter coincides with that of the original problem. Suppose the solution x' = {x*}; i = L/ri of the problem min/(x); x 6 A' = {x\gt(x) > 0,x > 0}; j = l,m is regular, i.e., it is located at the node of the convex polyhedral set X, (f(x),gj(x)) are convex differentiable functionals. On the set X it is possible to select a subset M C X of constraints g,(x) < 0; j = l,t> which form a bundle (roof) at the point a-': (Vj)ff,(x-) = 0 j = M -
(10.10)
Based on the Kuhn-Tucker theorem, the dual cone D(M) is defined as D(M) = {V(7,-(x*)}; j = l,v and has the property that the set of problems | m i n / ( p , ( x ) ; x € X\ has the general (regular) solution if the vector directions V/' p '(x*) belong to the
CHAPTER 10. OPTIMIZATION
220
OF NONLINEAR
PP
dual cone D(M). This condition can be written in terms of ALGOL operators as follows if V / W ( / ) £ D(M) then min/<">(*); x € X;
(V/M(«)),
(1
p = l,2,...
°'n)
If the set of vectors V/' p '(a:) is viewed, e. g., as a result of action of the fuzziness in crement operator, and if a subset of near-gradients < V/'"'(x*), V / ' J ' ( . T * ) | is selected, then the fulfillment of the condition A"™ C -D(Af), where Kf = = { V/'"'(x*), V/W(x*)j, is a necessary (and sufficient) condition for the nonuniquely defined functional optimization problem to have a unique solution. The fulfillment of the condition for inclusion of the near-gradient cone A'™ in the cone D(M) (the allowable variations cone) makes it possible to select in the fuzzy set of generalized gradients their fuzzy subset which has a unique solution if for the original fuzzy subset of gradients there is no unique solution. We will consider the estimation algorithms for existence of a unique solution in the nonuniquely functional optimization problem. Suppose the original problem can be solved by linear approximating programming methods [6]. Then we have the set of problems min / « (x, U(C)) = f (xW) + V / (xW>) (x - xW) ; XW = [x\g3 (jW) =
9i
x 6 X^
:
(*(/)) + V f t (*W) (x - i W ) > 0} ,
(10 12)
where z'^' is the vector in the neighborhood of which the original problem is linearized: / = 0,1, 2 , . . . If in order to select a vector x^+1' one may utilize the relation x " + 1 ' = where / is the number of a step, /'■" is the step whose value lies within x(f) _|_ lU)S''\ 0 —/max, and S^> is the vector of a feasible direction, then problem (10.12) transforms to the problem / (*<") + Z(»V/ (xW) 5 ( / ) -> Extr(min); S{!) = {*\gj {xu)) the sequence -jx^H
x, s G S<»;
+ / ( / ) V 5 t (*(/>) S<" > 0} ,
is convergent and forms the solution x*, j z ' ^ ' }
(10.13) —> s*,
/ = 0 , 1 , 2 , . . . The solution S*^\ f = 1,2,... of problem (10.13) is regular, which offers a means of constructing the cone of allowable variations Qx = n£*(M'- f '), / = 1,2,...,, where D(M^) = {Vg,{x^)], j C // is a special cone of allowable variation which is determined at each step; Ij is the number of hyperplanes which form a bundle at the /-th step. Since (10.13) is the convex optimization problem, then for each vector Su> it is possible to construct a special cone of near-gradients A'f = {V/W(x^)),V/M(a:W)} and a common cone: Kr = UfK»,
/=1,2,...
(10.14)
Test the condition K* £ Qx
(10.15)
10.3.
OPTIMIZATION
OF NONLINEAR
CONVEX
FUNCTIONS
221
for t h e entire fuzzy set of coefficients. T h e fulfillment of (10.15) shows t h a t in the nonuniquely defined functional optimization problem t h e r e exists a unique solution. Otherwise one has to reduce the fuzzy set of coefficients and restrict oneself to some value of t h e a-level for which a unique solution can be ensured (the a-level is a fuzzy subset for which t h e belonging functions /i(x) > ad). For t h e fuzzy set, which contains the belonging functions /J, > ad, one may define the original problem and t h e cones Qx and Kx for which the above algorithm can be utilized sequentially. So we have the vector optimization problem of functionals. Note t h a t in some cases it is advisable to verify the condition (10.6) at each step of the regular solution. Justification of the algorithm. As noted before, linearization of t h e original math ematical programming problem, where a regular solution is required at each step, leads to a set of problems (10.12), (10.13). T h e algorithm will be implemented in compliance with the operator scheme t h a t follows: A1A2A3P^ | ASA6A* t As, (10.16) where A\—the procedure of defining a feasible point x ' p ' in the original problem m i n / ( a : ) ; x£ X(p = 0); A2 — generation of t h e approximating LP problem (10.13) for determining a fea sible direction s'p' at t h e point x ' p ' ; A3 — solution of problem (10.13) and definition of vector 5 ' p ) P 4 —verification: V/(a- (p >)S (p > < 0; A5 — determination of the maximal step length / = maxj/|z' p > + / ■ S^l
£ A'
(P 4 = t r u e ) ; A6 — determination of t h e designed step line 0 < / ( p ' < / for which / ( z ( p ) ) + / (p) (p) 5 -► min;
A7~p
= p + 1;
A8—termination of the optimization search, storage of vectors x^\ S ' p ' , compu p tation of values / ( x ' ' ) ( P 4 = false). This scheme allows one to examine t h e definition of 5 ' p ' t h a t is a feasible direction at the point i ( p ) . By the regularity of the solution 5 ( p ) , a special cone D(NM) can be constructed as follows: D (M< p )) = {v9l(xM)}
;
j = TT^,
where vp is t h e n u m b e r of hyperplanes forming a bundle at the p-th step. T h e following proposition holds for t h e cone J D ( M ' P ' ) : if
V/< a >(z ( p ) ) € £>(M) (V/(3)(z)),
then
min (\7/<9>(.r• s^)
=l,2,...;
V
sw6Jl'l.
= 5* ( p '; '
(10.17)
T h e validity of (10.17) follows from the fact t h a t for regular solutions the nec essary e x t r e m u m conditions are also sufficient. T h e theorems proved below make
222
CHAPTER
10. OPTIMIZATION
OF NONLINEAE,
PP
it possible to formulate t h e existence conditions of a unique solution for t h e set of convex nonlinear programming problems. T h e o r e m 1 0 . 1 . Let f(x),gi(x) be bounded convex differentiable functions. The optim.iza.tion problem is min/(z);
x C X = {x\g,(x)
and the following assertions a) the sequence \x^p'\ from the relation:
x'
p+1
are true: is convergent p
p
j = l~m
> 0,x > 0} ;
and each of its elements
P
p
' = x ' ' + /' ' • ,S"' ', where 5 * ' ' is a solution Vf
(10.18)
is of the
determined problem:
(x ( p ) ) S{p} -> Extr ( m m ) ; s(p) G
5>b);
S{p) = { x | V 5 i (x ( p ) ) ■ S ( p ) > 0, x > 0 } ;
j = T77^;
b) t i e size of the step /' p ' is determined by the algorithm c) a, special cone of allowable variations is defined as D (M) = {V;(>>)} ;
(10.16);
3= 1^,
wiiere u p is t i e number of hyperplanes Vg,(x^) ■ 5 ' p ' = 0 forming a bundle for the ip solution S* l Then the special cones of allowable variations designated with numbers k and fc + 1 are related by the inclusion D (M
w
) D D (M(*+1))
(10.19)
Proof. By t h e regularity, the solutions for t h e steps k and k + 1 hold true
v / (iW) e {v5l(x<*>)}; j e 4; V/(x<^>)e{v 5 l (x<^>)} ;
J=
/, +1 ,
where ij., Zjt+j — a n index set of t h e hyperplanes V
(10.20) Vg,(x^k+1^)
If V / ( x ) is taken to be a linear functional, then since x'*"1"1' = £(*) • S' fc ' + x' fc ', one m a y obtain from (10.20) t h e pair of conditions
v/(x^)) e {v 5l (x< fc )),v/<*)(x)} ; V/(x( f c ») e {V 5 l (x<^)),-V/^(x)} ; where &fW(x)
= Vf{l^
■ S^).
,cii JC/HI,
10.3. OPTIMIZATION
OF NONLINEAR
223
CONVEX FUNCTIONS
The condition (10.20) can be written as V/(iW)£D(MW)u(A/W(,));
v/(x«) eo(MW))u(A/«w). Comparing the relations V / (*<*>) C D (M<*>)
and
A / (x(fc)) g Z) (A/(fc+1>) U ( - A / « ( x ) ) ,
on the one hand, and V/(x< f c + ») GD(M< t + 1 >)
and
V/(x< A + ») 6 D ( M « ) U ( A / W ( r ) ) ,
on the other, one may claim that this can hold only for D{MW) D Z)(M ( i + I ) ). If the functional / ( x ) is not linear, then by the theorem of inclusion [12], in terms of which V ( / i + f2 + ... + fm)(x) € Vfi(x) + Vf2{x) + ... + A / m ( x ) , one may obtain the following inclusion: V / (x* +1 ) € V / ( a « ) + V / (f<*> • S^)
.
(10.21)
It follows that if the condition £»(/W<';+1') C D(M^) holds for the linear func tional V / ( x ) , this is true for any smooth differentiable functional. Corollary 10.1. Suppose the problem Z', :
min/ifx);
x € X;
X = {x\9l(x)>0;x>0,j
=
l,2,...,m}
has a solution x*, < x' ' > —*• x*, where a; is the total number of iterations required to obtain a solution within a specified accuracy. For the problem Z" : min/ 2 (x); x E X to have a solution x*, it is necessary (and almost sufficient) that the vector direction V/ 2 (x*) should belong to the cone D(M^). The validity of this assertion follows from the condition that if the required accura cy 8 of extremum computations for problem Zg (the condition V / ( x ^ ' s * ^ ' — S\ > 0) is achieved in u> iterations, then the cone Qx =C\ D(M^h^), k = 1,2,.. . can be taken k
to be the cone £>(M M ) because Qx D D(M(u,)). Note that the above relations are valid where the original problem is solved by the linear approximating programming methods, because here each element of the sequence j x ^ ' \ —> x* is a regular solution of some convex programming problem. Furthermore, it should be noted that the initial point x' p ' must be selected to be a feasible (for the original problem) point and the constraint system must be
224
CHAPTER 10. OPTIMIZATION
OF NONLINEAR
PP
represented by inequalities alone. For the constraints of the form gj{x) = 0, the latter must be written as two equivalent inequalities: 5j(x)<0,
- & ( * ) < 0.
REMARK. In some cases it is advisable to obtain the solution x by employing the methods which preclude a regular solution at each step of the iteration procedure. These are the quadratic approximating programming, the penalty function methods, the generalized gradient descent method, etc. In order to obtain the cone Z)(M' P '), one has to determine a vector x* = x*±x such that expansion of the original problem into a Taylor series in the neighborhood of x* yields a linear programming problem with its solution as the vector x' ± 5x. The efficient algorithms for determination of the vector x* can be obtained by employing the penalty function methods. For actual estimations, it sometimes suffices to take the vector x* as the (w — 1) — e approximation of vector x* [u> being the total number of iterations).
10.4
Solution Algorithms for Nonlinear Programming Problems with Nonuniquely Defined Variables
F. Gill and W. Murrey [4] pointed out that although there are many various mathe matical methods, the issue of solution algorithms for nonlinear programming prob lems is far from being satisfactorily resolved. None of the rigorous formal methods for solving nonlinear programming problems can generally quarantee that the glob al extremum will be obtained. As noted by F. Gill and W. Murrey, those who are concerned about development of the algorithms and methods for solving nonlinear optimization problems should concentrate their attention on heuristic methods. The heuristic self-organization principle, as formulated in A. G. Ivakhnenko [7], has recently found wide application in the solution of such control problems as prediction, identification, and adaptation of systems. The heuristic self-organization is based on the nonfinal decision principle which was initially set forth by F. Rosenblatt in the analysis of perceptrons i.e. the devices designed to identify images. The nonfinal decision principle can be illustrated as follows. Suppose the investi gator is going to make a decision. This decision can be made either immediately or in several steps. In the latter case, it is to be successively improved (or at least not wors ened). According to this principle, at each step one has to select not a unique (the best) decision (e.g., in the statistical decision theory it is the best decision which is always selected), but a set of decisions, about 20-35% of the most realistic decisions. The nonfinal decision principle provides a means for multiple application of heuristic criteria in selecting decisions at each step, where both the decisions and the criteria require correction. Proceeding in this manner makes it possible to obtain decisions within a high accuracy.
10.4. SOLUTION ALGORITHMS
FOR NPP
225
It is essential that the heuristic self-organization method dealing with integral effects permits a reduction in the initial description which involves not a decrease, but an increase in the accuracy of problem solutions. Based on the above proposi tions for the heuristic self-organization principle, we shall consider the possibility of synthesizing the solution algorithm for the mathematical programming problem with nonuniquely defined variables. The core of the algorithm includes three steps: 1) definition of an integer solution of some uniquely defined problem which is di rectly related to the original problem; 2) estimation of the optimal solution invariance and definition of a unique solution; 3) definition of a limit solution set. Following the direct methods, we first consider the solution algorithm of the non linear programing problem, where the heuristic self-organization principle is used, assuming that the coefficients are all uniquely defined. As applied to the solution of this class of problems, the heuristic self-organization principle allows for implemen tation of the following procedures: 1. Select elementary algorithms for input generation: 1.1. Formulate the problem with nonuniquely defined coefficients that have been arbitrarily selected from the fuzzy set of coefficients. 1.2. Construct the linear approximation of the problem formulated in 10.1. 1.3. Formulate the primal and the inverse LP problem as in (9.12), (9.13). 1.4. Define an approximate integer solution of the uniquely defined problem in 10.1 (Algorithm A2). The integer solution is utilized as the starting point (an input) for defining a continuous solution of the uniquely defined problem. 2. Select the generation law of combinations: 2.1. Generate the sequence of vectors (9.14) each component of which is of the form x* ± /;;; i — l,n, where x* is an integer solution of the approximating LP problem. 2.2. Compute the functional value f(x) of the original problem at the points [x^\ . .., z ^ } (14), /(x ( j , ) ); j = T7ZZ7. In the array {/(z ( j ) )); j = T^J, define the maximal element / m ( x ( j ' = m a x j / ^ ^ ' ) } ; j = l,w and a suitable vector z<7n) 2.3. All the vectors x^K j = 1,2,... for which the values / ( x ' J ' ) exceed (0,7 — 0,8)/(a:' m '), are taken to be estimated for the next step. 2.4. Procedure 2.2. is carried out for the vectors x' J '; j = 1,2,... selected in 2.3. The termination criterion for generation of combinations is the condition that at the p-th step one has obtained the maximum value / ( x ( j ' ) which does not exceed a similar value at the step (p — 1). After solving the approximated problems, test the optimal solution for invariance or define the limit solution set (if the solution is not invariant). We shall consider the computational schemes of the algorithm. The original problem is m a x / (U(C).x); x € X; K ' X = {x\gdx) < 0,x > 0 } ; ;' = l , m .
(10.22)
CHAPTER 10. OPTIMIZATION
226
OF NONLINEAR
PP
where U(C) is the set of nonuniquely denned coefficients of the functional which is explicitly specified c,4
APPROXIMATELY
c=Uf^)/c^\
Vi,j =
1,2,...
gi(x) are convex functions. For some fixed value //'■", we formulate the uniquely defined problem max Jf(c, x): x € X; ' X = {x\g3(x) < 0,x > 0 } ;
j-\,m.
(10.23)
In the neighborhood of some arbitrary point x"', we perform linear approximation of the functionals (10.1): f(x) £ / (xW) + V / (xW>) (x - x<") ; gt(x) = , (*(«) + Vft (x ( / ) ) (x - . W ) ;
(10.24)
where V = —. ax In view of (10.23) and (10.24), set up the primal and the inverse approximating LP problem:
min/(x);
max fix)', x G X\ . ] _ xEX; X = {x\g,(x) < 0} ;
(10.25) j = 1,772.
Solve the pair of problems (10.25); the solutions are a;*'1' and T*' 2 ', respectively. Apply Algorithm A2 in order to determine the vector with integer components on the set X = {x\g,{x) < 0,z > 0J. The following assumption is based on the fact that the sets X and X are convex and related by the inclusion X C A'. ASSUMPTION 10.2. Suppose xW = {x{/]}, i = I ^ i s the vector in the neigh borhood of which the pair of the approximating LP problems (10.25) is constructed. If the set A' = {x\gi(x) = 5,(a;(/)) + Vgi(xW)(x - xW>) < 0; x > o} is closed and bounded above and below, then, regardless of the vector x"> selected, application of Algorithm A2 allows one to obtain the vectors with integer components that are taken to be the initial approximation of the solution of problem (10.22) (heuristic self-organization inputs). ASSUMPTION 10.3. If the set A' ={x\g,{x) < 0,x > o} contains at least one integer point, then application of Algorithm A2 allows one to determine this point if the vectors i*' 1 ' and x*'2' are taken to be the solution vectors of the approximating LP-problems. After defining the vector x'W that is an integer solution of problem (10.22), it is taken to be the initial point, and application of the random search methods allows one to define the vector x* that is a solution of the original problem.
10.4. SOLUTION ALGORITHMS
FOR NPP
227
The regularity properties of the solutions of problems (10.25) are utilized to es timate the optimal solution invariance. This estimation can be formulated as fol lows. If in the nonlinear programming problem (10.22) the functional coefficients are nonuniquely defined variables and the following conditions are satisfied: a) for the primal and the inverse approximating LP problem (10.25) there exist the solutions x' 1 ' and a;*'2', respectively; b) the dual cones D{Mx) and D(M2) are determined for the solutions x*(1) and x*2); c) the limit matrices (6), (7) and the cone A'+ are constructed, then the solution of the original problem (10.22) will be invariant with respect to nonuniquely defined coefficients subject to K+ £QX =
D{MX)C\{-D{M2))}
or C(k) € Qx,
V;'; k = T/2^.
If the solution is not invariant, then define the limit solution set. The algorithm for constructing the limit solution set is as above. Numerous experiments dealing with the solution of applied problems demonstrat ed high efficiency of algorithms. We shall now estimate the computational accuracy of "limit" solutions in the linear programming problems, where the coefficients of a linear form are defined at intervals. In the LP problem max(c,x); x € X = jx|Ax < b, x > 0 | , with the intervaldefined coefficients of a linear form C,mln < C; < CJmax, Vz, the assumption is made that the coefficients can assume only the maximum or the minimum value. Then 2" linear forms are respectively 2" LP problems that can be constructed in the general case. If the nodes of the polyhedral set .Y = jx|/lx < 6.x > 0[ are assumed to be densely "packed", then the number of solutions of the set of LP problems is equal to 2n The limit matrices (9.8), (9.9) provide a means of forming 2n LP problems. In vestigations show that such a replacement is allowable. The reference problem was considered to be the problem max(c, x);
c £ X;
X = |x|j](a! + l/2-xI)1/2<(a! + l/2-2t),x
>o|
The set X represents a multidimensional hyperball with a suitable hypersphere of radius r, r = fe(a; + 1/2)2 - 2b)
It is suggested that the hypefball is formed a x
by infinitely many linear constraints Yl ij ]
< b}, j = l,m. Variations of the t-tb
i
linear form are considered within ± 5 — 150%. The criterion of estimation is selected
228
CHAPTER 10. OPTIMIZATION
to be the value S% = \N(x°l) - N{x°2)\, where N(x°l) £i = £ x°^/2n;
U]
OF NONLINEAR
= (x,x);
PP
N(x°2) = {x,x)\
n
j = 172^; x, = £ x° /2 ; j = L/2™; x°^ is the solution vector of
the LP problem max(c"', x); x 6 X. Investigations show that the relation Sfj(AC) (AC represent variations of the functional) in the range of variation accounting for ±15 — 75% is within 0,2 — 0,8% and actually does not depend on the dimension of the problem being solved. The dependence 5jf(n, Ac = const) for n = 2,6,8,15 with Ac±150% is within 0,8 — 1,5%. In applications, this estimation seems to be satisfactory. The issues of control in large and complex systems-economic, social, engineering — are the currently central problems. A special feature of large and complex systems is that incomplete information at all control levels may preclude mathematically rigorous description and, more often than not, lead to erroneous results. Admittedly, the problem of control in large and complex systems (including the optimal control problem) must be viewed as insufficiently formalized, and this has for-reaching implications. The inadequately formalized problems (which are generally classed with the math ematically ill-defined problems) cannot be solved only on the basis of rigorous meth ods. Here, of course, one has to develop essentially new methods that are based on formal heuristic premises. Furthermore, the rational areas of application of rigorous and heuristic methods and their rational combinations have to be clearly outlined. Literatures have recently focused on development of the algorithms and methods for solving inadequately formalized control problems, including those of mathematical programming. Methods for solving these problems are likely to be found at the interface of such sciences as cybernetics, biology, and physiology. Of course, this way is not going to be a smooth one. We examined some special ways of solving inadequately formalized optimization problems with a view to application in large and complex systems.
Chapter 11 Optimization Problems in Function Spaces This chapter concentrates on the subject, stochastic optimization problems, which derives its source from two disciplines — theory of probability and mathematical pro gramming. However, it is far from being worthwhile to describe stochastic optimiza tion problems in terms of classical probability theory: such approach and statement of problem make it hard to analyze. In this case it may be wise to take a more gener al view of things replacing the original statement of the problem being discussed by its analog expressed in terms of function spaces. Thus, the study of some issues of modeling complex systems under uncertainty requires a familiarity with basic notions and results in topology and functional analysis. To make reading of this chapter more convenient, basic notions and results are summarized in sections 1-4. We shall not restrict our consideration to those insights without which the main body of this book may become unreadable. These sections 1-4 are intended for the reader who has not mastered the corresponding mathematics and whose perception of the text may be hampered by fragmentary presentation of the subject matter. Sections 5 and 6 deal with linear optimization problems and matrix games in the ordered spaces. These sections fit in well with the book because the space of random variables which is naturally linearized and ordered is the vector lattice and some of the stochastic problems are conveniently defined in terms of ordered spaces. This permits the use of the highly developed techniques of ordered spaces in inves tigations of stochastic optimization problems. The approach adopted here is a great opportunity, and its fruitfulness can be exemplified by the problems studied. Note that the statements discussed at the end of this chapter are rather general and involve the familiar statements of parametric problems.
11.1
Topological Spaces
The set E with some collection r of its subsets is called topological space if for any G\,Gi E T the set G\ f] G2 6 r;
$>$
230
CHAPTER 11. OPTIMIZATION
PROBLEMS
IN FUNCTION
SPACES
for any Ga G r, a G J (J is an arbitrary set of indices) \JGa G r; an empty set and the set S are contained in r. The set r is called topology, while its elements are open seis of the topological space. It is clear that various topologies can be introduced in the same set 5; therefore a topological space is defined by a pair ( E , T ) , where topology r is fixed. In view of this stipulation, however, we shall denote the topological space ( S , r ) by one letter E. The sets S \ G which are complementary to the open sets are called the closed sets of the topological space H. The closed set family ^ may satisfy the following conditions: for any VUV2 € $ the set Vx \}Vi € * ; for any Va G * , a G J the set f]Va £ 9; 0,EG*.
The family ^ may at times be called the closed topology of the topological space E. Evidently, topological space can also be defined in terms of the notion of a closed set, i.e. by the closed topology. The neighborhood of the point x in the topological space H (respectively the neigh borhood of the set M C E) is referred to as any open set G c H containing the point x (the set M). The point x is called the point of tangency of the set M C S if each neighborhood of this point contains at least one point of the set M. The set of all points of tangency of the set M in the topological space £ is called the closure of the set M and is denoted by M. The set M is closed if and only if it coincides with its closure. M is the smallest (in the closure) closed set containing M. The main properties of the closure operation are: 1. M C M; 2. M\JN
= l U l ;
3. M = Af; 4. 0 = 0. The point x G S is called the /imii pomi of the set M if each neighborhood of x contains at least one point of M other than x. The point x of the set Af C H is called the interior point of this set if there exists a neighborhood of the point x which is contained in M. The collection of all interior points of the set M is called its open kernel or the interior and is designated int M. The point x is called isolated in E! if the set consisting of this one point is open in H. The sequence of points {a-1}, i = 1,2,... in the. topological space S is called convergent to the point x G E! if each neighborhood of the point x contains all points of the sequence {x1} starting with some point. In general, the set which contains the limits of all convergent sequences of its elements does not have to be closed. For this reason, it is vital to distinguish the class
11.1.
TOPOLOGICAL
SPACES
231
of spaces in which the closure operation is equivalent to adding to the points of a given set the limits of convergent sequences of elements of this set. Let x be a point in the topological space S. The system O(x) of neighborhoods of the point x possessing the property that for any neighborhood V of the point x there is a neighborhood x contained in V is called the base of the space S at the point x or the neighborhood base of the point x. If the point i £ s has a countable base of neighborhoods, then the first axiom of countability is said to hold at this point. If the first axiom of countability holds at each point of the space H, then S is called the space with the first axiom of countability. In the space with the first axiom of countability, to obtain the closure A of the set A it suffices to complement A by the limits of the convergent sequences of elements from A. For specific spaces it may be difficult to describe the whole system of open sets determining the space topology. In such cases it may be more convenient to specify not the entire topology but merely a part thereof which determines the collection of all open sets of space in a unique way. The base of the topological space E is called a family B of open sets in the space E such that any open set in E can be represented as a union of some sets from B. The notion of a space base is one of the fundamental notions in topology, i.e. in many questions concerning the investigation of a topological space we may restrict our consideration to the elements of its base. The space may have many bases the largest of which forms the collection of all open sets in this space. Special significance is attached to the spaces with a countable base which are also called the spaces with the second axiom of countability. The topological space with the second axiom of countability is separable, i.e. in this space there is a countable everywhere dense set. The family of the sets which are open in E is the base if and only if it contains the base of neighborhoods of each point x € E. The collection of all possible unions of the sets making up the base of the space E which is complemented by an empty set coincides with the topological space E. In order that the family B of subsets of E be the base of a topological space in E, it is necessary and sufficient that any point x 6 E is contained in at least one G C E and if x S Gi OG2, Gi, G2 € B, then there exists G C B such that x 6 G C G± f] Gi ■ The above definition of topological space seems to be very broad, and hence it is usual to distinguish narrower classes of topological spaces which satisfy auxiliary conditions. Among such conditions, e.g., are axioms of countability which provide a means of investigating the space topology in terms of convergence. The availability of many "good" properties in topological spaces involves the possibility of separat ing one point after another by their neighborhoods. Such requirements imposed on topological spaces can be stated as axioms of separability. Null axiom of separability: of any two distinct points in a topological space, at least one point has the neigh borhood containing the other point
232
CHAPTER
11.
OPTIMIZATION
PROBLEMS
IN FUNCTION
SPACES
T h e spaces which satisfy the null axiom of separability are called To-spaces. First axiom of separability: for any two distinct points x and y in a topological space there is a neighborhood of the point x which does not contain the point y, and a neighborhood of t h e point y which does not contain t h e point x. T h e spaces satisfying the first axiom of separability are called Ti-spaces or sepa rable spaces. A very important feature of Ti-spaces is as follows: if A is a subset of Ti-space, t h e n for the point x to be limiting for t h e set A it is necessary and sufficient t h a t any neighborhood of this point contains infinitely m a n y points from A. Second (Hausdorff) axiom of separability: any two distinct points in a topological space have distinct neighborhoods. T h e spaces satisfying the second axiom of separability are called T2-spaces or Hausdorff spaces. The regular space is defined to be a Ti-space in which, for any point % and any closed set which does not contain this point, t h e r e exist disjoint neighborhoods. T h e normal space is defined to be a T\-space in which any two closed disjoint sets have disjoint neighborhoods. If t h e topological space H is normal and t h e set A is closed in E, then for any neighborhood V of the set A there is a neighborhood of Vo of t h e same set such that
V0c v Note t h a t all of the above classes of spaces, except To-spaces, satisfy the first axiom of separability which is equivalent to the following requirement: each set consisting of one point is closed. Let E be a topological space with topology r , M C E. T h e topology T then defines on the set M the induced topology which is composed of all sets of the form M PI G, where G 6 r . T h e thus defined space M is called a subspace of t h e topological space S. If the set V C M is open (closed) in S, then V is open (closed), but not the other way round. V C M is closed in M if and only if it coincides with the intersection of M and a closed set from S. T h e m a p / : E —>• Y of t h e topological space H into t h e topological space Y is called continuous in the point i £ 5 if for any neighborhood W of the point y = f(x) there is a neighborhood V of the point x such t h a t f{V)Y is called continuous if it is continuous in any point x £ S . If Y is a numerical line of R1, then we have a continuous function / . In order t h a t t h e m a p of the topological space 5 into t h e topological space Y be continuous, it is necessary and sufficient t h a t an inverse image of any open (closed) set from Y is open (closed) in S. T h e m a p / : E —> Y of the topological space E onto t h e topological space Y is called homeomorphism if it is one-one and mutually continuous (i.e. the continuous m a p / and the inverse m a p / : E —> Y). In this case t h e spaces E and Y are called homeomorphic with respect to each other. T h e topological space S is called disconnected if it can be represented as t h e union of two disjoint nonempty closed sets: S = V U W T h e sets V and W are thus closed
11.1.
TOPOLOGICAL
SPACES
233
and open simultaneously. Such sets are called open-closed. In any space there are at least two (trivial) open-closed sets, namely: an empty space and the whole space. The topological space is disconnected if and only if it contains a nontrivial open-closed set. The space which is not disconnected is called connected. The continuous image of the connected space is connected. Since any set M C S can be topologized by the induced topology, we may deal with the connectedness of the set (in the topological space) M. If we add to the connected set M any number its limit points, we obtain a connected set. The topological space H is called totally disconnected if it contains no nontrivial connected subspaces. The topological space is called extremally disconnected if it is regular and the clo sure of any its open set is open (open-closed). In the extremally disconnected space, the interior of any closed set is closed and the disjoint open sets have disjoint closures. The family of open sets in the topological space E is called the open covering of the space E if the union of these sets coincides with E. The space E is called bicompact if out of any open covering of E we may choose a finite system of sets forming the open covering of E. The set M C E is called bicompact if it is bicompact in its induced topology. Each infinite subset of the bicompact space has at least one limit point. The closed subset of the bicompact space is bicompact. The Hausdorff bicompact topological space is called a bicompactum. The bicompactum is closed in any Hausdorff space containing this bicompactum. The closed subset of bicompactum is a bicompactum. Any bicompactum is a normal topological space. The continuous image of a bicompact space is a bicompact space. The one-one and continuous map of bicompactum into the Hausdorff space is a bicompactum. The function which is continuous in the bicompact space E is bounded on E and achieves on E its maximum and minimum values. The set M is called everywhere dense in the topological space S if M = E. The set M C S is everywhere dense in E if and only if its intersection with any open set is nonempty. The set M is called nowhere dense in the space E if the complement to the clo sure ~M is everywhere dense in E. The closure of the nowhere dense set contains no nonempty open sets of the space E. The set M is called the set of first category if it can be represented as a countable union of nowhere dense sets; otherwise M is called the set of second category. Any complete nonempty metric space is the set of second category. If M is the set of first category in a bicompact topological space E, then E \ M is everywhere dense in E. Let {fn(x)} be a sequence of continuous real functions specified on the topological
234
CHAPTER 11. OPTIMIZATION
PROBLEMS
IN FUNCTION
SPACES
space H, and let a finite limit of this sequence be in every point x of the space S: lim/ n (a:) =
f(x).
The collection of all discontinuities in the function f(x) then forms the set of first category. Metric spaces constitute an important class of topological spaces. The set E is called a metric space if for every pair x, y G S there is a real number p{x,y) called a distance between the elements x and y, in which case 1) p(x,y)
> 0, p(x,y)
2) p(xty)
=
3) p{x,y)
< p(x,z)
= 0 if and only if x = y;
p(y,x); +
p(y,z).
The thus defined function ^ H x S ^ f l 1 is called a distance function (a metric). An open sphere S(XQ, r) in the metric space E with the distance function p is called a collection of points i £ 2 satisfying the condition p{x0,x)
< r.
The point XQ is called a center of the sphere S(XQ, r), with r as its radius. The collection of points x £ E satisfying the condition p{x0,x) is called a closed sphere
< r
S(x0,r).
The open sphere of radius e centered at x0 is called an £-neighborhood of the point x0. In any metric space, the collection of open spheres forms the base of a particular topological space. Moreover, the base of the thus defined topological space forms the collection of all open spheres with rational radii. However, it does not always happen that the metric space has a countable base. The following assertion is true: the metric space has a countable base if and only if it is separable. Note that any metric space (even that which is not separable) is the topological space with the first axiom of countability. Any metric space is Hausdorff space. Moreover, any metric space is normal. Thus, any metric space can be regarded as topological. The topological space whose topology is generated by some metric is called metrizable. In order that the topological space with a countable base be metrizable, it is necessary and sufficient that this space is normal (Urysohn's theorem). If between the elements of two metric spaces E and Y it is possible to establish a one-to-one correspondence such that the distances between the corresponding pairs of elements in the spaces S, Y are equal, then such spaces called isometric.
11.1.
TOPOLOGICAL
SPACES
235
Let E be a metric space with p as its distance function. As rioted before, E can be regarded as a topological space with its base made up of open spheres. In this case each point i £ H has a countable base, and hence to describe a topology in S, it suffices to restrict one's consideration to convergent sequences. If in the resulting topological space x" -> x, then p(xn,x) -» 0, and conversely p(xn,x) -> 0 implies xn -+ x. It can be shown that the function p(x,y) is continuous in that from xn -t x, y" -+ y follows p(xn,yn) -» p(x,y). The point sequence {x n } in the metric space E is called fundamental if p(x",a: m ) ->■ 0 as m,n -» oo. Using the notion of a fundamental sequence, we may separate out an important class of metric spaces called complete. The metric space E is called complete if its each fundamental sequence converges in E. The bicompact metric space is complete and separable. The complete metric space is the set of second category (Baire's theorem). In order that the metric space S be complete, it is necessary and sufficient that, in this space, each sequence of mutually embedded closed spheres whose radii approach zero has a nonempty intersection (lemma of embedded spheres). If the metric space is not complete, it can always be embedded in some space which is complete. Let S be a metric space. Completion of the space S is defined to be a complete metric space S* such that E is isometric with respect to a subspace Y of the space S* (i.e. Y C E* and specified on Y x Y is the metric which is a restriction of the distance function of the space E* to the set Y x Y) with Y being everywhere dense in E* Each metric space has a unique completion which is accurate to isometry. Note that some properties of metric spaces center around the notion of bicompactness. The metric space is bicompact if and only if each infinite sequence of its points contains a convergent subsequence. Bicompact metric spaces are also called bicompact. Let A be a set from the metric space S with £ as a positive number. The set M C A is called an e-network of the set A if for every x £ A there is at least one point y £ M for which p(x,y) < e. If for an e > 0 the set A has a finite e-network, then it is bounded. The set A is called completely bounded if it has a finite e-network for any e > 0. In order that the metric space be completely bounded, it is necessary and sufficient that a fundamental subsequence might be chosen out of any infinite sequence of points in this space. In order that the metric space be bicompact, it is necessary and sufficient that it is complete and completely bounded. We shall now provide examples of topological spaces. Let E be composed of two points a and b. All S, the set b and an empty set will be called open sets. The thus defined topological space is called a connected two-point
236
CHAPTER 11. OPTIMIZATION PROBLEMS IN FUNCTION SPACES
space. The connected two-point space provides an example of the To-space which is not a Ti-space. Let S be an arbitrary set. All E and an empty set are taken to be open. The thus obtained topological space satisfies none of the axioms of separability. Let us consider an interval [0,1], whose topology is generated by its naturally determined metric, and add to it a new point x. Neighborhoods of the point x are taken to be the sets made up of the point x and all points in the interval [0,1] except their finite number. The thus defined topological space is a TVspace, not a T2-space. As noted before, each metric space, can be regarded as topological. For this reason, in the examples given below, we merely indicate metrics of a particular space. In the space Rn of all n-dimensional vectors x = (x\, x2, ■ ■ ■, xn) each of the metrics 1 1/2
s>
Pi{x,y)
=
p2(x,y)
= max \xt - yt Ki
p3(x,y) = ^ l
x
--^
determines the same topological space. In the space C[Q, 1] of all functions which are continuous in the interval [0,1] the topology is generated by the metric ,1/2
j(f{t)-g(t)fdt
P(f,g)
Let S be the space of all positive numerical sequences x ='■ The metric in 5' is defined as follows: ,
N
^1
[xi,xt,..
■
,
x
|iW-V.|
+ \xi + y,\ Let Q be a bicompactum, and let C(Q) be the space of all continuous functions on Q. The distance between the functions f,g£ C(Q) can be defined as p(f,g)
=
tefl
sup\f(t)-g(t)\.
In connection with the last-named example, we shall make several remarks which may be useful in what follows. 1. The function / 6 C(Q) is uniquely determined by its values on the everywhere dense set G C Q.
11.2. LINEAR TOPOLOGICAL
SPACES. CONVEX
ANALYSIS
237
2. Let Q be a metric bicompactum with r as its metrics, and let C(Q) be the space defined as above. We shall provide bicompactness conditions for a set in the space C(Q). In order that the closed set F of continuous functions be bicompact in C(Q), it is necessary and sufficient that: the set of functions F is uniformly bounded, i.e. there exists a constant M such that |/(<)| < Af for all t 6 Q and all / G F; the functions of the set F are equicontinuous, i.e. for every e > 0 there is a 5 > 0 such that r(t',t") < 5 implies \f(t') - f(t")\ < c for all functions / G C(Q) (Arcella-Askoli theorem). 3. The convergence of the sequence of elements from C(Q) is nothing other than a uniform convergence of the function sequence.
11.2
Linear Topological Spaces. Fundamentals of Convex Analysis
The linear (vector) space over the field R1 of real numbers is defined to be a nonempty set L satisfying the following conditions: for every pair of elements x,y G L there is a uniquely determined element z G L referred to as the sum of elements x, y and designated x -j- y. In this case 1) x + y = y + x; 2) {x + y) + z = x + (y + z); 3) there exists an element 0 G L such that x + 0 = x for any x G L. The element 0 is called the null of the space L; 4) for every x € L there exists an element (—x) called the additive inverse such that x + (—x) = 0; for every number a e R1 and every element x e L there is a uniquely determined element ax G L called the product of the element x by the number a and 5) a(/3x) = (aP)x; 6) 1 x = x; 7) (a + P)x = ax + fix; 8) a(x + y) = ax + ay.
238
CHAPTER
11. OPTIMIZATION
PROBLEMS IN FUNCTION
SPACES.
A nonempty subset V of the linear space L is called a subspace if V is a, linear space with respect to the operations of addition and multiplication by a number as specified in L. Let S, T C L. The algebraic sum of the sets S, T is taken to be the set
S + T = {x: x = s +1, s e S,
teT}.
The product of the set S C L by the number a 6 R} is taken to be the set aS = {x: x = as, s € S} A nonempty subset L" of the linear space L is called a linear manifold if for a particular element x 6 L and subspace L' C L L" = U + x. The set S C L is called convex if, together with any pair of points s,t £ 5' it contains all points of the form As + (1 - X)t,
where 0 < A < 1. The space L is all convex. The empty set is considered convex by definition. The intersection of convex sets is convex, but the union of convex sets generally is not convex. Let S C L. The intersection of all convex sets containing S is the least among the convex sets containing S. This set is called a convex hull of S and designated co S. The convex hull may also be defined as the collection of all convex combinations of elements from S, i.e. combinations of the form \lSl
+ X2s2 + ••• + Xksk,
where Aj + Aj -f • ■ • + Ajt = 1, A3- > 0, s3 G 5, j = l,n. The elements x ,%,..., x of the linear space L are called linearly independent if for any numbers Ai, X%,..., Aj, 6 R1 from the equality A,x' + X2x2 + ••• + Xkxk = 0 it follows that Aj = A2 = ... = Xk = 0; otherwise these elements are called linearly dependent. The infinite system of elements x,y,. . . in the linear space L is called linearly independent if any one of its finite subsystems is linearly independent. If the maximum number of linearly independent space elements is finite and equal to n, the space L is said to have dimension n; otherwise the space L is called infinitedimensional. The map / of the linear space L into the set of real numbers is called a linear functional if f(ax + /3y) = af(x) + 0f{y) for all x,y e L, a,j3 £ Rl. A sublinear functional is taken to be a real-valued function p which is defined on L and is such that p(x + y) < p(x) + p(y) for any x,y 6 L and p(ax) = ap(x) for a > 0.
11.2. LINEAR TOPOLOGICAL
SPACES. CONVEX
ANALYSIS
239
Suppose that a sublinear functional p is specified on a linear space L, and M is a linear subspace in L. If / is the linear functional for which f(x) < p(x) with all x G M, then it can be extended to a linear functional g on L for which g(x) < p{x) with all i € I and g(x) = f{x) with x G M (the Hahn-Banach theorem). The linear space E over the field of real numbers which is endowed with a par ticular topology is called topological linear space if the operations of addition and multiplication by numbers in E are continuous in the topology specified in E, i.e. if z = x + y, then for every neighborhood V of the point z it is possible to select neighborhoods 5, T of respective points x, y such that S + T C V; if z = a0x, then for every neighborhood V of the point z there is a neighborhood S of the point x and a number £ > 0 such that aS C V for |a — a 0 | < e. From the definition of the topological linear space it immediately follows that: 1) if S, T are open sets in E, then the set S + T is also open; 2) if 5 is open, then with any A / 0 the set \S is also open; 3) if S is closed, then A5 is also closed for any A G R1 Thus, in order to specify a topology in E, it suffices to point out the base for neighborhoods of one of the points in E, e.g., a zero point: if V is a neighborhood of zero, then the set V + x is a neighborhood of the point x G E. In the topological linear space, each point and the closed set containing no such a point have disjoint neighborhoods; therefore the topological linear Ti-space is reg ular. Note some properties of convex sets in topological linear spaces. The close convex set S in the linear topological space E is convex. The interior points of the convex set form a convex set. Let S be a bicompact convex set and let T be a closed convex set in the space E. Then the set S + T is closed in E. The Hausdorff linear topological space is called locally convex if every neighbor hood of zero contains a convex symmetric neighborhood V of zero (i.e. from x G V it follows that —x G V). Below are given theorems of separability which are crucial it theory of convex sets. Let S, T be convex subsets of a locally convex space E such that intT / 0, Sf)intT / 0. Then there exists a continuous linear functional / and a real number c such that / ( x ) < c for x G T and f(x) > c for x G S (theorem of separability in weak form). Let E be a locally convex space and let 5 be a closed convex set which does not intersect a bicompact convex set T. Then there exists a continuous linear functional / and a real number c such that f(x) > c for x G T and f(x) < c for x G S (theorem of separability in strong form). Let S be a convex set considered in a linear space. The point z is called an extreme point of the set S if there are no points x,y € S (x ^ y) for which Ax + (1 - \)y = z for some 0 < A < 1.
240
CHAPTER
11.
OPTIMIZATION
PROBLEMS
IN FUNCTION
SPACES
Let 5 be a n o n e m p t y bicompact convex subset of a locally convex space. Then S has at least one extreme point. If So is t h e set of e x t r e m e points of S, t h e n co S = S (the Crain-Millman theorem). We now dwell on some properties of convex sets in ft". T h e convex set T C Rn is said to have dimension k if the m i n i m u m dimension of a subspace containing T is equal to k. Convex bicompact sets of the same dimension are homeomorphic. Of great importance is the following property of convex sets in ft": for T c i f each point of t h e set co T may be represented as a convex combination of at most n + 1 points from T If 5 is a bounded subset in Rn, t h e n the collection of e x t r e m e points of the set co S designated So has t h e property: co So = co So = coS. Let Si,S2}. .., Sk {k > 3) be convex bicompact sets in Rn whose union is convex. k
If each k — 1 of t h e m have a nonempty intersection, t h e n f] S, ^ 0. !=1
Let Si,S2, - ■ ■ ,Sk (k > n + 2) be convex sets in Rn If each n + 1 of them have a nonempty intersection, then the intersection of sets Si, S2, ■ ■ ■ , Sk is nonempty (Helly's theorem). Let 5' be a convex set and let / be a real-valued function on S. T h e function / is called convex if f(Xx + (1 - \)y) < Xf(x) + (1 - X)f(y) for all A € [0,1], x,y e S. T h e function / is called concave if the function —/ is convex. Let S be a convex set in the linear space L. T h e function / defined on S is convex if and only if the set {(x,r):xeS, f(x)0 on (a, b). In a more general case: if T C ft" is a n o n e m p t y open convex set, t h e function / is defined and twice continuously differentiable on T , t h e n it is convex if and only if its Hessian
H(x)
d2f dxidxj
(x)
in any point is a nonnegative definite matrix. In any interior point of its domain of definition S, t h e convex function / is con tinuous and has a derivative with respect to any direction. T h e function / which is differentiable on a closed convex set T C ft" is convex if and only if for any x,y G T
(gradf,y-x)
< f(y) -
(here (x, y) is the scalar product of vectors x, y).
f(x)
11.2. LINEAR TOPOLOGICAL
SPACES. CONVEX
ANALYSIS
241
Note that if the functions /,, i = l,k, are convex, then the following functions are also convex h
f(x) =
'£aift(x)
(o,-> 0 , i = I ^ )
•=i
tp(x) = max fi(x). l
Let T g Rn be a closed convex set and let the real-valued functions >,, i = l,m, be convex and continuous on F. Then the set
<5 = {x g r:y>i(x) <0, i = T~^} is convex and closed. We consider the minimization problem for a convex function
The point x* g fl" is called a local minimum for the problem of this form (not necessarily convex) if x* g Q and there is a neighborhood V C Rn of the point x" such that for every x 6 V f]Q the inequality y?(a~*) < v{x) ls true. If, however, this inequality holds for all x g Q, then i* is called a global minimum. Convex programming problems have the following important property: every local minimum in the convex programming problem is its global minimum. We say that the convex programming problem satisfies Slaiter's condition if there exists a point x g Q such that
L(x,y) = ip(x) + J2vi
considered for x g T, y > 0, y G Rm The function L(x,y) is called the Lagrange function for the convex programming problem. The saddle point of the Lagrange function L(x,y) on the set x E T, y > 0 is denned to be a pair ^",y* (x" G I\ y* > 0) such that the inequalities L(x*,y) < L(x\y")
< L[x,y*)
hold for all x € I\ y > 0. The following theorem is crucial in convex programming. Suppose the convex programming problem tp(x) —>■ min
x eQ satisfies Slaiter's condition. The point x* g i?" is a solution of this problem if and only if there is an y" > 0 such that the pair Z*, y* forms a saddle point of the Lagrange function L(x,y) on the set x g F, y > 0.
242
11.3
CHAPTER
11.
OPTIMIZATION
PROBLEMS
IN FUNCTION
SPACES
M e a s u r e Spaces. P r o b a b i l i t y Spaces. M o d e l i n g of R a n d o m Variables
Let us consider an arbitrary set T. A nonempty collection E of some of its subsets is called algebra if it satisfies t h e following conditions: for any A, B G E the set A \J B G E; if A e E, then T\ A e E. From the above definition it follows t h a t for any A, B € T, t h e intersection Af]B and the difference A \ B are also contained in E. Moreover, the e m p t y set and the entire set T are elements of E. By induction, we conclude t h a t algebra E is closed with respect to the operations of union and intersection of finitely m a n y sets. T h e nonempty collection £ of some subsets of t h e set T is called a-algebra if E is an algebra which is closed with respect to the union (and hence with respect to the intersection) of a countable number of sets. T h e pair of objects — the set T and some cr-algebra of its subsets E — forms a measurable space (T, E ) . Using cr-algebra of E of subsets in t h e set T, we define a real-valued function y which has the following properties: for every A G E y(A) > 0 (nonnegativity of /J,). Here it is assumed t h a t //(0) = 0; for every countable sequence {A,} of mutually disjoint sets from E
= EMA) 1i
i
(countable additivity or addititwity of the function y.). T h e thus defined function y is called a measure on cr-algebra E. If y.(T) < co, then the measure y is called finite. If the set T is representable as a union of the countable collection of sets A{ G E such t h a t y{At) < oo, then the measure y is called a-finite. If (T, E) is a measurable space and y is a measure on cr-algebra E, then the triple (T, £,/u) forms a measure space. The sets A G E are called y-measurable (or measur able in measure y.) and the value of y(A) is called a y-measure (or simply a measure) of the set A. T h e set B C T is called a null set (a set of measure zero) if B G E and y{B) = 0. A certain property pertaining to the points t of the set T is said to hold almost everywhere (in measure y) if it holds for all t G T exxept the null set. Let (T, E) be a measurable set and let
/(0 = X>U(0,
11.3. MEASURE SPACES. PROBABILITY
243
SPACES
where A, are real numbers, the disjoint sets A, are contained in E, ^A,{t) is a charac teristic function of the set A,-, i = 1, k. For every measurable function tp there exists a sequence of measurable finite-valued functions converging to tp for all t 6 T. Let (T, S,^i) be a finite /j-measure space. If {?„} is a sequence of measurable functions converging almost everywhere to a measurable function tp, then for every £ > 0 there exists a measurable subset of E such that p.{t \ E) < e and the sequence {fn} converges on the set E uniformly to tp (Egorov's theorem). Let us consider the space with measure (T, E, p>). The measurable finite-valued function
f(t) = J2\,xA,(t) is called integrable in the set T if k
^|At|M(A.)
k
The quantity J2 A,^(A,) is called the integral of the function F(t) over the set T and i=i
is designated as
or
fF(t)ij.(dt) T
I F{t)dp.. T
If the function ip(t) is measurable, the sequence {
[\
= Q,
n,k—J-co J
then tp is called integrable or summable over T. In this case, there exists a finite limit lim / tp(t) fi(dt), T
which is called the integral of the function tp in the set T and is designated jtp{t)fi(dt) T
or
jip{t)dp,. T
If the function tp is integrable, then for an arbitrary set S G E the integral fip(t)n(dt) over 5 is taken to be s j tp(t)li(dt) =
jxs{t)p{t)ii(dt),
244
CHAPTER
11.
OPTIMIZATION
PROBLEMS
IN FUNCTION
SPACES
where Xg(i) is the characteristic function of the set 5 . Suppose t h e functions ip{t), tfn(t), (n — 1 , 2 , . . . ) are measurable, A 6 E, fjt(A) < oo. The sequence {ip»(t)} converges in measure over t h e set ^4 to t h e function ip{t) of for every e > 0 lim fi({t g 4 : |y>(t) - ¥=„(i)l > e}) = 0. n-+oo
If the measure fj, is complete and the sequence {tpn(t)} of measurable functions converges to t h e function ip(t) almost everywhere, then (p(t) is measurable and the se quence {
= / ip{t) fj,(dt)
TH-OO J
J
T
T
(the Lebesgue theorem). Let fi, v be the finite measures specified on a measurable set (T, E ) . T h e measure JJ, is called absolutely continuous in measure u if v{A) = 0 implies (J,{A) = 0. If t h e measure /i is absolutely continuous in measure i/, t h e n t h e r e exists a finite nonnegative measurable integrable function ip{t) such t h a t
lx{A) = jV{i)v{dt) A
for every measurable set A G E (the R a d o n - N i k o d i m t h e o r e m ) . In the functional analysis, much importance is attached to the spaces LP(T, E , A O , 1 < p < oo defined as follows. Let (T, E, /i) be a measure space. For 1 < p < oo t h e space LP(T, E, y.) is composed of all measurable real-valued functions tp(t) for which t h e function | y ( i ) | p is integrable. The space Loo(T, E,/i) is composed of all measurable functions y> for which there is a member a^ such t h a t | y ( t ) | < av almost everywhere. T h e following inclusions are valid: Lp(T,£,M)cL,(r,E,^)
forp>g.
We shall now consider a random experiment. Let E be a set of possible events which may occur during this experiment. To specify t h e set E we need not always to describe each of t h e events: it suffices to define a particular subset fi of these events which, in a sense, generate the set E. In probability theory, the elements from U are called elementary events, and fi is chosen to be such t h a t every random event from E may be represented by a set of elementary events using basic set-theoretic operations. When E is finite it would appear natural t h a t t h e set E is an algebra of subsets from fi. It is possible to define over E a nonnegative function P , i.e. the probability of occurrence of events appearing in E and possessing t h e following properties: 1) t h e event fi is assigned t h e greatest possible value P(fl) = 1; 2) if A . S e E , A[)B =
11.3. MEASURE SPACES. PROBABILITY
SPACES
245
If, however, E is not finite, we have to consider infinite sequences of events and operations on them. The simplest of these are intersection and combination of in finitely many events (sets). We thereby arrive at the notion of cr-algebra. Among all cr-algebras containing this algebra it is usual to distinguish the least algebra rep resenting the intersection of all such algebras. Probability P specified in terms of cr-algebra E must be naturally placed in correspondence with operations on infinite set sequences. Namely, in this case we require its countable additivity. The thus defined probability is a, measure in cr-algebra E and is called a probability measure. The measurable space (ft, E) together with the probability measure P specified on E form a probability space (ft, E,P). Thus, the probability space is a special case of the measure space. Note that if ft contains no more than a countable number of elements and E coincides with the set of all subsets of ft, then the probability is completely defined by its values over elementary events. The quantity which varies under conditions of a random experiment is random. If a random experiment is described by a probability space (ft,E,P), then the random variable is a function of the elementary event u> £ ft. The assumption that such a quantity can be measured under conditions of this experiment implies that among the random events described in E there must be events as follows: for a given open set G of a real-valued straight line R1 the value of a random variable x(ui) (u> € ft) belongs to the set G. Thus the function x(u>) is measurable. The expectation of a random variable x is taken to be the integral fx{w)P{du). n The expectation of x is denoted by M i . The variance of a random variable x is taken to be the integral f(x{u>)-Mx)2P{du).
n The variance of i is denoted by D(i) or a\. The correlation moment of random variables x and y is taken to be the integral Kx,y = | ( x ( w ) - Mi)(j(w) - My) ?{du). a If Fx(t) is a distribution function of the random variable x, then oo
Mx= J tdFx(t). —oo
In the event that dFx(t) = p(t) dt, p(t) is called a distribution density of the ran dom variable x and x
Mx= [ tp(t)dt.
246
CHAPTER
11. OPTIMIZATION
PROBLEMS
IN FUNCTION
SPACES
Let us consider a probability space (fi, E,P). Let A, B 6 E, and P(B) ^ 0. The conditional probability of event A with a given B is taken to be P(A\B) = PB(A) =
P(Af]B)/P(B).
The function P B defined in terms of (j-algebra E is a probability measure and is called the conditional probability with a given B. We thereby introduce a probability space (fi, E, P B ) . If in this space there exists the expectation of a random variable x, then it is called the conditional expectation with a given B and is denoted as M{x\B)=
j a
x(uj)PB{du).
Here
M(x|B)=p^y/xHP(dw). \B)
The above-introduced notions of conditional probability and conditional expecta tion allow useful generalizations. For the probability space (fi,E,P) we shall consider the set A 6 E and some cr-algebra A C E (subalgebra E). Let A contain a countable collection of mutually exclusive events Bk, k = 1,2,... , and [j Bk = fi, P{Bk) > 0. Then F(A\Bk) =
P(Af]Bk)/P(Bk).
With the A fixed, the conditional probability P(A\Bk) depends on Bk and can be regarded as a random variable whose values are constant in the sets Bk, k = 1,2,. . . . This quantity is designated P(A\A) or P&(A). Since for a fixed u € fi the function PA is a probability measure on E, the conditional expectation of a random variable x with respect to A may be expressed as M(x|A)= f
x{u)?(du\&).
M(x\ A) is measurable in cr-algebra A by the function taking constant values in the sets Bk, k = 1,2,... With the Bk fixed, we have M{x\Bk)P(Bk)
=j
x(to)P(du,).
Bk
If, however, the set B € A can be defined as the union of countably many (disjoint) sets Bk, then f M{x\/\)P{du}) = f x(to)P(du B
B
Based on this relation, the conditional expectation can be defined for the general case.
11.4. ORDERED
SPACES.
247
Let (ft, £, P) be a probability space and let cr-algebra A be contained in E. The con ditional expectation M(x|A) of a random variable x for this cr-algebra A is taken to be a A-measurable function for which with each B 6 A I M{x\A)P(duj) B
=j
x(io)?{duj).
B
The random variable M(X,4|A) (XA(to) is the characteristic function of the set A) is called the conditional probability P(A|A) of the event A € £ with respect to A. Let i = x(u>) be a random variable taking values in Rn (a random vector) and let Ax be the least a-algebra in which the function x(io) is measurable ( the least (j-algebra generated by the sets {UJ: X{LO) £ G}, where G is the Borel set in Rn) The conditional expectation M(y\x) of a random variable y with respect to a random variable x is defined to be M(y\Ax). The conditional probability of a random variable y with respect to a random vari able x is taken to be P(y\x) = P(y\Ax). The conditional variance D(y\x) of a random variable y with respect to a random variable x is taken to be J[y-M(y\x)}2P(du).
D(y|x) = a
11.4
Ordered Spaces
The linear space E over a field of real numbers is called a vector lattice if E is partially ordered, for every I , J £ E there exists a supremum i V y £ H and infimum x A y € E, and the operations and the order are adjusted as follows: for every z £ H , from x > y follows x + z > y + z; for every real A > 0, from x > y follows Xx > \y. It is easy to see that the vector lattice has a supremum and infimum for any finite number of elements. Zero of the vector lattice 5 is designated 0. For i G H w e have the following notation: i + = iVO (a positive part of element x), x_ = ( — x) V 0 (a. negative part of element x), \x\ = x + + x_ (an absolute value of x). for every x € H, x = x+ + x_ (Jordan representation of element x). Denote by H + the s e t { x G S : i > 0 } ( the cone of positive elements of the vector lattice E). The elements x , j £ S are called disjunctive if |x| A |y| = 0. In this case we write x d y. The element x G S is called disjunctive with the set E C H if x d y for every y e E (we write x d £ ) . The sets ei,e 2 C E are called disjunctive if for every x € E^, y £ E2 x d y (we write e! d e 2 ). The disjunctive complement of the set E C E is taken to be
Ed = {xeE:xd
E}.
248
CHAPTER
11. OPTIMIZATION
PROBLEMS
IN FUNCTION
SPACES
The strip in the vector lattice E is taken to be the set Y C S such that (Yd)d = yid _ y- p o r e v e r y £ c H the set Ed is a strip in S. If V = i ^ } 1 " , x £ E, then Y is called a strip formed by the element x. The ideal of the vector lattice S is taken to be a linear subspace V C H such that the inclusion x e Y follows from x £ E, y £ V', |x| < |y|. The ideal Y is called an order-dense ideal if Yd = {0}. The element x £ S + is called a unity element and denoted by 1 if x A y > 0 for every y > 0. The element e £ S is called a unity element if e A (1 — e) = 0. The collection of all unity elements is called the base of the vector lattice E and denoted by 0 ( E ) . Let ei, e2 € 0 ( E ) . If ex > e2, then ex - e2 £ 0 ( E ) . If ex d e2, then ei + e2 £ 0(E). If the vector lattice has an element possessing the property of unity, then it has an infinite number of such elements. For this reason, in what follows the unity vector lattice is always interpreted to mean that the unity 1 in this lattice is fixed. If S is the unity vector lattice, then the set Eo composed of all x £ E such that |z| < Al for a real A > 0 forms a vector lattice which is called the vector lattice of restricted elements ( the elements x £ E 0 are called the restricted elements of the lattice E). S 0 is an ideal in E. Moreover, Eo is an order-dense ideal of the vector lattice E. The complete vector lattice or the conditionally complete vector lattice is taken to be a vector lattice in which every bounded-above (below) set has a supremum (infimum). The ideal in the conditionally complete vector lattices a conditionally complete vector lattice. The conditionally complete vector lattice is an Archimedean vector lattice, i.e. from the fact that inequality nx < y with any n = 1,2,.. . holds for x £ E + , y £ S it follows that x = 0. Let Y be a strip of the vector lattice E, x £ E + . If among all the elements y 6 Y satisfying the inequality 0 < y < x there is the largest element, then it is called a projection of the element x onto the strip Y and is denoted by [Y] x. If there exist [Y] x+ and [Y] x_, then for an arbitrary x £ E the projection of element x onto the strip Y is determined by [Y]x-
[Y]x+ - [ Y ] s _ .
If E is a complete vector lattice, then for any one of its elements there exists a projection onto any strip. Every unity element of the conditionally complete vec tor lattice with unity is the projection of unity onto a particular strip. Conversely, the projection of unity onto any strip Y is a unity element. The operator [Y] is called a projection operator or a projector onto the strip Y [Y] is a linear operator mapping E onto Y If Y is a strip of the conditionally complete vector lattice E, then every element x £ E can be uniquely expressed as x = [Y]x + [Y d ]x. Let S be a conditionally complete vector lattice. The strips E a , a £ A, form a decomposition of S if:
11.4, ORDERED
249
SPACES
1) these strips are mutually disjunctive; 2) the set of these strips is complete in S, i.e. in 5 there is no element other than zero and disjunctive with E a for all a G A. Suppose the elements x„ i G J, of the vector lattice E are mutually disjunctive and in E there exists sup(x,) + , sup(x,)_. The union of elements x„ i G J, is taken to be an element designated Q X; and expressed as i O X, = SUp(x;) + — S U p ( x , ) _ . z
The union of finitely many mutually disjunctive elements is equal to their sum. If some of the strips form a decomposition of the complete vector lattice, then every element of this lattice can be uniquely represented as a. union of its projections onto the strips of this decomposition. Each conditionally complete vector lattice is decomposable into strips which are conditionally complete vector lattices with unity. Also to be noted is that every con ditionally complete vector lattice is the order-dense ideal of a conditionally complete vector lattice with unity. The conditionally complete vector lattice E is called an extension if it contains unity and each set of its mutually disjunctive elements is bounded. The vector lattice S is said to be isomorphic to the vector lattice Y if there exists the one-one mapping of E onto Y retaining algebraic and lattice operations. If the conditionally complete vector lattice E is isomorphic in the order-dense ideal of an extended conditionally complete vector lattice E, then S is called a maximum extension of the conditionally complete vector lattice E. Every conditionally complete vector lattice has a maximum extension determined uniquely up to isomorphism. If two extended conditionally complete vector lattices have isomorphic bases (i.e. there exists a one-one mapping of one of the bases into the other retaining structural operations), then these vector lattices are isomorphic. Let S be a conditionally complete vector lattice with unity. The trace of the el ement x G E (designated ex) is taken to be a projection of lattice unity onto a strip generated by this element. Note some properties of the trace: 1) if |x| < \y\, then ex < ey; 2) the projection of element x on the strip generated by ex is x; 3) x d y if and only if ex d ey; 4) if x = supx,, Xi G E + , i G J, then ex = supe^,. Denote by e* the trace of a positive part of the element Al — x (e* G G(^))The system of elements ef, where A takes all values from - c o to +oo, is called a characteristic of the element x. The characteristic has the following properties:
250
CHAPTER 11. OPTIMIZATION
PROBLEMS
IN FUNCTION
SPACES
1) ej < e* for A < ,u;
2) supe^ = 1; A
3) i n f ^ = 0 ;
4) supe£ = e5; M
5) if Y\ is the strip generated by exx, then [Y\]x < \exx for any A. If two elements I , J E 5 are such that e* = e\ for all A, then x = y (it would merely suffice to have the equalities satisfied with rational A). If x = inf xa, a £ A, then &\ = sup ex" for any real A. a
From the results follows that the investigation of conditionally complete vector lat tices reduces to investigating conditionally compete vector lattices with unity which, in turn, are embedded in the extended conditionally complete vector lattices. It turns out that this chain of embeddings can be complemented by employing a spe cial way of "embedding" the Archimedean vector lattice in a particular conditionally complete vector lattice. Namely, the following statement is true. For every Archimedean vector lattice Z there exists a conditionally complete vec tor lattice S called a conditional completion of Z and possessing the following prop erties: 1) Z is isomorphic to a vector sublattice 1 Y C 5; 2) if E C Z and in Z there exists sup E = x (inf E = x), then the isomorphic image of x in y is a supremum (infimum) in H of the image of E in V; 3) for every x £ H there are sets Z\, Z2 C Y such that x = sup Z\ = inf ZiThe conditional completion of the Archimedean vector lattice is uniquely deter mined up to isomorphism. In the extended conditionally complete vector lattice S it is possible to introduce the commutative operation of element multiplication in such a way that the multipli cation unity coincides with the lattice unity. In this case, for any restricted elements I , I / 6 So their product x ■ y e S 0 . If x,y > 0, then x ■ y > 0. In S, for every element x there exists its inverse element x~l 6 E, i.e. an ele ment whose trace coincides with trace ex of element x and x ■ x~l = ex. Note that the introduction of the thus defined inverse element fails to change E to a field. If e g 0 ( 5 ) and Y is a strip generated by e, then for every e ■ x = [Y] x, e~l = e. We shall now point out two forms of convergence in vector lattices with the em phasis on sequences. 'A subset Y of the vector lattice S is called a vector sublattice in S if Y is a linear subspace of the linear space H, and x\/y£Y,xf\y£zY for any x,y £Y
11.4. ORDERED
251
SPACES
The sequence {xn} of elements of the vector lattice E is called (o)-convergent to the limit of x £ E (designated xn -4 x or (o)-limz" = x) if there exist two monotone sequences of elements from E — a decreasing {y n } and an increasing sequence {zn} — such that x = inf y" = sup 2" and z n < xn < y" for all n = 1,2,... . The element sequence {xn} of the vector lattice E is called (v)-convergent to the limit of x £ E if there exists an element u £ E + which is called a convergence regulator and has the following property: for every e > 0 it is possible to select N such that \xn — x\ < EU for n > N. The thus defined convergence is called a regulator convergence. The regulator convergence is only meaningful in Archimedean vector lattices. In general, the regulator convergence implies the (o)-convergence, and not the reverse. The lattice and linear operations are continuous in the above forms of convergence. An important property of the conditionally complete vector lattice with unity is its representability as a bicompact-continuous function space. Let Q be an extremally disconnected bicompactum. Denote by COB(Q) the set of all continuous real-valued functions defined over Q and capable of taking infinite values in nowhere-dense sets. In this case, the continuity of the function x(t) in the point r £ Q, where X(T) = ±00 as for the finite value function, implies that limi(r) = ±00. t-+T
^ '
We introduce an ordering in C(X(Q) setting x > y if x(t) > y(<) over the entire Q. Then we write x > y if x(t) > y(i) in at least one point. Let us linearize CrXl(Q). Let x,y £ Coz,(Q), X,fJ. £ R1 The set E of points, in which x(t) and y(t) simultaneously take finite values, is open and everywhere dense. There exists a unique function z(t) £ CX{Q) such that
z(t) = \x(t) + tiy(t) on E. This definition of linear operations in C oa (Q) makes possible all axioms of linear space over the field of real numbers. The partially ordered and thus linearized space CX,(Q) transforms to a vector lat tice. Moreover, the vector lattice Cao(Q) is an extended conditionally complete vector lattice. Here the function x(t) = 1 is taken to be unity in C^Q), and hence the col lection C{Q) of all bounded functions from Cro(Q) forms a conditionally complete vector lattice of restricted elements. We shall now formulate the theorem to represent the conditionally complete vector lattice as a continuous function space. For every conditionally complete vector lattice E with unity there exists an ex tremally unconnected bicompactum Q such that E is isomorphic to the order-dense ideal Y of the vector lattice CooiQ). In this case, isomorphism can be implemented in such a way that Crx,(Q) C Y and the identically unity function corresponds to unity of the conditionally complete vector lattice E. In what follows, isomorphism will be interpreted to mean exactly this form of isomorphism. Note that the bicompactum Q is invariant under any choice of unity in E and is determined up to homeomorphism.
252
CHAPTER
11.
OPTIMIZATION
PROBLEMS
IN FUNCTION
SPACES
T h e conditionally complete vector lattice with u n i t y is regarded as extended if and only if its image coincides with Coo(Q) when it is e m b e d d e d in t h e condition ally complete vector lattice CX)(Q) satisfying the conditions of this theorem. Thus, the m a x i m u m extension of the conditionally complete vector lattice with unity is the space C0o(Q) for the order-dense ideal of which 3 is isomorphic. T h e conditional completion of each Archimedean vector lattice and t h e embed ding of any conditionally complete vector lattice in t h a t with unity permit a partial extension of the above results to arbitrary Archimedean vector lattices. For every Archimedean vector lattice 3 there exists an extremally disconnected bicompactum Q such t h a t 3 is isomorphic with respect t o a vector lattice of C^Q). T h e use of t h e above theorem makes possible the following result which is called the relation conservation principle. Suppose there are expressions u ( x i , x2,.. . , xn) and v(xi, x2, ■ • •, xn) made up by means of finitely m a n y linear and lattice operations on variables x\, x%,,.., xn which m a y take values in an arbitrary vector lattice. In the Archimedean vector lattice S, t h e inequality u < v holds for any values of independent variables if and only if arbitrary real numbers are substituted for We shall now clarify some basic notions in t h e conditionally complete vector lat tice C „ ( Q ) . T h e disjunction of two functions from Cao(Q) implies t h a t t h e y are nonzero over disjoint sets. Therefore, if x d y, then there exists a partitioning of the bicom p a c t u m Q into two disjoint open-closed sets E\ and E2 such t h a t x(t) = 0 on E\ and y(t) = 0 on E2. T h e base of the conditionally complete vector lattice Coa(Q) is composed of char acteristic functions of all possible open-closed sets in t h e b i c o m p a c t u m Q. If the set E is open-closed and X% = e(t) is its characteristic function, t h e n t h e strip generated by the element e consists of all functions of t h e space Coo(Q) vanishing on Q\ E. T h e strip Y of the conditionally complete vector lattice C00(Q) can be regarded as an independent vector lattice Coo(E), where E is an open-closed set from Q. This fact will be used in the following notation. If the strip Coo(E) is generated by the element x G Coo(Q), then E = {t 6 Q: x(t) ^ 0}. T h e projection of y £ Cao(Q) onto the strip Cao{E) = Y is defined as follows lY]v(t)
= z(t) =
fy(t)
foteE;
If t h e strip C o t ) (£) is generated by the element x, t h e n t h e t r a c e ex of element x is a characteristic function of the set E. To be noted also is the following fact. If x, y, z are elements of t h e conditionally complete vector lattice Coo(<2), x-y = z, then at every point t G Q, where t h e functions x(t), y(t) are finite, the product is determined in a pointwise manner: x{t)-y{i) = z(t). If the element x~x is the inverse of x, then
x~\t) =
1
Wi
11.4.
ORDERED
SPACES
253
at t h e points, where the values of x{t) are nonzero and finite (the set of such points forms in Q an open everywhere-dense set). We shall now give examples of vector lattices. 1. If a "coordinate-wise" ordering of elements (vectors) is introduced in the linear space ff, i.e. (xi,x3,... ,xn) > {yl,y2,. ■ ■ ,y„) if Xj > y}, j = T~n, then Rn trans forms to a vector lattice. Moreover, Rn becomes an extended conditionally complete vector lattice. T h e extended conditionally complete vector lattice S of all infinite numerical se quences. 2. Let T be an a r b i t r a r y topological space and let C(T) be the set of all con tinuous real-valued functions on T. Here C(T) is linearized in the ordinary way: if x,y,z £ C(T), A,fi € R1, then z = Ax + p.y implies t h a t z(t) = \x(t) + fiy(t) for all t £ T. A partial ordering in C(T) is also introduced in a pointwise manner: x > y if x(t) > y(i) for every t £ T. T h e C ( T ) thereby transforms to a vector lattice. The vector lattice C(T) is not complete. 3. For an arbitrary set M, we denote by F(M) t h e collection of all real-valued functions specified on M. Linearizing F(M) and introducing in F(M) a partial ordering in t h e same way as for the space C(T), we transform F(M) into a vector lattice. F{M) is an extended conditionally complete vector lattice. 4. Let (T,Ti,fi) be a cr-finite measure space and let S(T, E,/i) be all measurable almost everywhere finite functions. Those functions from whose values on T coincide almost everywhere (such functions are called are identified among themselves and are regarded as t h e same element S(T, E,/Lt). T h u s the elements of the set S(T, E,/i) represent the classes of functions, and t h e function x(t) 6 S(T,T,,fj.) is interpreted to mean any equivalent functions.
the set of S(T, S,/i) equivalent) of the set equivalent one of its
Algebraic operations are introduced in S(T, S,/u) routinely, i.e. in a pointwise manner. A partial ordering is introduced as follows: we have x(t) > y{t) if this inequality holds for all t 6 T , except the set of measure zero, i.e. almost everywhere. Then we write x > y if x > y and x(t) > y(t) in t h e set of positive measure. T h e thus defined vector lattice S(T, £ , / i ) is an extended conditionally complete vector lattice whose unity is taken to be any measurable function x(t) for which x(t) > 0 almost everywhere. In general, t h e unity of this conditionally complete vector lattice is chosen to be t h e function taking on T the value of 1. If (T, £,/u) is a probability space, t h e n t h e space of r a n d o m variables S(T, T*,fl) may also be regarded as an extended conditionally complete vector lattice. This fact is very i m p o r t a n t and serves to use t h e vector lattice techniques for purposes of stating and investigating a variety of stochastic models.
254
CHAPTER
11. OPTIMIZATION
PROBLEMS
IN FUNCTION
SPACES
5. Each of the spaces LP(T, £,/u), linearized and ordered in accordance with the same rules as S(T, E, ft), is a conditionally complete vector lattice with unity. Here the conditionally complete vector lattice L0O(T, E,/u) is the conditionally complete vector lattice of restricted elements, while S(T, E,//) is the maximum extension of any one of the spaces LP(T, E , / J ) .
Note that the properties of the vector lattice are essentially dependent on the way in which its elements are ordered. For example, the vector lattice whose elements are lexicographically ordered is not a conditionally complete vector lattice; it is not even an Archimedean vector lattice.
11.5
Linear Optimization in Conditionally Complete Vector Lattices
Let E be an extended conditionally complete vector lattice, where the operation of multiplication of elements is introduced, and let h be the isomorphism of E on C^Q). If x 6 E, then the element ha; of the conditionally complete vector lattice COD(Q) is generally designated as x(t). We shall now consider the problem n
L(x) = 22 cjxj -> inf
(11.1)
3= 1 n
Y^a^Xj = 6,, £; S E+,
i=
l,m
(11.2)
j = l,n
(11.3)
where a,j,fe,, CJ £ S, i — l , m , j — l , n , m < n. The vector x = (xi} x 3 , . . . , xn), x2 6 E, j = l,n, satisfying the set of con straints (11.1)—(11.3) is generally called an optimal (feasible) solution of problem (11.1)—(11.3). If the solution x* is such that the inequality L(x") < L(x) holds for any feasible solution x, then x* is called an optimal solution of problem (11.1) - (11.3). Problem (11.1) - (11-3) represents one of the statements of a stochastic program ming problem when E is the conditionally complete vector lattice of measurable func tions defined over a particular probability space. Let Y be a strip generated by the element sup \atJ\ = a, Z = Yd If at least hi
one of the elements [Z]cj, j = l,n, is not nonnegative or [Z] b, ^ 0 for at least one i = l,m, then problem (11.1) - (11.3) obviously has no optimal solution, i.e. either the objective function L(x) is not bounded below on the set of feasible solutions or the set of feasible solutions to this problem is empty; otherwise we need to examine a problem of the form
M*) = £ M W - + i n f 3=1
(11.4)
11.5. LINEAR OPTIMIZATION
IN VECTOR
LATTICES
255
n
£ [ Y K * J = [Y]&1, x3eE+f]Y,
2 = T7^
(11.5)
j = T~Ti
(11.6)
Excluding the trivial case from consideration, we shall examine problem (11.1)— (11.3) on the assumption that the strip generated by the element a = sup \atJ\ coin ,J cides with 5. In the space CX(Q), problem (11.1) - (11.3) can be written as LQ(x) = '£cJ(t)xJ(t)^m{
(11.7)
J=I
n
Y,atl(t)x:(t)
= b,(t),
i = T^t
xjffy > o, zj{t) e Coo(), i = T ^
(11.8)
(n.9)
Let us denote by T the set of all points t £ Q, where all functions a,j(t), &i(i), j M (* = li m i J = 1> n ) assume finite values. T is a nonempty, open, and everywhere dense set. For every fixed t = r £ T we consider a numerical linear programming problem c
71
L T (x) = ^ c ; ( r ) y ; - ^ i n f
(11.10)
n
£ay(T)y,- = fii(i-), i = I~r^
(11-11)
3=1
fo-eiii,
j = T~^
(n.12)
Problem (11.1) - (11.3) will be denoted by I\ problem (11.4) - (11.6) by T{Y), problem (11.7) - (11.9) by V{Q), problem (11.10) - (11.12) by T(r). If Y = C„{E) is a strip in Coo(Q), then problem T(Y) is denoted by T(E). We shall now consider the set S" = {x: x = ( I 1 , I 2 , . . . , I „ ) , £j £ E, j = l , n } . Since the element multiplication operation introduced in E transforms E to a ring and the product Ax = [Xxi, Ax2,. . . , Axn) £ E n can be defined for arbi trary x £ E", A £ E n , the naturally linearized space E™ can be regarded as a vector space over the ring E. When E = C^Q) we have a vector space E n = C^(Q) over the ring Coo(<2). DEFINITION 11.1 The vectors x\x2,...,xk £ E" are called linearly independent in a (nonzero) strip Y of the conditionally complete vector lattice S if the fact that A[, A 2 , . . . , A*, £ Y, \\xl + A2x2 + - ■ - + \kXk = 0 implies that all elements Al5 A 2 ,..., Xk are equal to zero; otherwise these vectors are called linearly dependent in the strip Y In the case E = Rl his definition coincides with the conventional definition of linear independence of vectors in R1
256
CHAPTER 11. OPTIMIZATION
PROBLEMS IN FUNCTION
SPACES
Remark 11.1. If the elements A], A2,. . . , A& in Definition 11.1 are taken to be re stricted, then we get the definition of linear independence in a strip which is equivalent to this one. R e m a r k 11.2. If some vectors from S" are linearly independent in Y, then they are also linearly independent in every nonzero strip Z C Y, but may prove to be linearly dependent in some strip containing Y If, however, these vectors are linearly dependent in Y, then they may prove to be linearly independent in some strip Z C Y Further, whenever we are dealing with linear dependence or linear independence of vectors in a particular strip, this strip as taken to be nonzero. By Remark 11.2, some difficulties may arise in the investigation of vectors from S" which are linearly independent in a strip. We shall now show that these difficulties can be avoided. DEFINITION 11.2 Let a, £ E, i £ J, and let e* be a trace of the element a,. The strip V C H is called an indicator strip of the set of elements a*, i £ ■/, if for each of these elements the projection of [Y] e, either coincides with [Y] 1 or is equal to zero, and for every strip Z strictly containing Y there is an element a3, j £ J, for which 0 < [Z]e} < [Z]l. Lemma 11.1. For a finite set composed of n elements of the conditionally complete vector lattice S there exists a unique finite decomposition of E into indicator strips of this set which is made up of at most 2r strips. Proof. In the case r = 1 the desired decomposition is made up of the strips generated by the elements ej, 1 — e\. In the general case, the proof is obtained trivially, by induction. We shall now consider the set S™ o f m x n matrices whose elements are contained in the conditionally complete vector lattice E. DEFINITION 11.3 Let A £ S™ and let MA C E be the finite set of all possible mi nors of this matrix. The indicator strip of the set MA is called the matrix A-acceptable strip, while the decomposition of space E generated by indicator strips of this set is called the matrix A-acceptable decomposition. Now, if Y is the strip which is acceptable to matrix A, then the trace of every nonzero minor of matrix [Y] A = ||[Y] a;j|| (this notation will be retained in the fol lowing) coincides with [Y] 1. DEFINITION 11.4 Let Y be a strip which is acceptable to matrix A £ 5™. The maximum number of the rows in matrix A which are linearly independent in the strip Y is called the rank of A in the strip Y Lemma 11.2. In order that the rows (columns) of matrix A £ S™ be linearly inde pendent in the strip Y C E, it is necessary and sufficient that the determinant trace of matrix [Y] A, i.e. the trace of element det [Y] A C E coincides with [Y] 1 . Lemma 11.3. The rank of matrix A £ S™ in the acceptable strip Y is equal to the higher order of nonzero minors in matrix [Y] A. The above lemmas are proved by employing elementary means.
11.5. LINEAR OPTIMIZATION
IN VECTOR
LATTICES
257
R e m a r k 11.3. If a nonzero strip Z is contained in a matrix A-acceptable strip Y, then we are dealing with the rank of matrix A in the strip Z. In this case the rank of matrix A in Y coincides with the rank of A in Z. L e m m a 11.4. Let Y he a strip acceptable to the matrix A G E™. If some rows (columns) of the matrix A are linearly dependent in Y, then they are also linearly dependent in any strip Z C Y Proof. In fact, if some rows of the matrix A were linearly independent in a nonzero strip Z C V and linearly dependent in Y, then by Lemma 11.2 there would be a nonze ro minor in those rows whose trace would be smaller than [Z] 1. From the definition of the matrix-acceptable row it might follow that the trace of this minor coincides with [Y] 1, whence follows the linear independence of the rows involved in the strip Y In addition to Lemma 11.4, we may note that if some rows of the matrix A are linearly dependent in the strip Z, then they are linearly dependent in any strip Y C Z. L e m m a 11.5. Let x\t), x 2 (t),.. ., xk(t) G C£,(Q), and let S be a set of points t G Q in each of which all the components of these vector functions are finite. The vectors x 1 (i),x 2 (i),. . . , xk(t) are linearly independent in the strip Y = CX(E) if and only if in the points r of an open set P C S f]E which is everywhere dense in E the vectors I 1 ( T ) , S 2 ( T ) , . . . , x (T) G Rn are linearly independent. Proof. Let B(t) be a matrix whose rows are vectors x'(t), x2(t),.. ., x (t). Suppose these vectors are linearly independent in the strip Y = C 00 (£'). We decompose the Y regarded as an independent conditionally complete vector lattice into the strips which are acceptable to the matrix [Y] B(t). Let z = C^Ei) C Y be one of such strips. Then in the matrix [Z] B(t) there is a minor A(i) of the highest possible order whose trace coincides with the characteristic function of the set E\, i.e. with unity of the conditionally complete vector lattice Z. Hence A(£) ^ 0 in every point of an open set which is everywhere dense in Ey, whence (in view of the arbitrary choice of an acceptable strip Z) follows the necessity part of this lemma. The sufficiency part is obvious and weaker constraints can be imposed on the set P, namely: it is sufficient that the P be everywhere dense in E. Lemma 11.6. Let xl(t), x\t),... ,xk(t) G C™(<2) and let S be a set of points t G Q in each of which all the components of these vector functions are finite. The vectors xl(t),x2(t),. .., xk{t) are linearly dependent in the strip CX{E) = Y if and only if there is a nonzero strip C „ (Ei) C Y such that the vectors x 1 ( r ) , x 2 ( r ) , . . . ,i*(r) G Rn are linearly dependent in each point of an open set P C Sf]Ei which is everywhere dense in E\. This statement follows from the preceding lemma. L e m m a 11.7. Let A{t) € (CM(Q))™ and let Y = C^E) be a strip which is ac ceptable to the matrix A(t). If some rows (columns) of the matrix A(r) are linearly o f tile m a t r l x independent in at least one (§QW&rfqfaefr'M^BHaf1^ elements MT)
258
CHAPTER
11.
OPTIMIZATION
are finite, then these rows (columns) the strip Y
PROBLEMS
of the matrix
A(r)
IN FUNCTION
SPACES.
are linearly independent
in
Proof. We may say t h a t m = n and all rows (columns) of the m a t r i x A(r) are linearly independent. Let A(t) = det A ( i ) . By the definition of t h e function product in Coo(Q), det A(r) = A ( r ) / 0. Therefore [Y] A ( i ) ^ 0, i.e. t h e trace of t h e element [Y] A ( i ) is equal to unity of the space C00(E), and t h e validity of this l e m m a follows from Lemma 11.2. L e m m a 1 1 . 8 . The rank of the matrix A(t) in its acceptable strip Cao(E) coincides with the rank of each of numerical matrices A(r), where r belongs to an open set which is everywhere dense in E. Conversely, if in each point T of such a set the ranks of the matrices A(T) coincide, then the rank of the matrix A(t) in C^E) is equal to the rank of any of them. R e m a r k 1 1 . 4 . If some rows (columns) of the matrix A(r) are linearly dependent, then this does not imply linear dependence of these rows (columns) of the matrix A(t) in the strip Y R e m a r k 1 1 . 5 . Using Lemma plicitly relaxed.
8, the sufficiency
conditions
of Lemma
5 can be ex
T h e notion of a matrix-acceptable decomposition of space makes it possible to define the inverse of the square matrix in much t h e same way as the inverse element is defined. Let A 6 S™ and let Y be a strip acceptable to t h e m a t r i x A. We define the matrix A - 1 g S™ computing its projections onto the A-acceptable strips Y as follows: if det [Y] A = 0 , then we take the matrix [Y] A to be zero; if, however, t h e element trace det [Y] A is equal to [Y] 1 = e, then [Y] A - 1 is determined from t h e relations [Y] A-1-
[Y]A=||ey||,
where en = e, e^ = 0 with i ^ j (i,j = l , n ) and t h e m a t r i x multiplication is determined routinely. T h e matrix A - 1 is uniquely determined by
A-1
= £[YM-\ y y
the inverse of the zero matrix is a zero matrix by definition. T h e way in which the elements (ijj of the inverse matrix are computed is the same as in t h e numerical case: aJJ = A3i ■ (det A ) " 1 , where A 3 ; is the cofactor of a suitable element in t h e original m a t r i x . From the def inition of t h e inverse matrix it follows t h a t ( A - 1 ) - 1 coincides with A if and only if the element trace det A is equal to unity. More precisely,
(A-T' = IYM.
11.5. LINEAR OPTIMIZATION
IN VECTOR
LATTICES
259
where Y is a strip generated by the element det A. The following relations are also true: [Y]A-> =
([Y]A)-\
1
det A" = (det A ) - . 1
In the case m = n = 1, A" is simply the inverse of the element A £ S. R e m a r k 11.6. Let A(t), A~l{t) £ {C^Q})^ and let CM(E) be a strip generated by the element det A(t) ^ 0. Then at each point r of an open set which is everywhere dense in E a nonsingular numerical matrix A(T) is defined and its inverse coincides with A - ' ( T ) .
To look into problem T in detail, we have to study the constraint system (11.2)(11.3) for this problem. A preliminary stage for such consideration is provided by investigation of system (11.2) 5.1 or, what is the same, the system AxT = B,
(11-13)
where A € E™, B £ E™, m < n; T indicates transposition. In the space C^Q), system (13) becomes A{t)xT(t)
= B(t).
(11.14)
Denote the augmented matrix systems (11.13) and (11.14) by A and A(t), respec tively. Denote by S the set of points t £ Q in which all elements of the matrix A(t) are finite. S is an open, everywhere-dense set. For every fixed t = r £ S we consider a numerical system of equations A(T)y7 = B(T).
(11.15)
We also consider a system of the form [Y]AxT = [Y]B,
(11.16)
where Y is a strip of the space E. Here the solution of system (11.16) is interpreted to mean the vector i £ 5 ° whose components are contained in the strip Y The solution sets for the systems (11.13), (11.14), (11.15), (11.16) are denoted by w, u(Q), <^(T), U>(Y), respectively.
Before discussing the properties of solutions to the systems (11.13) - (11.16) we provide an auxiliary definition. DEFINITION 11.5 Let YUY2,. . . , Yk be pointwise disjunctive strips of the condi tionally complete vector lattice 5. The set k
x G E: x = D i j , X{ £ Yt, i = 1, k
260
CHAPTER
11. OPTIMIZATION
PROBLEMS
IN FUNCTION
SPACES
is called t h e union of strips Yi,Y2,. . . , Yk and is designated S(Yx, Y%,. . . , Yk). It is easy to see t h a t S(Yi, Y2,. . ., Yk) forms a strip in 5 . If Y, = Coa{Ei), i = 1, k, t h e n S(Y1,Y2,. ■ ■ ,Yk) coincides with t h e strip C^^Ei). i
Note some of t h e simplest properties of t h e solution sets of t h e systems being discussed. L e m m a 1 1 . 9 . Let as1 6 u(Yi), i = l,k, and let the strips Yj, Y2,..., Yk be disjunctive. IfY — S{Y%, Y2,. . . , Yk), then the vector x = x1 + x2 + is contained
Lemma
in
pointwise
\-xk
OJ(Y).
1 1 . 1 0 . Let x 6 w, and let Y be a strip in H. Then
[Y] x = ([Y]zi,
[Y]*,,...,|Y]«,) gw(y). L e m m a 1 1 . 1 1 . Let x1, x2,..., xk 6 u>, and let Y\, Y2,..., Yk be the strips a decomposition of the conditionally complete vector lattice S, Then
forming
x = [Y] j x 1 + [Y] 2x2 + ■ ■ ■ + [Y] kxk G w. L e m m a 1 1 . 1 2 . Let x(t) 6 C^Q). Now x(t) £ w(Q) if and only if for every point T of an open everywhere dense set X(T) £ cj(r). Note t h a t t h e sufficient conditions of t h e last s t a t e m e n t can b e relaxed: it is sufficient t h a t t h e inclusion X(T) £ w ( r ) hold for all points r 6 Q except t h e set of first category. L e m m a 1 1 . 1 3 . If the matrix of system (11-1) is square and the element trace det A = A is equaj to unity, then the solution of this system is unique and equals x = A~lB, in which case Cramer's formulas are applicable to computation of x. Note t h a t if t h e trace A is less t h a n unity, b u t is different from zero, t h e n t h e vector A~1B is a solution of system (11.4), where Y is a strip generated by t h e element A . T h e o r e m 1 . 1 . System (11.1) is consistent if and only if the ranks of matrices A coincide in every strip which is acceptable to the matrix A.
A and
Proof'will be carried out for t h e case S = CCX1(Q). T h e s t a t e m e n t necessity follows from Lemma 11.4 and Corollary t o Lemma 11.7. We shall prove t h e sufficiency of conditions set forth in this theorem. For simplicity, we assume t h a t t h e entire space E is a strip which is acceptable to t h e matrix A(t). Let t h e ranks of matrices A(i), A[t) in H b e equal t o each other and equal t o r. Then, in the m a t r i x A(t), we may select a nonzero minor of higher order r located in rows i1:i2,... ,ir and columns jl,J2, ■ ■ ■ , j r . Setting in (14) Xj(t) = 0 for j £ {ji,ji, ■ ■ ■ ,jr}, and discarding t h e equations with numbers other t h a n i\,i2,... ,ir, we obtain a system of r equations in r unknowns. T h e determinant trace of this system fs.equal to unity; therefore t h e resulting svstem
Copyrigntea Material
6 J
11.5. LINEAR OPTIMIZATION
IN VECTOR
LATTICES
261
has a unique solution (Lemma 11.5), thereby defining the solution x(t) of the system composed of equations iui2,...,ir. At the points of an open everywhere dense set, however, the numerical vectors X(T) are obviously the solutions of respective systems of the form (11.15). Hence x(t) satisfies all equations of system (11.14), thereby proving the theorem. Theorem 11.2. u>(Q) / 0 if and only if for every point of an open everywhere dense set P U>{T) ^ 0. Proof. The necessity part of this theorem has already been established (Lemma 4). The sufficiency follows from the preceding theorem and Corollary to Lemma 11.7. Remark 11.7. The set P from Theorem 11.2 can be selected in such a way that, in each of its points r, the rank of matrix A(T) is equal to that of A(T) and coincides with the ranks of A(t) and A(t) in the A(t)-acceptable strip Coo(E) such that r 6 E. Theorem 11.2 can be strengthened in its sufficiency part. Theorem 11.3. Let Y{ = C00(El), i = l,k, be the strips of a decomposition which is acceptable to matrix A(t), and suppose that in the strip Yt the matrix A(t) has the rank equal to r;, i = l,k. Let r, be a certain point from Etf]S, i = 1,/c. If the systems (15) with r = T, have an augmented matrix of rank rt, i = 1,/:, and are consistent, then u(Q) ^ 0. Proof. From Lemma 11.7 it follows that, in each of the strips Ylt the ranks of the matrices A(t), A(t) coincide, and it only remains for us to apply Theorem 11.1. Thus, the consistency of system (11.14) depends on that of finitely many numerical systems of the form (11.15). We now assume that system (11.13) is consistent. Let Yj, Y2,. .., Yj be the strips acceptable to the augmented matrix of system (11.13). Consider k systems of the form [Ys]Axr
= [YS]B,
s = l~k.
(11.17)
Let the rank of the matrix A in Ys be equal to r s , 5 = 1, k. To solve system (11.13), we may then restrict our consideration to any rs equations, in which the rows of coefficients with the knowns are linearly independent in Ys, and discard the other equations. Selecting in the thus obtained system of rs linearly Vs-independent coef ficient columns with the unknowns designated by numbers ji,j%, ■ ■ ■ ,jr, and setting Xj = 0 for j £ \j\ ,2i% ■ ■ • )ir,}i w e arrive at the system of rs equations in rs unknowns, whence we obtain a solution of system (11.17). If the components of the obtained solution vector are designated by numbers ji,J2, ■ ■ ■ > jP (p < r») other than zero, then we denote it by x(Ys; ji,j%,.. -,jp). In this case, x(Ys; 0) = 0. Let us sum up the re sulting solutions of the systems (11.17) which have the same collections of indices ii,j2i ■ ■ ■ ,JP- The resulting vector
x(Y;j1,j2,...,j,)
= 'Ex(Y.iJuJa,...,j,),
(11.18)
262
CHAPTER
11. OPTIMIZATION
PROBLEMS
IN FUNCTION
where t h e strip Y is t h e union of some of t h e strips Ys, is called Cramer's system (13). Note some properties of Cramer's vector.
SPACES
vector of
1. Each of t h e vectors x(Ys; j i , j j , . . . ,jp) is Cramer's vector and, simultaneously, a solution of t h e corresponding system (11.17). 2. x(Y;
ji,J2,...J,)€u>{Y).
3. x(Y; ji,js,.
.. ,j9) 6 to if and only if Y = S.
4. Each component of Cramer's vector (11.18) designated by number j £ {ji,jt, jp} has a trace which is equal to [Y] 1 .
• • ■,
5. Let E = Coo(Q), and suppose Cramer's vector of system (11.14) becomes xiCooiE);
ji,J2,.
. .,ip) =
x(t).
T h e n it is possible to select an open everywhere-dense set M such t h a t for every r 6 M x(r) £ W ( T ) and t h e components of vector X(T) enumerated as J G {ji»j2i • • •, h) are nonzero. 6. If x(Yl; ii, i2,. ■ ■, if),. . ., x(Yq; j\,j%, ■ ■ ■ , jh) are C r a m e r ' s vectors of (11.13), t h e n S is t h e union of t h e strips Y\, Y%,. . . , Yq. Note t h a t , for t h e numerical system of linear equations (11.15) o f r a n k r , Cramer's vector is obtained as a solution to a subsystem with a square nonsingular matrix of t h e same rank which is a submatrix of A(T). T h e o r e m 1 1 . 4 . Let tu(Q) ^ rem 11.2 and Remark 11.7. If x(t) 6 u(Q) such that x(r) x(t) € ui(Q) such that for any
0 and let the set P satisfy the conditions of Theo y £ u(r) for some r € P, then there exists a solution = y. Conversely, it is possible to select a solution r € P x(r) £ t j ( r ) .
Proof. T h e first part of this theorem is obvious. To prove t h e second statement of this theorem, it suffices to set E = CX)(Q) and to take x(t) t o be t h e sum of Cramer's vectors for each of t h e systems (11.17). T h e o r e m 1 1 . 5 . The conditions of the preceding theorem provide an open every where dense set P0 C P such that if x(Coo(E); ji,jt,...,jp) = x(t) is Cramer's vector of system (11.14), then for r 6 P0 X(T) is Cramer's vector of system (11.15) of the form y(Rx] juj-2, ■ ■ ■ ,jp). Conversely, if y = y(Ru, j l t j 2 , • ■ •, jp) is Cramer's vec tor of system (11.15), T 6 Po, then there exists Cramer's vector x(t) = x(C00(E); ji,J2, ■ ■ ■ ,jp) of system (11.14) such that X(T) — y. Proof. T h e first statement of this theorem is obvious. To prove t h e second state ment, it should be noted t h a t a guarantee is provided to ensure existence of an open everywhere dense set in whose p o i n t s , T the. following condition holds: if y = ytR1: y yy
Copyrighted Material
'
11.5. LINEAR OPTIMIZATION
IN VECTOR LATTICES
263
»i.*»,■••,-Si) is an arbitrary Cramer's vector of system (11.15), then it is possible to select Cramer's vector of system (11.14) x{t) = x(Cx{E); j u j 2 , . . . , jp) such that {ji,Sa,...,sj.} C {ji,j2, ■ ■ ■ ,J P } and X(T) = y. But in this case {a\iB%%.. .,$k] i= {jiiisi ■ ■ • > jp} some nonzero variables of the vector x(t) vanish at the point r. Since the trace of each nonzero variable is equal to unity of the space C^{E) and the num ber of Cramer's vectors is finite, the set of such points is nowhere dense, which proves this theorem. We shall now consider the set of feasible solutions to problem T. For brevity, the sets of feasible solutions to the problems T, T(Y), T(Q), T{E), T(r) will be denoted by n, w(Y), n(Q), n(E), 7r(r), respectively. The set n is convex over E, i.e. for any x,y £ 7r, a £ E, 0 < a < 1 ax + (1 — a)y £ n. The n is closed with respect to (o)-convergence in that if a. se quence x' = (x\,x'2,...,x'n) £ IT is such that (o)-limx* = xt, j = I7», then x = ( x i , z 2 l . ■ ■ ,xn) e n. In what follows we may need some auxiliary results. DEFINITION 11.6 The nonzero strip Y of the space E is called a comparability strip of elements Xi,x2, ■ ■ ■ ,Xk G E if the elements [Y] i^, j = l,k, are pairwise comparable. L e m m a 11.14. For a finite set of k elements of the space E tiiere exists a. finite decomposition of E into comparability strips of these elements which consist of at most fc! strips. Proof. For the pair of elements ai,a 2 £ E the required decomposition may be constructed as follows. Let Y be a strip generated by the element (ai —a 2 ) + , Z = Yd Then [Y] a3 > [Y]a 2 , [Z]ai < [Z]a2. We have thus constructed the decomposition. In the general case, the statement of this lemma is obtained by induction. L e m m a 11.15. Let xit x2, ■ ■ ■, Zyt S E and let in some nonzero strip Y C E, [Y] Xj 6 S + , j — l,k. Then there exists a strip Z, Z C Y such that [Z]xj £ E + , j = 1,fc,and for any strip K strictly containing Z it is possible to select an element Xj, j = l,h such that the projection [V]x, is zero-incomparable or [V]x} < 0. Proof. We correlate each element x3 with the strip Y} defined as follows: if Z3 is a strip generated by the element (XJ)-, then Y3 = Zf. The intersection of all thus obtained strips Yj is exactly Z. Note that the set Z is determined uniquely. DEFINITION 11.17 The nonzero strip Z described in Lemma 11.14 is called a positiveness strip for the elements xi,x2,... ,2fc € S. If x = (xi, i2> • • ■ , xn) £ —" and there exists a positiveness strip of elements xu x2,.. . ,xn, then we call it a positiveness strip of vector x. If z, y £ E and in any nonzero strip Y C E [Y] i > [Y] y, then we write s > y, The notation [Z] i > [Z] y means that in any nonzero strip Y C Z [x] > [y]. A similar interpretation is given to the notation x < y, [Z] r -C [Z] y. z » 0 if and only if x > 0 and ex = 1. Note that the inequality x{t)
264
CHAPTER
11.
OPTIMIZATION
PROBLEMS
IN FUNCTION
SPACES
Lemma 11.1 lead us to the following result. L e m m a 1 1 . 1 6 . Let Xj, x2,. .. ,xt € S. There exists a finite decomposition of S into comparability strips of elements X\,x2,... , xjt such that for any pair of elements Xi, Xj and any strip Y of this decomposition one of the following relations holds: either [Y] x, » [Y] x3 or [Y] x, « [Y] xj or [Y] x, = [Y] x3. T h e following s t a t e m e n t is the consequence of Lemma 11.16. L e m m a 1 1 . 1 7 . Let Xi,x2,.. . ,Xk 6 5 . There exists a finite decomposition of spaceE such that in each of its strips Y the elements [V]xj, j = l,k are zero-comparable, and it follows from [Y]xj / 0 that |[Y]x_,| ; » 0 . The number of strips in this decomposition does not exceed 3 . D E F I N I T I O N 11.8 T h e essential positiveness strip of elements x±, x2, ■ ■ ■, Xk € S is taken to be a m a x i m u m (in inclusion) strip Y, where the following conditions hold: for every xJt j = l,k, either [Y] x3 = 0 or [Y] x3 S> 0 . If x = (xltx2, ■ ■ . ,xn) € S™ and Y is the essential positiveness strip of elements Xi,x2,... ,xn then Y is called an essential positiveness strip of vector x. R e m a r k 1 1 . 7 . In order that there be an essential positiveness strip of the functions Xj(t) £ CoolQ), j = l,k it is sufficient that the strict inequalities XJ(T) > 0, j = l,k, hold in at least one point r € QD E F I N I T I O N 11.9 Let x = x(Y; ji,j2,... ,jp) be Cramer's vector of the system of equations appearing in the constraint set of problem F. If there exists an essential positiveness strip Z C Y of the vector x, then the vector [Z] x = x(Z) is called a base of the set n. It follows from the definition of the base of the set n t h a t t h e vectors of coefficients for unknowns in t h e constraint set of problem V are linearly independent in the strip Z. In general, the base x(Z) is not a feasible solution of problem T, but x(Z) 6 n{Z). In the case S = Ri1 the base is simply the basic solution of numerical LP problem. If in Definition 11.9 Cramer's vector x is zero, i.e. x = x(Y\ 0), t h e n x = x(Y) is called a zero base of the set n. T h e o r e m 1 1 . 6 . Let 7r ^ 0. Then there exist the bases xt(Yj), j = 1, k, i = 1, lj, and the sets IT, while the union of strips Y\, Y2,..., Yi coincides with E. Proof. Let E = COD(Q) and suppose the system of equations (11.8) in 5.1 has no Cramer's zero vector, for if x(Coo{E); 0) is Cramer's vector of system (11.8), then it is one of the bases of the set n(Q), and the problem reduces to V(Q \ E). Let x(t) £ TT(Q). For each point of an open everywhere dense set the problem T(r) then has a feasible solution. Therefore, each of such problems T ( r ) has at least one basic solution yT, in which case YT is simultaneously Cramer's vector of system (11.11). Suppose PQ is the open everywhere dense set satisfying the conditions of Theorem 11.5 and constructed for the systems (11.8) and (11.11). Since for any r € PQ for which 7T(T) ^ 0 for every basic solution (Cramer's vector) yT = yiR1; ji,J2>---,jp) °f
system (11.11) there is
Cr&r
^0sp^f^}ex^^r)^^^^
■ ■ ■ >h) = z ' M
such
that
11.5. LINEAR OPTIMIZATION
IN VECTOR
LATTICES
265
*(T) = V, r G fifl-Po, then there exists an essential positiveness strip of the vector £"■(*), i.e. there exists the base of the set K(Q) (Remark 11.7). The second statement of the theorem follows from the fact that the set {r £ P 0 :7T(T) ^ 0} is everywhere dense. X
Theorem 11.7. The set n(Q) is nonempty if and only if it is possible to select an open everywhere dense set P with TT(T) / 0 in each point r of this set. Proof. The necessity of this statement is obvious. The sufficiency was essentially obtained in the proof of the preceding theorem, because from 7T(T) / 8, T e P, we may obtain the existence of bases for the set n(Q) which permit an easy construction of a feasible solution to problem F(Q). Note that Theorem 11.7 provides the existence condition for solutions to systems (11.8) - (11.9) and (11.11) - (11.12) without any reference to the problems V(Q), r(r). The future references to this problem take account of this fact. Furthermore, let A3 be the vector of coefficients for Xj in the constraint set of problem F. DEFINITION 11.10 The feasible solution x = ( I 1 , I 2 , . . . , I „ ) of problem V is called basic if there exists a decomposition of the conditionally complete vector lat tice E into essential positiveness strips Y of vector x such that the vectors A3 corre sponding to the nonzero [Y] Xj are linearly independent in S. The nonzero variables of a feasible solution x for which [Y] x3 2> 0 is called basic in the strip Y In this case, we require that for different strips of the above decomposition there be different collections of variables of the solution x which are basic in these strips. This definition suggests that for each basic solution the above decomposition of space is unique and finite. In the case H = R1 we have a, definition of the basic solution of a numerical problem. Theorem 11.8. The set of basic solutions to problem V coincides with the set of feasible solutions of the form
E^'M). hi
where x'(Yj) is the base of the set of feasible solutions etJ £G(Yj),
(1L19) J2e>j = 1, J = 1)^>
i = uj. Proof. The fact that solutions of the form (11.19) are basic follows from the defini tion of the base of the feasible solution set, and from the definition of a basic solution to problem T. Conversely, let x be a basic solution to problem T and let Y\, Y2, , Yjt be the strips of the decomposition described in Definition 11.10. The vector [Yj] x then can be obtained as a projection of some Cramer's vector onto the strip Yj, and hence as a projection of some base of the set n. But the operation of projection onto the strip Y) is equivalent to multiplication by the unity element [Yj] 1, which proves the theorem. Corollary. Suppose the set ir is nonempty and has no zero base. The n then contains at least one basic solution.
266
CHAPTER
11. OPTIMIZATION
PROBLEMS IN FUNCTION
SPACES
T h e o r e m 11.9. Let ir(Q) / 0. Then there exists an open everywhere dense set P such that if x(t) is the basic solution of problem r(Q), then for any r 6 P X{T) is the basic solution of problem T(T), and if y is an arbitrary basic solution to problem T(r), r 6 P then it is possible to select a basic solution x(t) of problem V(Q) such that X(T) = y. In this case the numbers of basic variables of the solution X(T) coincide with the numbers of basic variables of the solution x(t) in a strip Coo(E) such that T £ E. The statement follows from Theorems 11.3 and 11.5. Before formulating the next theorem, we will make one remark. If the set M c ? , then by the extreme point of this set is usually meant the point which cannot be represented as a convex combination of two other points of this set with positive real coefficients. If M is the set which is convex over E, then it would appear natural that the elements of the conditionally complete vector lattice 5 would be considered in place of real coefficients. This correction, however, would have made the above definition incorrect. Indeed, suppose the elements x,y £ M fit such a definition of an extreme point. But then x = e\zl + e-^z2, where ei,e2 £ (3(3), £i + e2 — 1> z1 = C]X + e2y £ M, z2 = e2X + e\y 6 M, which is a contradiction to the definition. T h e o r e m 11.10. The set of basic solutions to problem T coincides with the collection of extreme points of the set jr. The proof is carried out along the same lines as in the numerical case (see, e.g. [1]) using Lemma 11.1. R e m a r k 11.8. If the set IT has a zero base x(Y), Z = Yd, then the collection of extreme points of the set n is exhausted by the extreme points of the set w(Z), i.e. the set n(Y) has a unique extreme point x(Y). T h e o r e m 11.11. The set n(Q) is bounded if and only if each of the sets v(r) is bounded in the points of an open everywhere dense set. The theorem is straightforward. T h e o r e m 11.12. If the set of feasible solutions to problem T is bounded, then any feasible solution x can be represented as
* = EV(l)-)i
(H-20)
where x'(Yj) is the base of a feasible solution set', Ay £ S + , 52 A,_, = 1, j = l,fc, >j
Proof. If the set of feasible solutions to problem V(Q) is bounded (we consider the case E = Coo(Q)), then, as is apparent from the preceding theorem, each of the sets 7r(r) is bounded in t^e points ofAn,open everywhere dense set P C Q. If
11.5. LINEAR OPTIMIZATION
IN VECTOR LATTICES
267
x(t) 6 x[Q), then the set P can be chosen such that in its points x(r) g TT(T), and hence the system J UijX'^r) = :X(T) C(T) E^«jZ' ("0 =
E, >,3 '•0
E^- = i Pi) u.tJ > 0,
jj=TJc, = ITfc,
i, = = 17, ~i
J
where x' (r) designate the bases of the set 7r(Q), is solvable for each r from the open everywhere dense set. Theorem 11.7 implies solvability of the system 3 J2\ x(t) EAy(t)x' (t) ) == x(t) ',3
EAu =- 1 A, > 0, A,,(t) 7 (t)>o,
I77~, ii == I7fc, TJ, i^ ~= IJy
in the unknowns A,_,(t) g Coo(Q), which proves the theorem. Corollary. If the set IT is bounded, then there exists a finite collection of its extreme points x1,!2,. . ., xk (which is generally not determined in a unique way) such that any feasible solution x can be represented as i1 x= = \A,! + A X22xs22 4+ ••• • ■■++\kxk,\kxk, 1x
(11.21)
where A, g H + , t = I7&, A1, A 2 , . . . , A* = 1. if A, g (3(H), t = I7&, then aii extreme points of the set IT are defined by (11.21). Remark 11.9. The elements Aj, A2,. . . , Xk in (21) cannot be replaced by real num bers, i.e. the bounded set IT is generally not coincident with the convex hull (with real coefficients) of its extreme points. To be noted also is that the set n has a zero base x(Y), then the set ir(Y), and hence the set IT, is obviously unbounded (except where x(Y) is a unique feasible solution of problem r ( F ) ) . Let IT ■£ 0. Denote by A the set of all feasible solutions of problem T which can be represented as (11.21) (or, what is the same, as (11.2020)). This leads us to the following theorem. Theorem 11.13. Each feasible solution x g IT can be represented as the sum x = u + v, where u g A, while v is a solution to the homogeneous system fitting the con straint set of problem I\ In the case E = R1 the theorem is known. In the general case, the proof is carried out in much the same way as for the preceding theorem. Theorems 11.4, 11.5 and 11.3 lead to the following result.
268
CHAPTER
11. OPTIMIZATION
PROBLEMS IN FUNCTION
SPACES
Theorem 11.14. Ifn(Q) ^ 0, then it is possible to select an open everywhere dense set such that the following condition holds in each point r of this set: if y G TT(T), then there exists a feasible solution x(t) £ n(Q) such that x(r) = y. We shall now consider the existence of optimal solutions to problem T. We denote by A0 the collection of extreme points of the set n. Lemma 11.18. Let n / 0 and let C be a restriction of the objective function L of problem T to the set A0. The function £ achieves on A0 its smallest value, i.e. for some solution x" € A0 and for any x 6 A0 C(x*) < £(x). Proof. Applying Corollary of Theorem 11.12, we consider a finite set of elements £(Xl), i = 1, k (see equation (11.21). By Lemma 11.14, there exists a finite decompo sition of S into comparability strips of elements of this set denoted by Z\, Z%,..., Zq. Since the operator £ is permutable with the operator of projection onto a strip, for each of the strips Zj it is possible to select a feasible solution x'>, j = 1, k such that L{[L^x'>) < L([Zj)xl) for any i = 1 7 1 . The feasible solution x' = [Z,]x"
+[Z2]x'i
+ ... + [Zq]x'<
is contained in A and, by employing Corollary to Theorem 11.12, we may verify the validity of this lemma. Note that if 7r has no zero base, then the solution x* is basic. T h e o r e m 11.15. Problem F is solvable if and only if the set of its feasible solutions is nonempty and the objective function is bounded below on this set. In this case x* 6 Ao is one of the optimal solutions of this problem. Proof. The necessity of the first statement is obvious. The sufficiency and the second statement follow from Theorem 11.13, the preceding lemma, and linearity of L over 2. Corollary. If problem T is solvable and the set of its feasible solutions has no zero base, then among optimal solutions of this problem there is at least one basic solution. Remark 11.10. Let x(Y) be a zero base of the set of feasible solutions to problem F. If problem T(Y) is solvable, then x(Y) is one of its optimal solutions. Theorem 11.16. Problem F(Q) is solvable if and only if each of the problems T(r), where T belongs to an open everywhere dense set, is solvable. Furthermore, if l(t) is the extreme value of the objective function LQ of problem T(Q), then l(r) is the extreme value of the objective function LT of problem F(r). Proof. Let the objective function LQ be bounded below on n(Q) by the element c(t). By Theorem 11.14, it is possible to select an open everywhere dense set in each point r of which y € IT(T) implies the existence of a solution x(t) 6 T ( Q ) such that X(T) = y. In this j;ase we.mav ififlMOfijfhat the coefficient functions of
11.7.
OPTIMIZATION
PROBLEMS
ON VECTOR
LATTICES
281
T h e o r e m 1 1 . 3 9 Let * = {p'r} £ § ( A ) , v = {pT} 6 0 ( A ) , i = 1 , 2 , . . . ,
M Xi,.
=
(x1,...
. . , [y]xnL
,xn)
£ 5 n and let Y be a strip in S.
Denote [y]x =
where [y] is t h e operator of projection onto t h e strip Y.
Pro-
jection is t h u s extended to cover t h e elements from E". T h e set AY = \z £ H" : z = [Y]x, x £ A > is a projection of A onto t h e strip Y D e f i n i t i o n 1 1 . 1 5 . Set tp £ $ ( A ) . T h e m a p ipY : A v —> Y is called a restriction of t h e m a p p i n g of tp onto t h e strip Y if for any x £ A < / J K ( [ V ] I ) = [Y](/?(x) 4 .. 5 We a d o p t t h e notation: $(AY)
= {tpY :
T h e o r e m 1 1 . 4 0 . Let tp £ <£(A), and let Y be a strip in H that is isomorphic to the strip CooiE) in CX(Q), y £ A K ,
¥>(*) = !>*([!«]*)■
( 1L24 )
1=1
T h e o r e m 1 1 . 4 2 . Let tp, tp £ $ ( A ) and suppose the strips Y, Z form an expansion ofE. Denote by IT a map of A into S, i. e., for any x £ A n(x) = c/>r([Y"];r)-|-V'Z([Z]x). 77ien n £ $ ( A ) . Proof. In fact, if V, Z are isomorphic to t h e respective strips Coo(-Ei), C00(E2), tp = {tpT}, i< = {^>r}> t n e n T = {"v}, where 7rT = tpT for r £ £?,-, and nr = t/>T for T £ £2. It can b e readily seen t h a t Theorem 6 can be extended to t h e case of an arbitrary finite set of m a p s . 5 If the stripy is zero, then AY consists of a single element —zero on which tpY takes the value 0.
270
CHAPTER
11. OPTIMIZATION
PROBLEMS
IN FUNCTION
SPACES
Proof. To prove this theorem, it suffices to refer to Theorem 11.13 and note that if at least one of the elements x,y € E is contained in S \ So, then x + y £ S \ So. The next result immediately follows from the formulated theorem. Theorem 11.20. The set of proper feasible solutions to problem F is nonempty if and only if the set of proper feasible solutions contained in A is nonempty. In the following we need one more auxiliary notion. DEFINITION 11.12 Let us consider the function w(A) of a real-valued argument A which is defined for A > 0 and takes values from O(S), in which case w(Aj) < UJ(X2) if Aj < A2 and infw(A) = 0, supw(A) = 1. The strip of the space S generated by A
A
the element 1 — LJ(S), where S is a fixed value of A, is called a S-strip and designated 3(LJ,S). The function w(A) is called the constraint function of the space E. If the space constraint function is defined for all real A, then it may be regarded as a decomposition of the space unity [1], p. 109. We shall list some properties of <5-strips. 1. 3 ( w , 0 ) = 3 ; 3 ( w , + o o ) = { 0 } . 2. 5^ < &2 implies S(w,<Ji) D S.(u>,62)3. u>\(5) < (x)2(5) implies 2(w 1 , S) D E(w2, S). k
_
„
4. f] H(w,:,^) = ZJ(U>,5), where ui(5) = supuj,(5). i=l
i
Lemma 11.19. Let the set M C H be bounded. There exists a constraint function ui(\) of space 3 such that for an arbitrariiy smaJl i > 0 the elements [Y]x, i 6 M, where Y = E(w,J), are contained in So. Proof. Since the set M is bounded, |x| < o for some a > 0 and all x g M. Let e° be the characteristic of the element a. Set w(A) = 1 — e ° , v From the properties of the characteristic it follows that u(t) fits the definition of the constraint function of the space E. Using the inequality [Y A ]a,5)]a < -eal/s. Since the element e\^ (S > 0) is called the unity of a conditionally complete vector lattice S(w, S) = Y, the last inequality implies that the element [Y] a is bounded in Y and especially in E. Since for every x € M |[Y] x\ < [Y] a, then [Y] x G E 0 . Note that if a > 0, then for some 5 = 50 > 0 (and with all 5 < 80) the strip S(w, S) is nonzero. This completes the proof of the lemma. This lemma implies the ty2$fl0tf&{pMaterial
11.5. LINEAR OPTIMIZATION
IN VECTOR LATTICES
271
Theorem 11.21. If the set of feasible (optimal) solutions to problem T is nonempty, then there exists a constraint function u>(\) of space E such that for an arbitrarily small 5 > 0 determining a nonzero S-strip of E(u>, 5) = Y the set of proper feasible (respectively proper optimal) solutions of problem T(Y) is nonempty and all basic solutions of the set n(Y) are proper. The theorems stated below suggest that the solvability of problem T is determined by the behavior of the objective function on the set of proper feasible solutions. Theorem 11.22. Let the set TT0 of proper feasible solutions to problem V be nonempty. If the restriction of the objective function L onto ir achieves on the set n0 its smallest value l0 = L(x°) (x° 6 7r0j, then x° is an optimal solution of problem T. Proof. Suppose that for some x G n L(x) — I < l0. Then there exists a strip Z such that [Z] /0 there exists a nonzero ^-strip V = E(ui, S) such that the solution [Y] x is a proper feasible solution of problem V(Y). For a sufficiently small 5 > 0 Yf]Z = V^ {0}. It is clear that the objective function Lv of problem T(V) being treated on the set of proper feasible solutions achieves on this set the smallest value [V] IQ and [V] x is a proper solution of this problem. Therefore, the following inequality is valid
M[v]*)>[v]/ 0 . But Ly ([V] x) = [V] /c for any x G n, and problem T is solvable. Proof. Let L(x) < c for some solution x which is not proper. Then, evidently, it is possible to select a strip 2 c H such that [Z]L(x) = L z ( [ Z ] x ) « [ Z ] / . Note that for every strip Z [ Z ] i £ K{Z) and for a sufficiently small 5 > 0 there exists a <5-strip of Y = E(OJ,5) such that the vector [Y] x is a proper solution of problem r ( V ) and Yf]Z = V / {0}. But the vector [V] x is a proper feasible solution of r(K) and
[V]L(x) = M[V]x)«[V]/. If y is an arbitrary proper solution of T(Vd), then [V]a; + y = x° is a proper solution of problem T for which the inequality L(x°) > / fails to hold. The resulting contradiction proves the theorem. We shall now turn to dual problems. So far we have dealt with problem T is canonical form. We shall now consider in a conditionally complete vector lattice S problem 5 of the form k
C(x)
= Y^ CJXJ —> inf
272
CHAPTER
11. OPTIMIZATION
PROBLEMS
IN FUNCTION
SPACES
k
= 6;,
z = 1, r
3=1
*
X xj < b,, i i== r r ++ 1,5 2 Ja >3 a-ij 1, s E 3 < 6;, 3=1
l,p, ij === l,p,
Zj *3 €e sS+, +>
pp << kfc
In the space Goo{Q), problem A assumes the form A(Q): k
£:«(z) = Y, CQ(X) At)xj(at*(*) ) -Mnf ->inf E cc3'W 3=1 k
= 177 i= l,r
J^,ai £« 3{t)x3(t) = k(t), %
3=1
^ a j ; , ( i ) x j ( i ) < 6j(t),
E
3= 1 3—1
2j Zj > 0,
1,5 i = r + l,s
pP < < k
j = l,jD, I,7,
At the points of an open everywhere dense set we obtain numerical problems A(r): k ft
£CT{X) T(S)
-►inf -+inf
T = J2cj(-]( = )yj E< T)yj
= 11 33 =
itk
=V
E a y( T )% = 6, 3 = 11 k 1:
E a.,(i")yj
3=1 3= 1
< b, (r),
%>o, %>0,
2 =
ji = =■■^P, l,p,
r
+ 1,5
p < fc
Problems A(V), A(J5) are constructed in much the same way as problems F(F), V(E). Introducing additional variables sj. + 1 ,Xk+2, ■ ■ ■ ,xk+r-l £ S-H a n d replacing vari ables x., which are not related by "nonnegativity" conditions, a difference x'3 — x", where x'},x" £ 5+, we transform as in the numerical case (e.g., see [3], pp. 42-43) problem A into canonical form T. In this case, problems A(Q), A ( T ) , A(Y"), A(E) transform to T(Q), T(r), F(Y), T(E), respectively. It is clear that the set of feasible solutions to one of the problems A is nonempty if and only if the set of feasible so lutions to the corresponding problem T is nonempty and, given the optimal solution of one of the problems A, it is possible to determine in terms of this solution a par ticular optimal solution of the corresponding problem T, and vice versa. In this case, the theorems previously formulated for problem T hold for problem A. For reference convenience, we reformulate some of these theorems.
11.5. LINEAR OPTIMIZATION IN VECTOR LATTICES
273
Theorem 11.24. Problem A(Q) is solvable if and only if each of the problems A(r), where r belongs to an open everywhere dense set, is solvable. Moreover, if l(t) JS an extreme value of the objective function CQ, then 1(T) is an extreme value of the objective function CT. Theorem 11.25. The feasible solution x(t) of problem A(Q) is optimal if and only if X(T) is an optimal solution of problem A(r) at the points of an open everywhere dense set. Note that, in place of the minimization problem which has been considered so far, we may consider a maximization problem and the previously obtained results may have analogs for this case. Denote by A* the dual of problem A: M
M(y) (v) == ^2*>ty« yt -> -> sup sup 3= 1 Ss
a
i = TTP
v
E 'J 3= 1
^ a 1 J j / J = c J ,i-.
E O -J/j = C
3=1
3 ■
= Pj=p+l,k + l, fc
v
y3 ee E+, % s.
i= = r+ r + l,s
By problem A*, we construct A*(Q), A * ( T ) , A*(V), A*{E). The dual of the nu merical problem A(r) coincides with A*(r). This enables one to apply "numerical" duality theorems to the problems A(r), A*(T). It turns out that the duality theory of numerical problems applies to the case of conditionally complete vector lattices. Theorem 11.26. (First theorem of duality) If one of the problems A, A* has an op timal solution, then the other problem is solvable, in which case the extreme values of objective functions coincide. If the objective function in one of the problems is not bounded, then the system of conditions in the other problem is inconsistent. If, however, the system of conditions in one of the problems is inconsistent, then the dual problem either has an unbounded objective function or is also inconsistent. Proof. If one of the pair of dual problems, e.g., A(Q) is solvable and the extreme value of the objective function is equal to /(£), then in the points of an open everywhere dense set the problems A(r) are solvable and the extreme values of the corresponding objective functions are equal to / ( T ) (Theorem 11.24). Then, by the first theorem of duality for numerical problems, the problems A*(r) are solvable and the extreme values of their objective functions are equal to /(r). Theorem 11.24 implies solvability of problem A*(Q), and since the extreme value l*(t) of the objective function of problem A*(Q) is equal to l(r) at the points T of an open everywhere dense set, then /*(£) = l(t). Suppose the objective function of problem A(Q) is unbounded below on the set of feasible solutions. Then the objective function of each of the problems A ( T ) is unbounded below at the points of a nonempty open set; therefore the set of feasible
274
CHAPTER
11.
OPTIMIZATION
PROBLEMS
IN FUNCTION
SPACES.
solutions to problems A ( r ) is empty, and this implies t h e emptiness of t h e set of feasible solutions to problem A*(Q). T h e last s t a t e m e n t of t h e theorem is proved in a similar way. T h e o r e m 1 1 . 2 7 . (Second theorem of duality) In order that the feasible solutions x = (x1) x2,. . . , xk), y = (yi,y2,. . . ,ys) to the pair of dual linear programming problems A, A* be optimal, it is necessary and sufficient that the product of any variable xt or y} of one problem by the difference between the values of the left and right parts of respective constraints in the dual problem is equal to zero. Proof. If x(t), y(t) are respectively optimal solutions of the problems A ( Q ) , A*(Q) t h e n at the points of an open everywhere dense set x(r), y(r) are optimal solutions of the problems A ( r ) , A*(r) and for each of these numerical problems the theorem is known; whence it follows t h a t the theorem applies to our case. Conversely, if x(t), y(t) are feasible solutions to the problems S(Q), A*(Q) for which t h e conditions of the the orem hold, then at the points of an open everywhere dense set a similar condition holds for the feasible solutions X(T), y(r) of the problems A ( T ) , A*(r) The second duality theorem for numerical problems implies the optimality of t h e solutions X(T), y ( r ) , while Theorem 5.7.2 implies the optimality of the solutions x(t), y(t). We shall now formulate some implications of duality theorems. T h e o r e m 1 1 . 2 8 . In order that there be an optimal solution to any one of the dual problems, it is necessary and sufficient that the sets of feasible solutions of these problems are nonempty. T h e o r e m 1 1 . 2 9 . In order that a feasible solution x of a problem A be optimal, it is necessary and sufficient that there exists a feasible solution y of problem A* such t h a t C(x) = A4(y). In this case y is an optimal solution of problem A*. T h e o r e m 1 1 . 3 0 . In order that a feasible solution x of problem A be optimal, it is necessary and sufficient that there exists a feasible solution y of the dual problem A* such that for every strict inequality in the conditions of this problem the dual problem has an equality in the adjoint condition.
11.6
Matrix Games in Conditionally Complete Vector Lattices
Let E be a conditionally (not necessarily universally) complete vector lattice with unity and let S be its m a x i m u m extension. We may assume without loss of generality t h a t E C E. We shall consider the game with the payoff matrix A = ||a,j|| € S™. Let us define some forms of players' strategies. T h e pure and the mixed strategy are defined routinely. Let e 6 0 ( H ) , a = (ai,a2,. ■ ■, an) € Rn Denote by cte the vector (a^e, a2e,. .., ane) 6 H". D E F I N I T I O N 11.13
11-6. MATRIX
GAMES IN VECTOR
275
LATTICES
1) —-pure strategy for Player 1 is taken to be a vector (ej, e 2 ,. . . , e m ), e, G G(H), x = l , m , ei + e2 + h e m = 1; 2) FLi-strategy for Player 1 is taken to be a vector from A r " which can be rep resented as ale\ + a 2 e 2 + • • • + ctkek, where a' G Rm is a mixed strategy for Player 1, e, e (3(H), i = TJc, ej + e2 + ■ ■ ■ + ek = 1, 3) ^,-rnixed strategy for Player 1 is taken to be a vector x = (xi,a>2,.. . , z m ) G E m , i i e H + , i = l , m , xi + x2 + ■ ■ ■ + xm = 1. Strategies for Player 2 are defined in a similar way. In the case E = R1 the notions of a mixed strategy, E-mixed strategy and flE-strategy coincide. R e m a r k 11.11. The pure (mixed) strategy a for one of the players can be regarded as E-pure strategy (KE-strategy) Q 1 . The sets of mixed strategies, /JE-strategies and E-mixed strategies for the i-th player are denoted by Vi(R), i'.(RE), f;(S), respectively (i = 1,2). u,(R) C u,(RE) C ".(H), t = l , 2 . The payoff function for Player 1 is given as = ^ayijj/j,
H{x,y)
where x € ^i(E), y G ^ ( E ) . The payoff to Player 2 is —H(x,y). It can be readily seen that for any S-mixed strategies x,y of the players H(x,y) G E. E-mixed strategies i*,y* for respective Players 1 and 2 are called optimal if H(x", y") = sup inf H(x,y) = inf sup i/(z, y). The element H(x*,y") of the conditionally complete vector lattice E is called the value of the game. The above game in which the players select their strategies from the sets ut{R), Vi(KE), Vi(E) will be denoted by II(-R), I I ( i E ) , 11(E), respectively. If Y is a nonzero strip of the conditionally complete vector lattice E, then we denote by U(R; Y) the game \l(R) with the payoff matrix [Y] A = ||[Y] a , J examined in Y as in the independent conditionally complete vector lattice. In the space C^Q) the game 11(E) is designated U(Q). At the points r of an open everywhere dense set P we examine the numerical matrix games U(R,T) with the matrices A(T) = ||O,J(T)||.
We shall now make several observations about /JE-strategies. If /?E-strategies x, y are of the form 1
/
i
2 /
i
i
k
i
x = a e1 + a e2 + ■ ■ ■ + a ek
y = /?V/ + ^e 2 ' + -"+/F<>
276
CHAPTER
11.
then the situation (x,y) (x,y)
OPTIMIZATION 6 Em+n
PROBLEMS
IN FUNCTION
SPACES
can be written as l
+ (a2,b2)e2
= (a\b )er
(ar,br)er,
+ ■■■ +
where (a',fe*) £ / ? m + " is a situation in the space of mixed strategies, e, £ (3(B), i = l , r , ei + e 2 4- ■ ■ • + e r = 1. Indeed, it suffices to decompose E into indicator strips of the set {e[,e'2,... ,e'k, e", e' 2 ',..., e"} denoted by Yi, Y j , . . . , Yr) and set e, = [Yj 1 , t = l , r . Here a ! coincides with one of the strategies a', and 6* coincides with one of the strategies j3' Note that in this case H(x,y)
= H(a\bl)ex
+ H(a\b2)e2
+ ■■■ +
H{ar,br)er.
This fact permits the following interpretation of the choice by players of REstrategies x, y. This choice is equivalent to decomposing the game into a finite number of the games Yl(R;Yt), i = 1, r (Vj is a strip generated by e,, i = l , r ) in each of which Player 1 chooses a mixed strategy a' and Player 2 chooses a mixed strategy b'. One more fact is worthy of notice. If the game being examined has optimal strategies, then, as in the numerical matrix game, by adding to each element of the m a t r i x A a fixed element w and retaining optimal strategies we may change the game value v t o v + w. In this case the value of the payoff function for any E-mixed strategies x, y is equal to H(x, y) + w. We shall present without proof the principal results relative to t h e existence of optimal strategies in the games under examination. T h e o r e m 1 1 . 3 1 . The game 11(E) has a pair of optimal
'E-mixed
strategies.
Note t h a t this theorem can be obtained as a consequence of the following result. T h e o r e m 1 1 . 3 2 . The game 11(E) reduces to a pair of solvable dual LP problems the conditionally complete vector lattice E.
in
T h e o r e m 1 1 . 3 3 . Let x'(t), y'(t) be E-mixed strategies for players in the game 11(E). These strategies are optimal if and only if at the points T of an open everywhere dense set x*(r), y*(r) are optimal strategies for players in U(R;T). T h e o r e m 1 1 . 3 4 . For an arbitrarily small e > 0 and any situation (x, y) in the space of E-mixed strategies there exists a situation (xe,yc) in RE-strategies such that \H(x,y)-H(x%y*)\<el. C o r o l l a r y . Let v be the value of the game 11(E). Then for every e > 0 there a situation (x £ ,y £ ) in R^E-strategies such that
exists
\v-H{x*,y')\<e\. Thus, the game Il(i?S) is "approximating" in t h a t sense for the game 11(E). T h e o r e m 1 1 . 3 5 . Optimal
in the game U(R) exist if and only if
E-pure strategies
s u p i n f a . j = infsupa.-j. .'
3
3
i
1L7.
11.7
OPTIMIZATION
PROBLEMS
ON VECTOR
LATTICES
277
Optimization Problems on Vector Lattices
Representation of vector Lattices as a, Spaces of Functions t h a t are Contuous on a compact Set An i m p o r t a n t property of a complete vector lattice (CVL) with unity is the pos sibility to represent it as a space of continuous functions on some compact set. We shall provide strict formulations. T h e compact set Q is called extremely disconnected if t h e closure of any open set from Q is open-closed (i.e., open and closed at t h e same t i m e ) . Let Q be an extremely disconnected c o m p a c t u m . Denote by C^Q) the set of all continuous real functions t h a t are defined over the c o m p a c t u m Q and can take infinite values on the nowhere dense sets. Here, as in the case of functions with finite values, t h e continuity of functions x{i) at t h e point r £ Q, where X(T) = oo, implies that l i m x ( i ) = oo. We introduce in Caa{Q) the ordering: x(t) > y(t) if x(t) > y(t) on t h e entire Q and x(t) > y(t) at least at one point. By naturally linearizing C00(Q) and introducing a partial ordering, we transform C00(Q) to a vector lattice. T h e following assertion is true: C^Q) is an augmented vector lattice. T h e role of unity in CX(Q) is played by the function 1(2) = 1, and hence t h e collection C(Q) of all bounded functions from Cra(Q) forms a complete vector lattice of bounded elements. We now formulate t h e theorem on representation of a vector lattice as a space of continuous functions. T h e o r e m 1 1 . 3 6 For any complete vector lattice S with unity 1 there exists an extremely disconnected compact set Q such that S is isomorphic to some base Y Coo(Q). Isomorphism can be realized here in such a way that C(Q) C Y and to unity 1 6 E there corresponds a function that is identically equal to unity. It is this kind of isomorphism t h a t will be meant in what follows. Note t h a t t h e compact set Q is invariant with respect to t h e choice of unity in E and is determined up to isomorphism. 2 T h e isomorphic image of E in C^Q) is denoted by £-«,((?) (where the vector lattice E is not augmented and is distinct from E 0 ) . T h e strip Y of one of the vector lattices C[Q), £«,(£?), C « J ( Q ) can be considered to be an independent vector lattice C(B), L^E}, Coo(rJ), respectively (E is an open-closed set from Q\ if E = 0, then the strip Y is zero, i. e. Y = {0}. This fact will be used in notations. For t h e further discussion, we need some auxiliary s t a t e m e n t s (see [1], [2]). Suppose t h a t an arbitrary set of continuous functions x,-(f) G C00(Q), i € J is given on t h e extremely disconnected c o m p a c t u m Q. T h e n 1) t h e pointwise computed function x(t) = supx;(£) is lower semicontinuous; 3 2 A similar representation as a vector lattice Ctxs(Q) can also be obtained for the u-complete vector lattice. Here the compactum Q proves to be
278
CHAPTER 2) if y(t)
11.
OPTIMIZATION
PROBLEMS
IN FUNCTION
SPACES
= s u p i i ( i ) (the supremum is defined in terms of t h e ordering of the J
vector lattice Coo(<2)), then y(t) is t h e smallest of the functions CX(Q) majorizing above x{t), i. e., if z(t) £ Coo(G) and z(t) > x(t) for all t £ Q, t h e n z(t) > y{t) > x{t); 3) t h e set it £ Q '■ y(t) > x(t)\ is of the first category in Q. Functions of Vector Lattice Elements Let 5 be an augmented complete vector lattice with a fixed unity 1, S ( H ) its base, and h the isomorphism of H on CxiQ)If * € H (x = (xi,. . . , i . ) £ S n ) , then an element /is of t h e space Cao(Q) (respectively, (hxi,. . . ,hxn) £ Coo(Q) X Coo(Q) X • • • x Coo(<2)) is denoted by x(t), t € Q. Further, let T be an arbitrary nonempty set in Rn t h a t is naturally topologized, and let S(T) be algebra of all real functions on T which assume finite values. For j f H " , denote
H(x) = {teQ
x(t)eT}.
Denote by A the set of those as for which Q\H(x) is of t h e first category in Q. We associate each point r £ Q with the function (pr 6 S(T). Denote the resultant family of functions by {E is called the m a p generated by the family {fT} if f ° r a n Y s G A there is a set - P ^ ) C H(x) such t h a t A) Q\P(ip) is of the first category in Q; B) if z = f>(x), then for any r £ P(y>) « ( T ) = V 5 r(a;( r ))The collection of maps, thus defined, is denoted by $ ( A ) . 4 R e m a r k 1 1 . 1 2 . If x,y £ A, i J ( l ) = i / ( y ) then t h e sets P(y>) defined for respec tive as, y coincide. Let a = ( o i , . . . , a n ) £ T. Denote by a\ t h e vector ( a i l , . . . , a n l ) £ A. For any a £ T H(a\) = Q, and hence for ant fixed (p £ $ ( A ) the set P(y>) is identical for elements of t h e form o l , a £ T. R e m a r k 1 2 . 1 3 As is seen from the definition, each of t h e functions tp £ $(A) is uniquely determined by the family { ^ T } , T £ Q- { ^ T } and {/ T } define the same function from $ ( A ) if and only if the set {T £ Q : tpT / fT} is of t h e first category. Here we identify [tpT} and {/ T } and write
: {/} € * ( A ) }
is nonempty and contains all Baire's functions on T. T h u s , $ ( A ) is nonempty and is not restricted to functions of t h e form { / } . Trivial example: let Ei,...,Ek be disjoint open-closed sets in Q which cover Q, and let Q, / i , . . . , / j . be pairwise distinct functions from U(T). Then the m a p { y T } determined by t h e condition tpT for r £ Z3; is contained in $ ( A ) . Linearize $ ( A ) setting for a,fl £ R1, if, f £ § ( A ) i £ A (atp + 0f)(x) = af 4
T h e term "map" will be used along with the term "function"
11.7. OPTIMIZATION
PROBLEMS
ON VECTOR
279
LATTICES
if for all x g A tp(x) > f(x). Here tp > f if tp > / and for some x g Af(x). It is clear that tp > f if and only if for any x g A with T £ P(y) H P ( / ) y T > fT (i. e., for y g T y?T(y) > fT(y)). It can be readily seen that $(A) thus becomes a vector lattice. T h e o r e m 11.37. $(A) —an augmented a-complete vector lattice. Proof. Let x g A, tp\ f g $(A), < / , ¥>'(x) = ZA J g J and suppose the index set J is countable. Define the map tp : A —> E as ¥>(x) = sup {x) = y g 5.
(11.22)
For any x g A there exists y>(x), because y'(x) < / ( x ) g E. For each r g ff(x), compute 2(T)
=(ypT(x(r)).
l
By the definition of functions ip z(t) for t g n P(
The function y*(t) is lower semicontinuous on Q, y*(t) < y(t) and the set Q0 = \t g Q : J/*(<) < y(0f ' s °f the first category (see above). Since J is countable, Q\P is of the first category, therefore the set P(■ E which is defined on each x g A, because by the assumption of ip' the elements y' = ip'(x) g E are pairwise disjunct, and hence there exists supy' j
Proceeding in the same manner as in the proof of the first part of the theorem allows us to conclude that ip g $(A). Clearly, all p' < tp, i.e., the vector lattice $(A) is augmented. R e m a r k 11.14. From the proof of Theorem 11.37 it follows that if tp = supy 1 j
and J is countable, then ip = ls\i-pp\\
(11.23)
A similar formula holds for the infimum. The above definition and theorem have interesting applications some of which are important to us. In E, it is possible to introduce the commutative operation of multiplication, in which the unit of multiplication coincides with the unit of space (see [1], [7]). Indeed, let T = Rn and the function / g S(T) be defined as follows: f{yi,V2,---,y») = Vi ■ Vi ■ ■■■ ■ V*- B y t h e continuity of / tp = {/} g $(A) and the function tp(xx,Xi,... ,xn) has the above properties of multiplication. Then we write tp(xu x2,...,xn) = xl ^6pyrfghfed F ^ a n y ^ x = ( , l f . . . , xn) g B, and for a
280
CHAPTER
11. OPTIMIZATION
PROBLEMS
IN FUNCTION
SPACES
given tp, the set P(tp) consists of those t G Q in which Xi(t),..., xn(t) are finite. So p(p) = # ( x ) is open and everywhere dense. Note that if xi,x2,- ■ ■ , x„ G E 0 (S 0 is the vector lattice of bounded elements), then xi • Xa • . . . • as» € Eo. R e m a r k 11.15 In [1], it is shown that if the multiplication with the above prop erties is defined for ant pair of elements of the vector lattice with unity, then the vector lattice is either augmented or a vector lattice of bounded elements. 2. For each x G 3 there exists a unique element x _ 1 G E such that x ■ x - 1 = ex (ex being a trace of x). Indeed, let n = 1, T = Rl and / G S{T) be defined by the equality f(y) = i if y ^ 0 and /(0) = 0. Then / G C/(T) and the map tp = {/} is such that ip(x) = x'1 for any i £ S . The element x _ 1 is called the inverse of x (see [8]). Note that if for any element of some vector lattice there exists the inverse, then the vector lattice is augmented [8]. So the vector lattice is required to be augmented even for the simplest functions of its elements. We shall now examine the limit properties of functions from $(A). Let x' G H n , x* = (x'j,... ,x\), i = 1,2,..., x = ( x l 5 . . . , xn) G E" We write (o) — limx' = x if for any j = 1,... ,n (o) — limxj- = Xj. L e m m a 11.20 Let x' G E, i = 1, 2,. . . For the existence of (o) — limx' it is necessary and sufficient that for all t G Q, except the first category set, there exists a Unite limit limx*(i) = x(t). Lemma 11.20 is an explicit generalization of the same assertion as was proved for n = 1 in [6]. Definition 11.14 The map
282
CHAPTER 11. OPTIMIZATION
PROBLEMS IN FUNCTION
SPACES
The Feasible and Optimal Solutions of problems In what follows it is assumed that in (11.22)-(11.23) x G A, g' G $(A) or x G Ay', g G $ ( A r ) , i = l , . . . , m (Y is a strip in S). In one case, problem (l)-(2) is denoted by T, and in the other by T(Y), Clearly, the problems F, F(E) coincide. By the feasible solution of problem T (T(Y)) is meant the vector x G A x G A y which satisfies the constraint system of a suitable problem. Let x* be a feasible solution of problem T (T(Y)). If for any feasible solution x of this problem g°{x*) < g°{x), then x* is (as usually) called the optimal solution of problem T (resp. problem T(Y)). Note some properties of the feasible solutions of problem T. Definition 11.16. Let Y\,Yi, •.. ,Yk be the strips of the vector lattice S that are isomorphic to the strips Coo(Ei), Coo(E2), ■ ■ ■, C^E^) of the vector lattice Coo(Q), x
respectively. Then CooiflEA
is a strip in Coo(Q) that is isomorphic to some strip
Y C H. The strip Y is called the union of strips Yi,Y2,..., Yk and is denoted by S(YuY2,...,Yk). Lemma 11.21 Let x1,x2,...,xk be the feasible (optimal) solutions of the re spective problems F(Yj), T(Y2),..., r(YJt) where the strips Y\, Y2,..., Yk are pairwise disjunct, Y = S(Y1 ,Y2,...,Yk). Then x =Si= xl + x2 + ... + xh i
is a feasible (optimal) solution of problem T(Y). Lemma 11.22 Let x be a feasible (optimal) solution of problem T, and let Y be a strip in S. Then [Y}x is a feasible (optimal) solution of problem T(Y). Lemma 11.23 Let x 1 , x2,..., xk be the feasible (optimal) solutions of problem T, and Yj, Y2,..., Yk the strips which form an expansion of the vector lattice E. Tien [Yi]xl + [Y2]x2 + ... + [Yk]xk is a feasible (optimal) solution of problem T. Lemmas 11.21, 11.22, are obvious, and Lemma 11.23 is their consequence. Note that, by Lemma 11.23, the solvable problem T has generally either a unique optimal solution or infinitely many such soluitons (because the number of all possible strips, which form an expansion of the vector lattice, is finite only if the space of JE is finitedimensional). Since in (11.22)—(11.23) g' = {g'T}, i = 0 , 1 , . . ., m each point T £ Q can be related to a numerical problem F(T) of the form 9° -►inf g'T(y)<0,
i=
l,2,...,m,
where y G T. Between the feasible solutions of the problems T and T(T) there exists a relationship which is expressed in the assertions below. Lemma 11.24. The vector x G A is a feasible solution of problem F if and only if at each point r G Q, except the set of first category, x(r) is a feasible solution of the respective problem T(T).
11.7.
OPTIMIZATION
PROBLEMS
ON VECTOR
LATTICES
This l e m m a is trivial: its assertions hold for r g f] P(g'). o L e m m a 1 1 . 2 5 . Let x be a feasible solution of problem V, and suppose m
F C f l P(g')
C H(x)
is such that Q\P
283
the set
is of the first category and for any r g P X(T)
is an optimal solution O I T ( T ) . Then x is an optimal solution of problem F. Proof Let g°{x) = z and, on some feasible solution x* of problem T, g°(x") = 2* < z. T h e n at t h e points t of the nonempty open set V C Q z' < z(t). But at the points of the everywhere dense set (which is said to coincide with P) x*(r) is a feasible solution of problem T ( T ) , and since g° = {g°}, we have <7°(X*(T)) = Z*(T) < z ( r ) = # ° ( Z ( T ) ) for r € P (1 V But the set P n V is of the second category, which is inconsistent with the assumption of the lemma. Classification of Optimization Problems T h e functions from $ ( A ) admit classification according to t h e properties of con vexity which allows us to properly classify optimization problems on vector lattices. We shall consider these properties. Suppose t h e set T is convex. We establish a relation between the properties of the sets T C R' and A C E". L e m m a 1 1 . 2 6 A is convex only if T is convex. Additionally, A is convex over E, i.e., for any x,y g A, a , 0 g E, ce,0 > 0, a + 0 = 1, ax + 0y g A. 6 T h e l e m m a is obvious if it is considered in Coo(Q). T h e sets T, A are assumed to be convex t h r o u g h o u t this subsection. D e f i n i t i o n 1 1 . 1 7 . T h e m a p0, a + 0 — 1,
0, X + /J. = 1,
fi{bl) and it remains to take into account Remarks 1,2 to Definition 11.13. T h e sufficiency of the assertion is obvious. T h e o r e m 11.43 admits a similar formulation for concave and linear functions. Then t h e following theorem is true. T h e o r e m 1 1 . 4 4 Let A = En and suppose
ax = ( M I , . . . , ax„) = a(x\,..., tion 3 of this Section.
xn). The product of elements from E is determined in Subsec
284
CHAPTER 11. OPTIMIZATION
PROBLEMS IN FUNCTION
SPACES
Note that ip = {v T }, where
11.8
Generalized Parametric Programming Problem
This section focuses on the relationship between a generalized (in a sense) paramet ric programming problem and the numerical optimization problems obtained from the parametric problem by fixing some values of parameters. The form "paramet ric problem" is interpreted here more generally than it is usually understood: the dependence on parameter can be rather arbitrary, and the parameter can be taken to be the points of a fairly abstract set, e.g., the points of some compact set. In this case, the functions of parameters involved in formulation of the problem can be measurable functions, functions of a constrained variation, arbitrary real functions given on a vector set, etc., i. e., the elements of some linear space. In such a space, it would be natural (because the optimization problem is considered) to contemplate a partial ordering of elements which agrees with linear operations. We are thus brought to the idea of utilizing Riesz spaces [1] and formulating the problem of mathematical programming in Riesz space. A special case of this problem, namely the linear pro gramming problem in Riesz space, is discussed in [1]. Note that no special constraint is imposed on the functions forming the problem, e. g., their linearity or convexity (or even continuity) is not contemplated here. This is due to the fact that the absence of constraints imposed on such functions is motivated by the main purpose of this section, namely: to distinguish a wide class of parametric problems, where there is a relationship between the parametric LP problem and the numerical LP problems obtained by fixing the values of parameters. At the same time certain requirements (e.g., the existence of the iteration algorithms for solving these problems) can be imposed on the numerical problems thus obtained. Although such a requirement re stricts the class of problems, it is justified by the idea that many specific problems, which, e.g., are not convex and do not belong to any class, admit construction of an algorithm for their solution in the general form. Such problems, however, can be solved by some specially constructed algorithms. It turns out that the availability of these algorithms for solving numerical problems involves the existence of the solution algorithm for the generalized parametric problem. And again we must remember that here the term "parametric problem" is used conventionally. For example, when the functions of parameters are measurable, one may consider the stochastic programming problem with the implication that these functions are some random variables. To read this section, one has to know only the basic facts from the Riesz space
11.8. GENERALIZED
PARAMETRIC
PROGRAMMING
PROBLEM
285
theory as presented in the first five chapters of monograph [1]. Let E be a Riesz space (a universally complete vector lattice) with a fixed unit e, h : S -» Coa{Q) being the isomorphic mapping of E onto the Riesz space Coo(<2) of all real functions that are continuous on the extremely disconnected compactum Q and allow for infinite values on nondense sets. Here he 6 CX{Q) is the function that is identically equal to unity. If x g E, then hx = x(t), t 6 Q and conversely, if x(t) g CaalQ), then h~1x(t) = x g E. For the purposes of further discussion, we need to distinguish the class of functions of elements of the Riesz space E with values in S. We shall use the findings in G. J. Lozanovsky [5]. For some integer n > 0 in Rn, we fix the set T and elements x^,. . .,x„ 6 E. Set x = (xuxn), x(t) = ( i i ( i ) , . . ., xn(t)) and H = {t g Q : x(t) g T). Let S(T) be the algebra of all real finite functions on T, f g S(T). Further, suppose that A) z g E and B) there exists a set P C H such that Q\P is of the first category in Q and for any t g P f(x(t)) = z(t). Then, according to [3], and by the definition, we set T{x) = z. Denote by U„{T) the set of all / € S{T) for which there exists JF(x) g E (for any x satisfying conditions B). In [3], it is demonstrated that Un(T) 7^ 0. Additionally, Un(T) contains all Baire's functions on T. Let / g S(T), Cj,. . . ,Ck be some constants involved in generation of the function / , and ck = ( c i , . . . ,Cjt), y g T, i.e., / is of the form f(ck;y). Parametrization of the function / involves replacement of vectors c* and y with some vector functions of parameter t, which leads to some function of the form f(ck(t);y(t)j. This procedure is sufficiently general, because the set T C Rn, constants c i , . . .,<% and their number are selected in an arbitrary way. If t is selected to be the points of the compact set Q, this naturally gives rise to the following definition of the function of "parameter" t. Let ci(i),. . ., Cjt(i) g Coo(Q), ck(t) = lci(t),... ,Ck(t)j and suppose the set H is determined for a given x(t). The function
286
CHAPTER 11. OPTIMIZATION
PROBLEMS IN FUNCTION
SPACES
function g$(ak(t);x(i))^M
(11.25)
with the constraints on x(t) of the form: g?((3<(t);x(t))<0, (H-26) g%(Y(ty,x(t))<0, where gf G U$(T), i = 0 , 1 , . . . , m. Denote problem (11.25)-(11.26) by TQ. Passing (by means of isomorphism h) to the space H yields the problem denoted as FH- Fixing the finite values ah(t), /?''(*),... , 7 P (t) at the points T of an open everywhere dense set, we obtain the numerical problems g0{inf
(11.27)
5m(7 (^);y)can be viewed as an independent Riesz space COQ(E), where E C Q is an open-closed set. Denote by YE(W = Coo(-S)) the problem [w}g$(ak(ty,x(t))^mi
Hff?(^(t);<*))*(*); *(*))<«, where [iu] is the operator of projection onto the component io, while the auxiliary constraint x(t) = (zi(i),. . •, xm(t)j, Xj(<) 6 w, j = 1 , . . . , n is imposed on TE(W = Coo{E)). If z is a component of 2, then projecting T= onto z yields, in a similar manner, problem Tz. The feasible solution of problem VQ is interpreted to mean the vector function x(t) (with components from Ctx,(Q))) satisfying the constraint system of the problem. The feasible solution i*(i) of problem TQ is optimal if for any feasible solution x(t) go(x(t)) > g^ (x*(t)). In a similar manner, define the concepts of the feasible and the optimal solution for problem r=, T ^ T E . We shall now formulate some statements about the set of feasible solutions of problem FQ (where it is possible, formulations are given in terms of the space 2). (1) If x(t) is a feasible solution of problem TQ, then at the points r of an open, everywhere dense set x(r) is the feasible solution of a suitable problem F T .
11.8. GENERALIZED
PARAMETRIC
PROGRAMMING
287
PROBLEM
(2) If x(t) = Xi(t),. . .,xn{t)), Xj(t) £ C 0o (Q), j = l , . . . , n and at the points T of an everywhere dense set X(T) is the feasible solution of a suitable problem TT, then x(t) is a feasible solution of problem TQ. (3) If x = (Xi,. . ., xn) is a feasible solution of problem F= u> is a component in H, then [w]x = ^[tu]xi,..., [u>]xnJ (this notation will be preserved in what follows) is a feasible solution of problem Tw. (4) Let x' = (x\,. .. ,x'n)t i = 1 , . . . , r be optimal solutions of problem IV, and T
suppose the components u»i,..., wr form an expansion of space !E. Then J2 [wi]x' = 1=1
J2 v\[wi]x\,...,
v=i
[to.'lxJJ is an optimal solution of problem I V '
(5) If x' is an optimal solution of problem Tw, i = 1 , . . . , T , and components Wi,... ,wT form an expansion of E, then x1 -f . . . + xT = x is an optimal solution of problem I V The last two assertions hold for the feasible solutions that are not necessarily optimal. (6) If x*(t) is a feasible solution of TQ such that X*(T) is an optimal solution of a suitable problem TT for any r from an everywhere dense set, then x*(t) is an optimal solution of TQ. Statements (11.25), (11.26) follow from the definition of functions gf, i = 0,1,... ,m. The proofs of statements (11.27)—(11.28) are based on the obvious formula
fix) = r (Hx + [z)x) = {w\r(x) + [z]f(x), where components w, z form an expansion of H. Let us prove statement (11.30). If x{t) is an optimal solution of TQ and g$ (x(t) < goix'it)), then at the points r of an everywhere dense set X(T) is a feasible solution of TT (statement (11.25)), therefore go(ak(T); X(T)) < g0{ak{r); X'(T)J at the points of a nonempty open set, which yields a contradiction. Let x'°)(£) be a feasible solution of problem FQ. Then, according to statement (1), Z'°)(T) is a feasible solution of TT, where r runs over an open, everywhere dense set. If z'°'(£), i ' ° ' ( r ) are the initial approximations of optimal solutions of suitable problems, we assume that for each of the problems F T , where r belongs to an everywhere dense set 5, there is an iteration process described by the formulas X^+1\T)
= u\ ( z w { r ) ) ,
i = 0,l,...
(11.29)
Here it is assumed that for any i = 0,1,... the values i ( , + 1 ' ( r ) at the points T £ S coincide with those at these points of some vector function x' , + 1 '(t) with components from Coo(Q). Thus, the values I < ' + 1 ' ( T ) T £ S uniquely determine z<'+1>(tj that is a feasible solution of problem TQ (statement (11.26)). The above assumption leads to the existense of the maps Ui acting from C^Q) x Coo(Q) x . . . x CoofQ) into themselves (at t — r from an everywhere dense set the values U{(x(t)) coincide with UJ(X(T)).) Then we have
^"teB^ 0 - 1 '-
(1L30)
288
CHAPTER 11. OPTIMIZATION
PROBLEMS IN FUNCTION
SPACES
The relations (11.30) determine the iterative solution process of problem TQ, and if the iterative sequences X'(T) for each r G S converge to X(T), then the sequence i'*'(t) converges in a componentwise (O)-manner to some optimal solution x(t) of problem TQ (convergence of this sequence follows from statement 2.33 in [4]). Here X(T) = xT at the points of an everywhere dense set. From this we may deduce the inverse statement to (11.20): if x(t) is an optimal solution of problem TQ, then at the points of an everywhere dense set x(r) is the optimal solution of a suitable problem r T . So the assumption that for problem F, r 6 S there exist convergent iterative processes (11.29) is true. Theorem 11.45. Problem TQ is solvable if and only if to the points of some everywhere dense set there correspond solvable problems TT. To prove this theorem, it sussifices to construct convergent iterative sequences for the problems TQ TT by the formulas (11.29) and (11.30). Passing to the space TT and making suitable replacements in (11.30), we obtain iterative process formulas for problem T=. In closing this section, it may be said that the results presented here are applicable both to unconstrained minimization problems (where there are no constraints (11.26)) and inequality systems (the objective function is an identical constant).
Conclusion In the final analysis , human activities manifest themselves in two trends: creation of material and spriritual values and their distribution. Natural sciences have a great impact on development of the former trend, whereas humanitarian sciences are responsible for distribution of the values within a society. The standard of living largely depends on how the values, goods and services are distributed in the society. The end of this century is characterized by penetration of natural sciences into the sphere that is traditionally related to humanitarian investigations. In this respect, the efforts made to extend the techniques developed by natural sciences seem to be valid for solving problems in allocation of profits and resources. There are many problems that can be adequately described in terms of system optimization. The entire classical mathematics, even in its abstract sections, describes the category of motion. Characteristic of the society are goal-oriented processes, and hence conflicts of interests. Then the problem arises as to how the best (in a sense) or the optimal behavior, decision or control can be selected from the set of feasible ones. This monograph provides an outline of some trends in the systems optimization methodology. Naturally, the monograph is far from being universal, but it provides orientation in the problems that are not very simple.
BIBLIOGRAPHY for chapters 1 and 2
[1] Ben-Israel, A., On some Problems of Mathematical Programming, Science. Evanston 1961, Northwestern University.
Ph.D. Thesis in
Engeneering
[2] Charnes, A., Cooper, W. W., and Symonds, G. N., 'Cost Horizons and Certainty Equivalents: an Approach to Stochastic Programming of Heating Oil', Management Sci. 4 (1958), 235-263. [3] Charnes, A., Cooper, W. W., and Symonds, G. N., 'Cost Horizons and Certainty Equivalents: Heating Oil', Management Sci. 4, No. 12 (1958). [4] Charnes, A. and Cooper, W. W., 'Chance-Constrained Programming'. Management (1959), 73-79.
Sci.
6
[5] Charnes, A. and Cooper, W. W., 'Chance-Constraints and Normal Derivatives', ,/. Am. Assoc. 9 (1962), 33-49.
Sratist.
[6] Charnes, A. and Cooper, W. W., 'Deterministic Equivalents for Optimizing and Satisficing under Chance-Constraints', Operation Res. 1 1 (1963), 18-39. [7] Dinkelbach, W., C. On Non-Linear Fractional Programming , Management 492-497.
Sci.
13 (1976),
[8] Dragomirescu, M. and Lidaroin, C , 'Modele stochastice peutru determinaer lastructuri optime a productili agricole', Revista de Statistica, No. 6 (1968). [9] Friedland, A. J., 'Methods of Optimal Annual Planning for Coal Output at a Group of Mines: Probabilistic Approach', Ph.D., Moscow 1971. [10] Holstein, E. G. and Judin, D. B., New Approaches Soviet Radio.
of Linear Programming,
Moscow 1966,
[11] Joffe, A. and Judin, D. B., 'On some Stochastic Non-linear Programming Problems', J. Com puting Mathematic. Phys. 10, No. 1 (1970). [12] Judin, D. B., 'Stochastic Quadratic Programming', Rep. [13] Judin, D . B . , Mathematical Radio.
Methods of Management
U.S.S.R. Acad. Sci., No. 3 (1969).
under Uncertainty,
Moscow 1974, Soviet
[14] Judin, A. D., 'Stochastic Variant Model for Working out of the Plan of Functioning and Production O u t p u t for the Firm', Econ. Mathematic. Methods X , No. 6 (1974). [15] Kataoka, S., 'On Stochastic Programming. I: Stochastic Programming and its Application to Production Horizon Problem', Hitotsubashi Journal Arts Science. No. 2 (1962), 23-35. [16] Kataoka, S., 'On Stochastic Programming. II: A Preliminary Study of a Stochastic Program ming Model', Hitotsubashi Journal Arts Science, No.2 (1962), 36-44. [17] Kataoka, S., 'On Stochastic Programming. Ill: A Stochastic Programming Model', HUotsubaxhi Journal Arts Science, No. 2 (1962), 44-55. [18] Kataoka, S., 'A Stochastic Programming Model', Econometnca
3 1 (1963), 181-196.
[19] Kolbin, V. V., 'Stochastic Programming', Theory of Probability and Mathematical Theoretical Cybernetics. Results of the Science, VINITI, Moscow, 1968, pp. 5-68. [20] Kolbin V. V. 'Stochastic Programming' D.Reidel Publ. Company, 1976, 196 p.
289
Statistics.
290
BIBLIOGRAPHY
[21] Kolbin, V. V., 'Investigation of Absolute Stability of Planning for the Solution of Stochastic Programming Problems', Proceeding of the 2-nd All-Umon Symposium Problems of Systemotechmque, Vol.2 (1972), pp.96-102. [22] Kolbin, V. V., 'On the Possibility of the Optimal Decision Application to the Chance-Constrai ned Mathematical Models', Systems of Management Economic and Organizational Problems, Leningrad Engeneering and Economic Institute, Vol.107 (1975). [23] Kolmogorov, A.N., 'Basic Notions of the Theory of Probabilities', M.-L., ONTI, 1936. [24] Lavrinenko, E. P., 'Stochastic Programming Application to the Air Trafic Control Problem', Techonog. Cybernetics 16 (1970), 50-56. [25] Madansky, A., 'Methods of Solution of Linear programs under Uncertainty', Operation 10 (1962), 463-471.
Res.
[26] Miller, B. L. and Wagner, H. M., 'Chance-Constrained Programming with Joint Probability Constraints', Operation Res. 13 (1965), 930-945. [27] Polyak, B. T., 'Non-Smoothing F u n c t i o n a l Minimization', J. Comput. Phys. 9 (1969), 509-521.
Math.
Mathematic.
[28] Prymak, M. E., 'Numerical Methods for the Solution of Some Problems of the Theory of Approximations', Ph.D. Voronezh University, 1966. [29] Sengupta, J. K., 'Distribution Problems in Stochastic nd Chance-constrained Programming', in Economic Models, Estimation, and Risk Progtammmg, New York 1969, Springer-Verlag, pp. 391-424. [30] Sengupta, J. K., 'Safety-First Rules under Chance-Constrained Linear Programming', Opera tion Res. 17 (1969), 112-132. [31] Shchepakin, M. B., 'Multi-Stage Stochastic Problem on Mixture', Mathematical Systems' Investigations and Modeling, No. 4 (1970), pp. 3-15.
Metods of
[32] Soldatov, V. E., 'A problem of Chance-Constrained Linear Programming', Siberian Math. 6 (1965), 705-710.
J.
[33] Soldatov, V. E., 'On the Chznce-Constrained Linear Programming Problem', Math. Opt. Plann. Models, pp. 54-64, Novosibirsk 1966.
Models
[34] Soldatov, V. E., 'On Some Stochastic Programming Problem', Siberian Section of the Academy of Sciences Inxtitute of Mathematics pp. 78-89, 1966.
U.S.S.R.
[35] Symonds C , 'Deterministic Solutions for a Class of Chance-Constrained programmong Prob lems', Operation Res. 15 (1967), 495-512. [36] Wessels, J., 'Stochastic Programming', Statistica
Neerlandica
2 1 (1967), 39-53.
[37] Zhulenev, S. V., 'Chance-Constrained Problem and Normal Model', Econ. Mathematic. ods 6 (1970), 757-764.
Meth
[38] Zhulenev, S. V., 'Normal Model for the Chance-Constrained Problem Normal Distribution', Economic Process Modeling, pp. 403-414, Moscow 1971, Moscow university Press. [39] Zellmer, C , 'Uber Problemstellungen in der stochastischen linearen Optimierung', tische Wirtschaft6 (1968).
Mathema-
[40] Zuhovitsky, S. I., Polyak, R. A., and Prymak, M . E . , 'Numerical Methods for the Solution of Convex Programming', Rep. U.S.S.R. Acad. Sci. 1 6 3 , No. 2 (1965).
BIBLIOGRAPHY
291 BIBLIOGRAPHY for chapter 3
[1] Arbouzova, N. I., Vereskov, A. I., and Nikolayeva, N. D., 'Some Stochastic Programming Problems (Review)', Econ. Math. Methods 5 (1969), 412-430. [2] Beale, E. M. L., 'On Minimizing a Convex Function Subject to Linear Inequalinies', J. Roy. Statist. Soc. Series B 1 7 (1955), 173-184. [3] Berkovitch, E. M., 'On t h e Two-stage Problem of Stochastic Optimal Management', University, Series: Math. Mech. 4 (1970), 9-17. [4] Berkovitch, E. M., 'On the Approximation of Two-Stage Stochastic Problems', J. Math. Mathematic. Phys. 1 1 (1971), 1150-1165.
J.Moscow
Computing
[5] Berkovitch, E. M., 'On the Existence of an Optimal Solution for a Class of Two-Stage Stochas tic Extremum Problems', Approzivate Methods of Solution for the Problems of Optimal Man agement and for Some Incorrect Inverse Problems. Translations of the Computing Center of Moscow University, Moscow 1972, Moscow University Press. [6] Berkovitch, E. M., 'On the Nimerical Solution of a Class of Stochastic Optimal Management Problems', Approximate Methods of Solution for the Problems of Optimal Management and for Some Incorrect Inverse Problems, Transcastions of the Computing Center of Moscow Uni versity, Moscow 1972, Moscow University Press. [7] Bereanu, B., 'Stochastic Transportation Problem: I, II. Random Costs', Comunicavile R.P.R. 1 3 (1963), 325-337.
Acad.
[8] Charnes, A., Cooper, W. W., and Thompson, G. L., 'Constrained Generalized Medians and Hypermedians as Deterministic Equivalents for Two-Stage Linear Programs under Uncertqinty', Management Sci. 2 1 (1965), 83-112. [9] Charnes, A., Kirby, M., and Raik, W., 'Chance-Constrained Generalized Networks', Res. 14 (1966), 1113-1120.
Operation
[10] Charnes, A., Kirby, M., and Raik, W., 'Solution Theorems in Probabilistic Programming: a Linear Programming Approach', J. Math, Appl. 20 (1967), 365-382. [11] Charnes, A., Littlechiled, Kirby, M., and Raik, W., 'Chance-Constrained Model on Transport Prices and Scheduling under Comprtition', Transportation Science No. 1. 1968. [12] Courtillot, M., 'Contribution a la theorie de la programmation lineaire et de la programmation stochastique', Theses Doct. Sci. math., Fac. Sci. Umv. Paris, p. 96, Paris 1963. [13] Dantzig, G. B., 'Linear Programming under Uncertainty", Management
Set. 1 (1955), 197-206.
[14] Dantzig, G. B., 'Recent Advances in Linear Programming', Management 144.
Sci. 2 (1956), 131—
[15] Dantzig, G. B. and Madansky A., 'On the Solution of Two-Stage Linear Programs under Uncertainty', proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probavility, Vol. 1, pp. 165-176. Berkeley 1961, University of California Press. [16] Dempster, M. A. H., 'On Stochastic Programming. I, Static Linear Programming under Risk', Math Anal. Appl 2 1 (1968), 304-343. [17] Dvoretsky, A., 'On Stochastic Approximation', Proceedings of the 3-rd Berkeley Symposiom Mathematical Statistics and
of
292
BIBLIOGRAPHY
[18] Efimov, V. M., 'Optimal Estimations under Uncertainty', Econ. (1970).
Math.
Methods 6, No. 3
[19] El-Agizy, M., 'Two-Stage Programming under uncertainty with discrete Distribution Function', Operation Res. 15 (1967), 55-70. [20] Elmaghraby, S. E., 'An Approach to Linear Programming under Uncertainty', Operation 7 (1959), 208-216.
Res.
[21] Elmaghraby, S. E., 'Allocation under Uncertainty when the Demand has Continuous Distribution Function', Management Sci. 6, No. 3 (1960). [22] Ermolyev, J. M., 'An Approach to the Solution of Stochastic Programming Problems in Mixed Strategies', Cybernetics, No. 6 (1969). [23] Ermolyev, J. M., 'On of Stochastic Programming General Problems', Cybernetics, (1971), 47-50.
No.
3
[24] Ermolyev, J. M., 'On the Method of Generalixed Stochastic Gradients and Stochastic QuasiFeier Sequentialities', Cybernetics, No. 2 (1969), 73-83. [25] Ermolyev, J. M., 'Stochastic Quasi-Gradient Methods and Applications', PH.D. in the Institute of Cybernetics Ukrainian S.S.R. Academy of Sciences, Kiev 1970. [26] Ermolyev, J. M., 'On Some Stochastic Programming Problems', Cybernetics, 1-5.
No. 1 (1971),
[27] Ermolyev, J. M. and Nekrylova, Z. V., 'Method of Stichastic Gradients and Application', Theor Opt. Decis. Institute of Cybernetics, Kiev 1967. [28] Ermolyev, J. M. and Shor, N. Z., 'Random Search Method for Stochastic Programming Twostage Problems and Generalization', Cybernetics, No. 1 (1968), 90-92. [29] Hadley, G., Non-Linear 484 pp.
and Dynamic
Programming
[30] Holstein, E. G. and Judin, D. B., New Approaches Soviet Radio.
Reading,
Mass, 1964, Addison-Wesley,
of Linear Programming,
Moscow 1966 ,
[31] JofFe, A. and Judin, D. B., 'On Some Stochastic Non-Linear programming Problems', / . Computing Matherat. Mathematic. Phys. 10, No. 1 (1970). [32] Judin, D. B. and Zoy, E. V., 'On Weak Solution of Stochastic Programming Two-Stage Problems', Economics and Mathematical Models, Moscow 1973, Moscow University Press. [33] Judin, D. B., Mathematical Radio.
Methods of Management
under Uncertainty,
Moscow, 1974, Soviet
[34] Kail, P., 'Quahnanive Aussagen zu eingen Problemen der stochastischen Programmierung', Z. Wahrschemlichkeitsheorie und verw. Geciete. 6 (1966), 246-272. [35] Kail, P., 'Das zweistufige Problem der stochastishen linearen Programmierung', Z. lichkeitstheorit und verw. Gebiete. 8 (1967), 101-112. [36] Keri, G., 'On the Two-Stage Programming under Uncertainty', Studia Set. No. 5 (1970), 37-40.
Math.
[37] Kolbin, V. V., 'Stochastic Programming', Theory of Probability and Mathematical Theoretical Cybernetics. Results of the Science, VINITI, Moscow, 1968, pp. 5-68. [38] Kolbin V. V. 'Stochastic Programming' D. Reidel Publ Company, 1976, 196 p.
WahrscheinHunger., Statistics.
293
BIBLIOGRAPHY
[39] Kolbin, V. V., and Tanskaya, V. N., 'Some Stochastic Linear Programming Problems and Algorithm for the Solution', Economic Processes Models, Moscow University Press. [40] Kolbin, V. V., 'Investigation of Absolute Stability of Planning for the Solution of Stochastic Programming Problems', Proceeding of the 2-nd All-Union Symposium Problems of Systemotechnique, Vol.2 (1972), pp. 96-102. [41] Kolbin, V. V., 'On the Possibility of the Optimal Decision Application to the Chance-Const rained Mathematical Models', Systems of Managment Economic and Organizational Problems, Leningrad Engeneering and Economic Institute, Vol. 107 (1975). [42] Kolbin, V. V., 'On a Stochastic Programming Problem', Management vostok, U.S.S.R. Academy of Sciences, Vol. 16 (1975).
an Information,
Vladi
[43] Kuhn, H. W. and Tucker, A. W., 'Non-Linear Programming', Proc. on Math., Stat. and Prob., Univ. California Press, 1951
Second Berkeley
Symp.
[44] Lieu, B. T., 'On a Problem of convexity and its Application to Nonlinear Stochastic Program ming', / . Math. Anal. Applic. 8 (1964), 177-187. [45] Madansky, A., 'Inequalities for Stochastic Linear Programming Problems', Management 6 (1960), 197-204.
Set.
[46] Madansky, A., 'Methods of Solution of Linear Programs under Uncertainty', Operation 10 (1962), 463-471.
Res.
[47] Madansky, A., 'Dual Variables in Two-stage Linear Programming under Uncertainty', J. Math. Anal. Applic 6 (1963), 98-108. [48] Madansky, A., 'Linear Programming under Uncertainty', Recent Advances in Math. ming, R. L. Graves and P.Wolfe (eds), McGraw-Hill, New-York 1963, pp. 103-110.
Program
[49] Mangasarian, O. L., 'Nonlinear Programming Problems with Stochastic Objective Function', Management Set. 10 (1964), 353-359. [50] Mangasarian, O. L. and Rosen, I. B., 'Inequalities for Stochastic Ninlinear Programming Prob lems', Operation Res. 12 (1964), 143-154. [51] Mihoc, G. and Nadejde, Ileana, ' P r o g r a m a r e a m a t h e m a t i c a . Programareastochastica', Editura stintifica, Bucuresti 1967, 407 pp. [52] Milder, W. and Wollnar, 'Stochastic Programming Models for Scheduling Airlift Operation', Nav, Res. Log. Quart. 16, No. 3 (1969). [53] Murty K a t t a , G., 'Linear Programming under Uncertainty: a Basic Property of the Optimal Solution', Z. Wahrschemlichkeitstheorie verw. Geb. 10 (1968), 284-288. [54] Sengupta, J. K., 'Economics of Decomposition and Divisionalization under Transfer Pricing', Z. fur die gesamte Staatswissenschaft 127 (1971), 50-71. [55] Sengupta, J. K., Stochastic Programming: Holland Publishing Company.
Methods and Applications,
Amsterdam 1972, North-
[56] Shahidi, A. A., 'Models and Solutions of Some Stocastic Programming Problems', Rep. Sci. Tadjic S.S.R 1 1 , No. 3 (1968). [57] Szwarc, W., 'The Transportation Problem with Stochastic Demand' Management (1964), 33-50.
Set.
Acad. 11.
[58] Tinn, K. A. and Tnugu, E. H., 'On Non-Linear Programming with Chance-Constraints', Cy bernetics, No. i, (1968), hi-f&opyrighted Material
294
BIBLIOGRAPHY
[59] Vajsa, S., 'Inequalities for stochastic Linear Programming', Bull. 36 (1958), 357-363.
J. Internat.
Statist.
Inst.
[60] Van Slyke, R. and Wets, R. J. B., 'Programming under Uncertainty and stochastic Optimal Control', J SIAM Control. 4 (1966), 179-193. [61] Walkup, D. W. and Wets, R. I. B., 'Stochastic Programs with Recourse', SIAM J. Appl. 15 (1967), 1299-1314. [62] Wessels J., 'Stochastic Programming', Statistica
Neerlandica
Math.
2 1 (1967), 39-53.
[63] Wets, R. I. B., 'Programming under Uncertainty: The Complete Problem', Zeitschrtft scheinhchkeitstheorie verw. Gebiete 4 (1966). 316-339.
Wahr-
[64] Wets, R. I. B., 'Programming under Uncertainty: The Equivalent Convex Program', J. Appl. Math. 14 (1966), 1143-1151.
SLAM
[65] Wets , R. I. B., 'Programming under Undertainty:' The Solution Set , SIAM 14 (1966), 1143-1151.
Math.
J. Appl.
[66] Wets, R. I. B., 'Stochastic Programs with Recourse; a Basic Theorem for Multistage Problems', Zeutschrift Wahrscheinlichkeitstheorie verw. Gebiete 2 1 (1972), 201-206. [67] Williams, A. C , 'A Stochastic Transportation Problem', Operation Res. 1 1 (1963), 759-770. [68] Williams, A. C , 'Approximation Formulas for Stochastic Linear Programming', SIAM J, Appl. Math. 14 (1966), 668-677. [69] Zellmer, G., 'Zu Sensibilitatsuntersuchungen in der stochastischen linearen Optimierung', Wiss Zeitschrift der Hochschule fiir Okonomte, Berlin, No. 4 (1968), pp. 495-502.
295
BIBLIOGRAPHY BIBLIOGRAPHY for chapter 4
[1] Bellman, R., 'Decision Making in Face of Uncertainty, II,' Naval Res. Logist. Quart. 1 (1954), 327-332. [2] Charnes, A., Cooper, W. W., and Symonds, G. N., 'Cost Horizons and Certainty Equivalents: Heating Oil', Managment Set. 4, No. 12 (1958). [3] Charnes, A. and Kirby, M., 'Optimal Decision Rules for the B-Model of Chance-Constrained Programming', Cah. Centre Etud. Recherch Operation. 8 (1966). 5-44. [4] Charnes, A. and Kirby, M., 'Some Special P-Models in Chance-Constrained Programming', Management Set. 14 (1967), 183-195. [5] Charnes, A. and Kirby, M., 'Optimal Decision Rules for the Triangular £-Model of ChanceConstrained Programming', Cah. Centre Etud. Rech. Operation. 11 (1969), 215-243. [6] Efimov, V. M., 'Perspective Planning: Stochastic Model', Theory of Optimal Solutions, Kiev 1969. [7] Efimov, V. M., 'Stochastic Economical Models', Stochastic Economical Models, Moscow 1971, Moscow University Press. [8] Eisner, M. J., Kaplan, R. S., and Soden, J. V., 'Admissible for the S-Model of ChanceConstrained Programming', Management Set., No. 17 (1971). [9] Ermolyev, J. M., 'Stochastic Quasi-Gradient Methods and Applications', PH.D. in the Institute of Cybernetics Ukrainian S.S.R. Academy of Sciences, Kiev 1970. [10] Ermolyev, J. M. and Shor, N. %., 'Random Search Method for Stochastic Programming Twostage Problems and Generalization', Cybernetics, No. 1 (1968), 90-92. [11] Fan Tsi, 'On Systems of Linear Inequalities', Linear Inegualities and Related Problems, Moscow 1959, pp. 214-262. [12] Joffe, A. and Judin, D. B., 'On Some Non-Linear Stochastic Programming Problems', Rep. U.S.S.R. Acad. Sci. 186 (1969), 16-18. [13] Judin, D. B., 'New Approaches of Stochastic Programming', Econ. Mathemat. Methods 4, No. 6 (1969). [14] Judin, D. B., 'Duality in Stochastic Programming', Econ. Mathemat. 280-284.
Methods 5 (1969),
[15] Judin, D. B., 'Stochastic Quadratic Programming', Rep. U.S.S.R. Acad. Sci., No. 3 (1969). [16] Judin, D. B., 'Multi-Stage Planning Problems undert Risk and Uncertainty', Tech. Cyberne tics, No. 6 (1972). [17] Judin, D. B., 'Dynamic Problems of Stichastic Programming', Rep. U.S.S.R. Acad. Set. 210, No. 3 (1973). [18] Judin, D. B., 'Methods for the Construction of Decision Rules for Stochastic Programming Dynamic Problems', Rep. U.S.S.R. Acad. Sci. 210, No. 4 (1973). [19] Judin, D. B. and Zoy, E. V., 'Stochastic Programming Problems under Preiliminary', Econ. Mathematic. Methods 9. No. 5 (1973). [20] Judin, D. B., Mathematical Methods of Management under Uncertaity, Moscow 1974, Soviet Radio.
296 [21] Judin, D. B., Mathematical Radio.
BIBLIOGRAPHY Methods of Management
Moscow 1974, Soviet
under Uncertainty,
[22] Judin, A. D., 'Duality in Milti-Stage Stochastic Programming Problems', Tech. No. 6 (1973).
Cybernetics,
[23] Judin, A. D., 'Descent by Coordinates in the Problems of Infinite-Dimensionality Programming and i-th Applications', Tech. Cybernetics, No. 1 (1974). [24] Kaplan, R. S. and Soden, I.V., 'On the Objective function for the Sequential P-Model of Chance-Constrained Programming', Operation Research 19 (1971), 105-114. [25] Kolbin, V. V., 'Stochastic Programming', D. Reidel Publ. Company, 1976, 196 p. [26] Kolbin, V. V., 'Stochastic Programming', Theory of Probability and Mathematical Theoretical Cybernetics. Results of the Science, VINITI, Moscow, 1968, pp. 5-68.
Statistics.
[27] Kolbin, V. V., 'Investigation of Absolute Stability of Planning for the Solution of Stochastic Programming Problems', Proceeding of the 2-nd All-Union Symposium Problems of Systemotechnigue, Vol.2 (1972), pp. 96-102. [28] Kolbin, V. V., 'Investigation of the Conditions of the Optimal Decisions Existence in the Stochastic Non-Linear Programming Problem', Problem of Economics, Organization, Infor mation and Mathematics in the System of Management, Leningrad Engineering and Economic Institute, Vol. 105 (1975). [29] Kolbin, V. V., 'On a Stochastic Programming Problem', Management vostok, U.S.S.R. Academy of Sciences, Vol. 16 (1975).
an Information,
Vladi
[30] Lemaire, J. M., 'Prevision et decision en programmation lineairestochastique', These, Grenoble 1967. [31] Liapounov, A. A., 'On Completely Additive Vector Functions', Rep. Series Mathematics 4 (1940), 465-478.
U.S.S.R.
[32] Manne, A. S., 'Linear Programming and Sequential Decisions', Management 259-267.
Set.
Acad,
Sci.,
6 (1960),
[33] Mihoc, G. and Nadejde, Ileana, ' P r o g r a m a r e a m a t h e m a t i c a . Programareastochastica', Editura stintifica, Bucuresti 1967 , 407 pp. [34] Midle, P., 'Zu einigen Problemstellungen der stochastischen linearen Optimierung', Wiss. Z. Hochsch. Arch, und Bauw, Weimar 18 (1971), 73-75. [35] Rogovsky, E. A., 'Dynamic Multi-Product Stochastic Perspective Planning Model', Investi gations in Mathematical Economics and Akin Problems, pp. 99-119, Moscow 1971, Moscow University Press. [36] Shchepakin, M. B., 'Multi-Stage Stochastic Problem on Mixture', Mathematical Systems' Investigations and Modeling, No. 4 (1970), pp. 3-15.
Metods of
[37] Sion, M., 'On General Minimax Theorems', Pacific Math. 8, N 1 (1958). [38] Wolfe, P. and Danrzig, G. B., 'Linear Programming in Markov Chains', Operation Res. 10, No. 5. (1962).
297
BIBLIOGRAPHY BIBLIOGRAPHY for chapter 5
[1] Charnes, A., Kirby, M., and Raik, W., 'Chance-Constrained Games with Partially Controlalble Strategies', Operation Res. 16 (1968), 142-149. [2] Fromovitz, S., C. A Class of One-Period Inventory Models , Operation Res. 1 3 (1965), 779-799. [3] Iosfescu, M. and Theodorescu, R., 'Linear Programming under Uncertainty', Colloquim on Appl. of Math to Economics, Budapest 1965, pp. 133-139. [4] Iosifescu, M. and Theodorescu, R., 'Sur la programmation lineaire', Comptes Rendus madaires des Seabces de I'Academie des Sciences 2 5 6 , No. 253 (1963). [5] Judin, D. B., 'Multi-Stage planning Prlblems under Risk and Uncertainty', Tech. No. 6 (1972). [6] Judin, D. B., Mathematical Radio.
Methods of Management
under Uncertainty,
Hebdo-
Cybernetics,
Moscow 1974, Soviet
[7] Kolbin, V. V., 'Stochastic Programming', Theory of Probability and Mathematical Theoretical Cybernetics. Results of the Science, VINITI, Moscow, 1968, pp. 5-68.
Statistics.
[8] Kolbin, V. V., 'Stochastic Programming' D.Reidel Oubl. Company, 1976, 196 p. [9] Kolbin, V. V., and Tanskaya, V. N., 'Some Stochastic Linear Programming Problems and Algorithm for the Solution', Economic Processes Models, Moscow University Press. [10] Kolbin, V. V., 'Investigation of Absolute Stability of Planning for the Solution of Stochastic Programming Problems', Proceeding of the 2-nd All-Union Symposium Problems of Systemotechnique, Vol.2 (1972), pp. 96-102. [11] Lebedev, V. N., 'Convex-Concave problems of Stochastic Programming under Uncertainty', Rep. U.S.S.R. Acad. Sci. Technolog. Cybernetics, No. 1 (1969), pp. 9-15. [12] Theodorescu, R., 'Minimax Solutions of Random Convex Programs', RebdicontiXLVI 689-692.
(1969),
[13] Zackova, J., 'On Minimax Solutions of Stochastic Linear Programming Problems', Casopis pro pestovam matem. 9 1 (1966). 423-429.
BIBLIOGRAPHY
298 BIBLIOGRAPHY for chapter 6 [1] Hanson, M. A., 'Stochastic Nonlinear Programming', J. Austral. 353.
Math.
Sec. 4 (1964), 347-
[2] Kolbin, V. V., Stochastic Programming D. Reidel Publ. Company, 1976, 196 p. [3] Kolbin, V. V., 'Stochastic Programming', Theory of Probability and Mathematical Theoretical Cybernetics. Results of the Science, VINITI, Moscow, 1968, pp. 5-68.
Statistics.
[4] Kolbin, V. V., 'On the Possibility of the Optimal Decision Application to the Chance-Const rained Mathematical Models', Systems of Managment Economic and Organizational Problems, Leningrad Engeneering and Economic Institute, Vol. 107 (1975). [5] Kolbin, V. V., 'Investigation of the Conditions of the Optimal Decisions Existence in the Stochastic Non-Linear Programming Problem', Problem of Economics, Organization, Infor mation and Mathematics in the System of Management, Leningrad Engineering and Economic Institute, Vol.105 (1975). [6] Kolbin, V. V., 'On a Stochastic Programming Problem', Management vostok, U.S.S.R. Academy of Sciences, Vol. 16 (1975).
an Information,
Vladi
[7] Stoddart, QA. W . I . 'An Existence Theorem for Optimal Stochastic Programming'. J. Austral Math. Soc. 8 (1968), 114-118. [8] Tintner, G., 'On Some Theorems of Stochastic Linear Programming with Applications', Man agement Sci. 10 (1963), 143-159.
BIBLIOGRAPHY
299 BIBLIOGRAPHY for chapter 7
[1] Arbuzova, N. I. and Danilov, V. L., 'On a Problem pf Stochastic Linear Programming and its Stability', Rep. U.S.S.R. Acad. Set. 162 No. 1 (1965). [2] Arbuzova, N. I., 'Correlation of Stochastic instability in Linear Planning Dual Problems', Econ. Math. Methods 2 (1966), 558-562. [3] Arbuzova, N. I., 'Correlation of Stochastic ^-stability on Linear and Some Special Linear Fracrional Programming Problems', Econ. Math. Methods 4 (1968), 108-110. [4] Babbar, M. M., 'Distributions of Solutions of a Set Linear Equations with Application to Linear Programming', J. Am. Statist. Assoc. 50 (1955), 854-869. [5] Barnett, S., 'Stability of the Solution to a Linear Programming Problem', Operat. Res. 13 (1962), 219-228.
Quart.,
[6] Bereanu, B., 'On Stochastic Linear Programming: I. Distribution Problems: a Single Random Variable', Rev. Math. Pures Appl. 8 (1963), 683-697. [7] Bereanu, B., 'On Stochastic Linear Programming. The Laplace Transform of the Distribution of the O p t i m u m and Application', J. Math. Anal. Appl. 15 (1966), 280-294. [8] Bereanu, B., 'Regions de decision et repartitions de 1'optimum dans la programmation lineare,' Compt. Rend. Seans. Acad. Sa Paris 2 5 9 (1964), 1383-1386. [9] Courtillot, M., 'On Varying All the Parameters in a Linear Programming Problem and Sequential Solution of a Linear Programming Problem', Opeartional Res. 10, N o . 4 ( 1 9 6 2 ) . [10] Kolbin, V. V., 'Stochastic Programming', Theory of Probability and Mathematical Theoretical Cybernetics. Results of the Science, VINITI, Moscow, 1968, pp. 5-68.
Statistics.
[11] Kolbin, V. V., 'Stochastic Programming' D. Reidel Publ. Company, 1976, 196 p. [12] Kolbin, V. V., investigation of the Conditions of the Optimal Decisions Existence in the Stochastic Non-Linear Programming Problem', Problem of Economics, Organization, Information and Mathematics in the System of Management, Leningrad Engineering and Economic Institute, Vol.105 (1975). [13] Kolbin, V. V., 'On a Stochastic Programming Problem', Management vostok, U.S.S.R. Academy of Sciences, Vol. 16 (1975).
an Information,
Vladi-
[14] Kolbin, V. V., 'Some Problems of Planning and Functional Stability of the Stochastic Programming Problems Solutions', Management and Information, Vladivostok, U.S.S.R. Academy of Sciences. Vol. 16 (1975). [15] Kolbin, V. V., investigation of Absolute Planning Stability of the Stochastic programming Problems Solution', Management and Information, Vladivostok, U.S.S.R. Academy of Sciences, Vol. 16 (1975) [16] Kolbin, V. V., 'Planning and functional Solutions stability at the i-th Constraint in the Stochastic Programming Problem', Management and Information, Vladivostok, U.S.S.R. Academy of Sciences, Vol. 16 (1975) [17] Sengupta, J. K., Tintner, G., and Millham, C , 'On Some Theorems in Stochastic Linear Programming with Applications' Management Set 10 (1963), 143-159.
300
BIBLIOGRAPHY
[18] Shetty, C. M., 'A Solution to the Transportation Problem with Non-Linear Costs', Res. 7 (1959), 571-580.
Operation
[19] Tintner, G., 'On Some Theorems of Stochastic Linear Programming with Applications', Man agement Sci. 10 (1963), 143-159. [20] Tintner, G., 'On Some Theorems of Stochastic Linear Programming with Applications', Man agement Sci. 10 (1963), 143-159. [21] Vajda, S., 'Stochastic Linear Programming', Ch. II, in Mathematical Wesley, Reading Mass, London 1961, pp. 206-217.
Programming,
Addison-
301
BIBLIOGRAPHY BIBLIOGRAPHY for chapter 8
[1] Anderson, E. J., 'A Review of Duality Theory for Linear Programming over Topological Vector Spaces.' J. Math. Anal, and A p p l , 1983, N 2, pp. 280-392. [2] Anderson, E. J., Lewis A . S . 'An Extension of the Simplex Algorithm for Semi-infinite Linear Programming'. M a t h . Prog., 1989, 44A, N 3, pp. 247-269. [3] Anderson, E. J., Philpot A. Infinite Programming. Berlin. Heildelberg, N.Y.: Springer, 1984 [4] Arrow, K. J., Gurwitz L., Uzawa H. Investigations of Linear and Nonlinear programming, M.: IL, 1962, p. 334 [5] Asic, M . D . , Kovacevic-Vujcic, V. V. An Interior semi-infinite Programming Method. J Opt. Theor. and Appl., 1988, 59, N 3, pp. 353-367. [6] Astafiev, N . N . Infinite-diminsional Programming Problem (Approximation, Duality Gap, Fi nite Models) 27. Int. Wiss. Koll. Ilmenau, 1982. Hf. 5, s. 23-24. [7] Astafiev, N. N. Infinite Lp Problems with a Duality Gap. USSR Ac. of S c , 1984. 275, N 5., pp.1033-1036. [8] Astafiev, N . N . Approximation of Semiproper Infinite-dimensional Lp Problems. Approxima tion Methods for Semiproper MP Problems. Sverdlovsk, 1984, pp. 26-33. [9] Astafiev, N . N . To Regularization of semi-infinite-dimensional LP Problems. Koll. Ilmenau, 1985. Hf., s. 3-6.
30 Int. Wiss.
[10] Astafiev, N . N . Quasifinite Infinite-dimensional LP Problems. Parametric Optimization and Approximation Methods for Improper MP Problems. Sverdlovsk, 1985, pp. 54-62. [11] Bakhovets, E. B., Tsitritsky O. E. On the Lagrange Method in Vector Optimization Problems. Comp. and Appl. Math. Kiev, 1984, N 54, pp. 82-86. [12] Beale, E. M. Quadratic Programming NRLQ, 1959, 6, N 3. [13] Bell, Bradley. Global Convergence of a Semi-infinite Optimization Method. Appl. Math, and Opt. 1990, 21, N 1, pp. 69-88. [14] Ben-Tal, A., Teboulle M., Zowe, I. SecondOorder Necessary Opnmality Conditions for Semimfinite Programming Problems. Semi-infinite Programming, Berlin, Heidelberg, New York: Springer, 1980, pp. 17-30. [15] Blankenship, J . W . , Falik, J . E . Infinitely Constrained Optimization Problems. JOTA, 1976, 19, pp. 261-281. [16] Blatt, H. P. Characterization of Strong Unicity of semi-infinite Optimization by Chain of Ref erences. P a r a m e t r . Optimiz. and Approxim., pp. 301-318. [17] Bobilev, N. A. On the Gradient Method in Infinite-demensionsl Optimization Problems. Au tomat, and Telemech., 1985, N 12, pp. 34-42. [18] Boltiansky, V. G. Separanility of Cones in Linear Topological Space. Dynamics of Nonhomogeneous Systems. M., 1984, p p . 68-76. [19] Boltiansky, V. G. Marquee Theory as a Means of Solving Extremal Problems. M.: VNII of Systems research. 1985, p.& 6. [20] Boltiansky, V. G. Marquee Method in Topological Vector Spaces. USSR Ac. of S c , 1986, 289, N 5, pp.1036-1039.
302
BIBLIOGRAPHY
[21] Boltiabsky, V. G. Marquee Method in Systems Research. Appl. Prob. of Macrosystem Control. M., 1986, pp. 36-43. [22] Boltiansky, V. G. Marquees in Topological Vector Spaces. VNII of Systems Research. 1986, N 19, pp.4-12. [23] Borwein, J. M. Direct Theorems in Semi-infinite Convex Programming. Math. Program., 1981, 21, N 3, pp. 301-318. [23] Borwein, J.M. Semi-infinite Programming Duality: how special is it? Semi-infinite Program ming and Applications. New York: Springer, 1983, pp. 10-36. [24] Buie, R. N., Abragam, J. Postoptimality Sensitivity Analysis in Abstract Spaces with Appli cations to Contimuous-time Programming Problems. JOTA, 1985, 45, N 3, pp. 347-373. [25] Burev, D. D. Analysis and Synthesis for Adaptive Control Systems in Semi-infinitive Nonlinear Programming Problems. Techn. Misl. 1983, N 6, pp. 7-13; 1984, N 1, pp. 15-20. [26] Burev, D.D. New Semi-infinitive Optimization Problems. Statement and Algorithms. Techn. Misl. 1985, 22, N 2, pp. 15-20. [27] Burev, D. D. Upward Convergence to a Solution for a Unique Nonlinear Semi-infinitive Pro gramming Problem. Techn. Misl, 1985, 22, N 4. pp. 15-18. [28] Charnes, A., Cooper, W.W., Kortanek, K.O. Duality, Haar Preograms and Finite Sequence Spaces. Proc. Nat. Acad. Sci. USA, 1962, 48, N 5, pp. 783-786. [29] Charnes, A., Cooper, W.W., Kortanek, K.O. Duality in Semi-infinite Programs and Some Works of Haar and Caratheodory, Management Science, 1963, 9, pp. 209-228. [30] Charnes, A., Cooper, W. W., Kortanek, K.O. On Representations of Semi-infinite Programs which have no Duality Gaps. Management Science. 1965, 12, pp. 113-121. [31] Charnes, A., Gribik, P. R., Kortanek, K. O. Separably-infinite Programming ZOR, 1980, A 24, N 1, pp.33-45. [32] Conn, A. R., Gould, N. I. M. An Exact Penalty Function for Semi-infinite Programming. Math. Program., 1987, 37, N 1, pp. 19-40. [33] Coope, I D . , Watson, G.A. A Projected Langrangian Algorithm for Semi-inginite Program ming. Math. Program., 1985, 32, N 3, pp. 337-356. [34] Curtian, R. F., Pritchard, A. J. Infinite-dimensional Linear Systems Theory. Berlin: Springer, 1978. [35] Dantzig, G.B. Programming in a linear Structure. Washington: Comptroller, 1948. [36] Darkhovsky, B. S., Levitin E.S. necessary and Sufficient Auadratic Optimality Conditions for Semi-infinite MP Problems. Proc. Moscow math. Soc, 1985, N 48, pp. 163-210. [37] Duffin, R. D. Infinite Programs. Linear Inequalities and Associated Problems. M.: IL, 1959, pp.263-276. [38] Duffin, R. J., Karlovitz, L.A. An Infinite Linear Protam with a Duality Gap. Management Science, 1965, 12, pp. 122-134. [39] Duffin, R. J., Jeroslav, R.G., Karlovitz, L.A. Duality in Semi-infinite Linear Programming. Semi-infinite Programming and Applications. Berlin, Heidelberg, New York: Springer, 1983, pp.50-62.
303
BIBLIOGRAPHY
[40] Elzinga, J., Moore, T.G. A Central Gutting Plane Algorithm for the Convex Protramming Problem. Math. Program. 1975, V. 8, pp. 134-145. [41] Endelmann, B. Modifizirte Lagrangefunktionale und Dualitat fur nichtkonvexe Optimierungsprobleme in abstrakten Raumen. Wiss. Z. TH Leipzig, 1983, 7, N 6, s. 365-374. [42] Faibusovich, L. E. The Wolfe Algorithm for a Class of Infinite Programming Problems. USSR Ac. of So,, Tech. Cybernetics, 1982, N 3, pp. 25-35. [43] Faibusovich, L. E. Coordinateless Representation of Simplex Method Kybernetika. Kiev, 1983, N 2, pp. 78-81. [44] Faibusovich, L. E. The Dual Simplex Method: an infinite-dimensional case. Kybernetika. Kiev, 1984, N 1, pp. 81-84. [45] Ferris, M.C., Philpott, A.B. An Interior Point Algorithm for Semi-infinite Linear Program ming. Math. Program., 1989, 43A, N 3, pp. 257-276. [46] Fiaccl, A. V., Kortanek, K.O. Semi-infinite Progr amming and Applications. Berlin, Heidel berg, New York: Springer, 1983, p. 322 [47] Fiacco, A., McKormic, G. Nonlinear Programming. M.: Mir, 1972. [48] Frank, M., Wolfe, P. An Algorithm for Quadratic Programming. NRLQ, 1956, 3, N 1-2. [49] Georgibiani, D A . , Zhuzhinashvili, A.Sh. Stability in Infinite-dimensional MP Problems. 30. Int. Wiss. Koll. Ilmenau, 1985. Hf. 4, s. 55-56. [50] Gerstewitz, Ch. Dualitat bei nichtkonvexen Vektoroptimierungsproblemen in Math., 1983, N 50/1, s. 75-81. [51] Gerstewitz, Ch., Iwaniw, E. Dualitat fur nichtconvexe Vektoroptimierungs probleme. Wiss. Z. TH Ilmenau, 1985, 31, N 2, s. 61-81. [52] Gfrerer, H., Guldat, J., Wacker, Hi. and Zulehner W. Globalization of Locally convergent Al gorithms for Nonlinear Optimization Problems with Constraints. Semi-infinite Programming and Applications. Berlin, Heidelberg, New York: Springer, 1983, pp. 128-137. [53] Glashoff, K. Duality Theory of Semi-infinite Programming. Semi-infinite Programming. Berlin, Heidelberg, New York: Sptinger, 1979, pp. 1-16. [54] Glashoff, K., Gustafson, S.A. Linear Optimization and Approximation. Berlin, Heidelberg, New York: Springer, 1983. [55] Goberna, M.A., Jornet, V. Geometric Fundamentals of the Simplex Method in Semi-infinite Programming. OR. Spectrum, 1988, N 3, pp. 145-152. [56] Goberna, M.A., Lopez, M. A. Reduction and Discrete Approximation in Linear Semi-infinite Programming. Optimization, 1987, 18, N 5, pp. 643-658. [57] Goberna, M.A., Lopez, M.A. The Optimal Value Function in Semi-Infinite programming. JOTA, 1988, N 3. [58] Goberna, M. A., Pastor, J. Conditiones suficientes para la existencia de solucion optimaen un programe semi-infinito. Trab. estatist. y invest, oper., 1983, 34, p. 3-20, N 1. [59] Golden, R. Zur Stanilitat separabel-infiniter linearer Optimierungsprobleme. Berlin: Springer, 1970, pp. 66-82.
Oper. Res.,
[60] Golstein, E.G. Dual Convex Programming Problems. Econ. and Math. Methods. 1965, 1, N3.
304
BIBLIOGRAPHY
[61] Golstein, E . G . Dual Problems of Convex and Convex-fractional Programming in Function Spaces. Math. Prof. Investigations. M.: Nauka, 1968, pp. 10-108. [62] Gribik, P. R. A Central-cutting-plane Algorithm for SIP Problems. Semi-infinite Programming. Berlin: Springer, 1979, pp. 66-82. [63] Gurwitz, L. Programming in Linear Spaces. Investigations in Linear and Nonlinear Program ming. M.: IL, 1962, pp. 65-115. [64] Gustafson, S. A. On the Computational Solution of a Class of Generalized Moment Problems. SIAM J. Numer. Anal., 1970, 7, pp. 343-357. [65] Gustafson, S.A., Kortanek, K . O . Numerical Solution of a Class of Semi-uinfinite Problems. NRLQ, 1973, V. 20, pp. 477-504. [66] Gustafson, S. A. A General Three-phase Algorithm for Nonlinear Semi-infinite Programming. Operational Research. Amsterdam, 1981, pp. 495-508. [67] Gustafson, S.A., Kortanek, K . O . Semi-infinite Programming and Applications. Math. Pro gram., Berlin, 1983, pp. 132-157. [68] Gustafson, S. A. Investigating Semi-infinite Programs Using Penalty Functions and Lagrangian Methods. J. Austral. Math. S o c , 1986, B 28, N 2, pp. 158-159. [69] Gwinner, J. Results of Farkas type Numer. Funct. Anal, and Optim. 1987, 9, N 5-6, p p . 4 7 1 520. [70] Hettich, R., Gramlich, G. A Note on the Implementation of a Method for Quadratic Semiinfinite Programming. Math. Prog., 1990, 46, N 2, pp. 249-254. [71] Hettich, R., Jongen, H. Th. Semi-infinite Programming: Conditions of Optimality and Appli cations. Optimization Techniques, part 2. Berlin: Springer, 1978, pp. 1-11. [72] Hettich, R. Semi-infinite Programming. Berlin: Springer, 1979, p. 178. [73] Hettich, R., Honstede, W. On Quadratically Convergent Metnod for Semi-infinite Program ming. Semi-infinite Programming. Berlin, Hedelberg, New York: Springer, 1979, pp. 97-111. [74] Hettich, R. A. Comparison of Some Numerical Methods for Semi-infinite Programming. Semiinfinite Programming. Berlin: Springer, 1979, pp. 112-125. [75] Hettich, R., Zencke, H. Numerische Methoden der Approximation und semi-infinite Oprimierung. Stuttgart: Teubner, 1982. [76] Hettich, R. A Review of Numerical Methods for Semi-infinite Optimization. Programming and Applications. Berlin: Springer, 1983, pp. 158-178.
Semi-infinite
[77] Hettich, R. An Implementation of a Discretization Method for Semi-infinite Programming. Math. Program., 1986, 34, pp. 354-361. [78] Himelblau, D. Applied Nonlinear Programming. M.: Mir, 1975, p. 534. [79] Hoffman, K . H . , Klostermeir, A. A. Semi-infinite Linear Programming Procedure and Applica tion to Approximation Problems in Optimal Control. Approximation Theory. II, New York, San Francisco, LondonL Grow-Hill, 1976, pp. 379-389. [80] Honstede, W. An Approximation Method for Semi-infinite Problems Semi-infinite Program ming. Berlin, Heidelberg, New York: Springer, 1979, pp. 195-216. [81] Hurwiez, L. The Minkowski-Farkas. Lemma for Bounded Linear Transformations in Banach Spaces. Cowles Commiss.
^ S ^ k e W k M h t
1
6
BIBLIOGRAPHY
305
[82] Ibieiugba, M. A. The Role of the Multipliers in the Multiplier Method. JOYA, 1985, 47, N 2, pp. 195-216. [83] Itskovich, I. A. Infinite-dimensional Economic and Mathematical Models. Math. Analysis of Regional Production System Models. Novosibirsk, 1984, pp. 119-131. [84] Jahn, J. Dualitat der Vektoroptimierung mit Anwendung in der vektoriellen Approximation. 27 Int., Wiss. Roll., Ilmenau, 1982, N 5, s. 7-9. [85] Duality in Vector Optimization. Math. Program., 1983, 25, N 3, pp. 343-353. [86] Jeakumar, V. Nonlinear Alternative Theorems and Nondifferentianble Programming. ZOR, 1984, A28, N 5, pp. 175-187. [87] Jeroslow, R. G. Uniform Duality in Semi-infinite Convex Optimization. Math. Program., 1983, 27, N 2, pp. 144-154. [88] Joffe, I. D., Tikhomirov, B. M. Theory of Extremal Problems. M.: Nauka, 1974, p. 192. [89] Jongen, H.Th., Wetterling, W., Zwier, G. On Sufficient Conditions for Local Optimality in Semi-infinite Programming. Optimization, 1987, 18, N 2, pp. 165-178. [90] Kantorovich, L.V. Mathematical Methods of Production Organization an Planning. L.: LGU Press, 1939. [91] Kaplan, A., Tikhachke, R. Adaptive Solution Methods for Ill-defined Semi-infinite- dimansional Convex Optimization Problems. USSR. Ac. of Sc, 1992, 322, N 3, pp. 460-464. [92] Karney, D. F. A Duality Theorem for Semi-infinite Convex Programs and Their Finite Sub programs. Math. Prog., 1983, 27, pp. 75-82. [93] Karney, D. F. Clark's Theorem for Semi-infinite Convex Programs. Advances in Appl. Math., 1981, 2, pp. 7-12. [94] Karney, D. F. Duality Gaps in Semi-infinite Linear Programming — and Approximation Prob lem. Math. Program., 1981, 20. N 2, pp. 129-143. [95] Karney, D.F., In a Semi-infinite Program only a Countable Subset of the Constraints is Es sential. J. Approxim. Theory, 1985, 44, N 1, pp. 69-72. [96] Kawasaki, H. An envelope-like effect of Infinitely Many Inequality Constraints on Secondorder Necessary Confitions for Minimization Problems. Math. Programming, 1988, 41, N 1, pp.73-96. [97] Kelley, J. E. The Gutting Plane Method for Solving Convex Programs. J.SIAM, 1960, 8, 4. [98] Kenderov, P S . Most of the Optimization Problems have a Unique Solution. Parametr. Optimiz. and Approxim., 1985, pp. 203-216. [99] Klee, V. A. Note on Convex Cones and Constraint Qualifications in Infinite-dimensional Vector Spaces. JOTA, 1982, V 37, N 2, pp. 277-284. [100] Kolbin, V. V. Investigation of Infinite-dimensional Programming Problems. Problems of Me chanics' and Control Processes. L.: LGU Press. 1984, N 7. pp. 184-189. [101] Kolbin, V. V. Investigation of Infinite-dimensional Programming Problems "Problems of In formatics", Ukrainian Ac. ofSc, V l.Kiev, 1990, pp. 116-121. [102] Kolbin, V. V. Infinite-dimensional stochastic Programming Problems. "Problems of Informat ics", Ukrainian Ac. of Sc, VOO^glWdPM^tMMI
306
BIBLIOGRAPHY
[103] Kornstaed, H. I. Necessary Conditions of Higher Order for Semi-infinite Programming. Semiinfinite Programming. Berlin: Springer, 979, pp. 31-50. [104] Kortanek, K. 0., Stroiwas, H. M. On Extreme Points of Bounded Sets of Generalized Finite Sequence Spaces. ZOR, 1983, A27, N 5. pp. 145-147. [105] Kretschmer, K. S. Programmes in Paired Spaces. Canad. J. Math., 1961, 13, N 2, pp. 221-230. [106] Kuhn, H. W., Tucker, A W . Nonlinear Programming. Proc. Second Berkley Symp. Math. Statist, and Probability. Berkley and Los Angeles: Univ. California Press, 1951, pp. 481-492. [107] Laban, M.M. On Some Numerical Properties of Infinite-dimensional Symplex. Numer. Meth. and Approxim. Theory Ni s: 1984, pp. 7-12. [108] Levin, V. L. Extremum Conditions in Infinite-dimensional Linear Problems with Operator Constraints. Math. Program. Investigation. M.: Nauka, 1986, pp. 159-197. [109] Lopez, M.A., Vercher, E. Optimality Conditions for Nondifferentiable Convex Semi-infinite Programming. Math. Program., 1983, pp. 307-319. [110] Nebeling, V. Uber eine schnittmethode zur Losung konvexer quadratischer semi-infiniter Optimierungsaufgaben 30 Int. Wiss. Koll. Ilmenau, 1985, N 4, s. 91-93. [Ill] Niirnberger, G. Unicity in Semi-infinite Optimization. Parametr. Optimiz, and Approxim., 1985, pp.231-247. [112] Niirnberger, G. Global Unicity in Semi-infinite Optimization. Numer. Funct. and Optim., 1985, 8, N 1-2, pp. 173-191. [113] Pinsker, A.G. A Linear Programming Problem in Compact Metric Spaces. Optimization. Novosibirsk, 1985, N 35, (52), pp. 24-27. [114] Polak, E., Tits, A. L. A Recursive Quadratic Convergent Algorithm for Semi-infinite Problems. Semi-infinite Programming and Applications. Berlin: Springer, 1983. [115] Polak, E., Mayne, D. Q., Stimler, D. M. Control System Design via Semi-infinite Optimization: a Review. Proc. IEEE, 1984, 72, N 12, pp. 1777-1794. [116] Polak, E., Wardy, Y. Y. A Study of Minimization Sequences. SIAM J. Contr. and Optim., 1984, N 4, pp. 599-609. [117] Price, C. J., Coope, I. D. An Exact Penalty Function Algorithm for Semi-infinite Programmes. BIT, 1990, 30, N 4, pp. 723-734. [118] Roleff, K. A Stable Multiple exchange Algorithm for Linear SIP. Semi-infinitive Programming. Berlin: Springer, 1979, pp. 83-96. [119] Rosen, J. The Gradient Projection Method for Nonlinear Programming. J. SIAM, 1960, 8, pp.181-217. [120] Rubinstein, G. Sh. Duality in Nathenatical Programming and Some Issues of Convex Analysis. Progress in Math. Sc. 1970, V 25, N 5, pp. 171-198. [121] Shapiro, A. SecondOorder Derivatives of extremal-value functions and Optimality Conditions for Semi-infinite Programs. Math. Oper. Res., 1985, 10, N 2, pp. 207-219. [122] Stolp, D. Zum von Tucker bei linearen Optimalproblemen in halbgeordneten Vektor aumen Optimization, 1984, 15, N 4, s. 531-544. [123] Sturm, N. Semi-infinite Optimierung. Fachverlagfur Revisions — und treugandfragen Schawarzenbek, 1978.
BIBLIOGRAPHY
307
[124] Thamelt, W. Uber Losbarkeitseigenschaften homogener linearer Ungleichungen in halbgeordneten Raumen Optimization, 1976, 7, s. 281-294. [125] Thamelt W. Zum Satz von Tucker in halbgeordneten Vektorraumen Wiss. Z. TH Leuna-Merseburg, 1980, 22, N 1, s. 107-114. [126] Thamelt, W., Stolp, D. Einige Bemerkungen zum Satz von Tucker Wiss. Z. TH Liena-Merseburg, 1981, 23, N 2, s. 294-302. [127] Thamelt, W. Infinite-dimensional LP Problems, Duality, Approximation, and their Relation to the Theory of Games. L.r LQU Press, 1966. [128] Thamelt, W. Duality of Infinite-dimensional LP Problems and some Related Issues. Kybernetika. Kiev, 1969, N 4, pp. 80-87. [129] Tichatschke, R., Lohse, Th. Eine verallgemeinerte Schnittmethode fur konvexe semi-infinite Optimierungsaufgaben Wiss. Z. TH Karl-Marx-Stadt, 1982, 24, N 3, s. 332-338. [130] Tichatschke, R. Uber Schnittmethoden fur konvexe semi-infinite Optimierungsaufgaben seminarber. Humboldt Univ., Berlin. Sek. Math., 1983, N 50/2, s. 339-350. [131] Tichatschke, R., Nebeling V. A Cutting-plane Method for Quadratic Semi-infinite Program ming Problems. Optimization., 1988, A 19, N 6, s. 803-817. [132] Timokhin, S. G., Shapkin, A. V. Multiple Specification of Matrices of Conditions for LP Prob lems. VINITI, 1980 (N 2609-80-DEP). [133] Timokhin, S.G., Shapkin, A.V. On the LP Problem with Inadequately Defined Data. Eco nomic and Mathematical Methods, 1981, V 17, N 5, pp. 955-963. [134] Todorov, M.I. Generic Exastence and uniqueness of the Solution to Linear Semi-infinite Op timization problems. Proc, Bulgarian Ac. of Sc, 1985, 38, N 8, pp. 989-991. [135] Todorov, M.I. Generic Existence of the Solution to Linear Semi-infinite Optimization Prob lems. Numer. Funct. Anal, and Optim., 8, N 5-6. pp. 541-556. [136] Todorov, M. I. Generic uniqueness of Saddle Points in the Linear Semi-infinite Optimization. Proc. Bulgarian Ac. of Sc, 1986, 39, N 4, pp. 27-29. [137] Tran, Q. Ch. Duality in Vector Optimization. Pt. 1. Abstract Duality Scheme. Kybernetika, 1984, 20, N 4, pp. 304-313. [138] Tran, Q. Ch. Duality in Vector Optimization. Pt. 2. Vector Quasiconcave Programming. Kybernetika. 1984, 20, N 5, pp. 386-404. [139] Tran, Q. Ch. Duality and Optimality Conditions in Abstract Concave Maximization. Kyber netika, 1985, 21, N 2, pp. 108-117. [140] Tretiakov, A. A. The Necessary Kuhn-Tucker Conditions for Optimization in Hilbert Space. Collected Works of MGU Computation Center, 1983, N 39, pp. 190-196. [141] Troktzch, F. On Lagrange Multipoier Rules and Their Applications to Nonlinear Control Problems in Function Spaces. Seminarber. Humboldt Univ. Berlin Sekt, Math., 1984, N 64, s. 121-129. [142] Uzawa, H. The Kuhn-Tucker Theorem of Convex Programming. M.: IL, 1962, pp. 57-64. [143] Valyi, I. Approximate solutions of Vector Optimization Problem. Math. Res., 1985, 27, n 1, pp.246-250.
308
BIBLIOGRAPHY
[144] Watson, G. A. Globally Convergent Methods for Semi-infinite Programming. sknaft for Informationsbehandling (BIT), 1981, 21. p p . 3 6 - 3 7 3 .
Nordisk Tid-
[145] Watson, G . A . Numerical Experiments with Globally Convergent Methods for Semi-infinite Programming Problems. Semi-infinite Programming and Applications, 1983, pp. 190-205. [146] Weber, R. Entscheidungsprobleme bei Unsicherheit und mehrfacher Zielsetzung Math. Syst. Econ., 1982, N 80, VIII, s. 172. 147] Weber, R. Entscheidungsprobleme bei Unsicherheit und mehrfacher Zielsetzung Ein Absatz mit Hilfe der semi-infiniten linearen Vektoroptimierung Konigstein: Verl. Anton Hoim, Meisenhiem, 1982, VIII, s. 172. 148] Weber, R. Decision Making under Uncertainty: a Semi-infinite programming Approachl Eur. J. Oper. Res., 1985, 19, N 1, pp. 104-113. 149] Wolfe, Ph. A Duality Theorem for Nonlinear Programming. 150] Yeremin, 1.1., Astafiev, N. N. Introduction to the Theory of Linear and Convex programming. M.: Nauka, 1976, p. 170. 151] Yeremin, 1.1., Vatolin, A. A. Duality for Improper Infinite-dimensional Problems of Linear and Convex Programming. Approximation Methods for improper MP Problems. Sverdlovsk, 1984, pp. 3-20. 152] Yeremin, 1.1., Vatolin, A. A. Duality for Improper MP Problems. Mech. Institute, 1985, p. 50.
Sverdlovsk: Math,
and
153] Zaboieva, O.A., Thimokhin, S.G., Shapkin, A . V . On the Search of Guaranteed Optimal Solutions to Large-scale SIP Problems. J. Comp. Math, and Math. Physics. 1985, 25, N 1, pp. 45-52. 154] Zontendeik, G. Feasible Direction Methods. M.: IL, 1963. 155] Zukhovitsky, S. I., Poliak, R. A., Primak, N. E. The Solution Algorithm for Convex Program ming Problems. Proc. of USSR Ac. of S c , 1963, 153, N 5.
309
BIBLIOGRAPHY BIBLIOGRAPHY for chapters 9 and 10
[1] Automated Management Systems for Municipal Economy. Kiev, Budivelnik, 1978, p. 143. [2] Andronov, A. A., Pontriagin, L.S. "Structurally Stable" Systems. USSR Ac. of Sc. Reports, 1937, V.14, 5, pp. 521-523. [3] Buslenko, N. P., Kovalenko, I. N. Lectures on the Theory of Complex Systems. Soviet Radio, 1973, p. 352. [4] Dantzig, G.B. Linear Programming and Extensions. Princeton University Press. 1963, pp. 625. [5] Gill, F., Murrey, W. Numerical Methods for Optimization Conditions. M.: Mir, 1977, p. 290. [6] Himmelblau, D. Applied Nonlinear Programming. M.: Mir, 1976, p. 526. [7] Ivakhnenko, A. G. Heuristic Self-organization systems in Technique Cybernetics. Kiev, Teknika 1971, p. 165. g8 Judin, D.B., Golstein, E.G. Linear Programming. Theory and Finite Methods. M., Fizmatgiz, 1963, p. 524. [9] Karr, W., Houv, W. Quantitative Methods of Decision Making in Economy and Management. M., Mir, 1966, p. 127. [10] Kolbin, V.V. Stochastic Programming. D. Reidel. Publ. Cs., 1976. [11] Likhtenstein, V. E. Discrete Programming Models. M., Nauka, 1971, p. 240. [12] Minaev, Y.N. Stability of Optimal Solutions in Convex MP Problems. Automated Plant Management Systems. Kiev, IC of Ukranian Ac. of Sc, 1977, pp. 73-84. [13] Minaev, Y. N., Vilinskaia, V. M., Iliashenko, V. A. Relianility of Optimal solutions in automat ed Management Systems for Construction and Building. M., CBNTI of the USSR Minpromstroy, 1977, p. 90. [14] Minaev, Y. N. (Ed.) Stability of Economic and Mathematical Models of Optimization. Moscow, Statistika, 1980. [15] Mikhalevich, V.S. Computational Methods for Selection of Optimal Design Works. Kiev, Naukova Dumka, 1977, p. 132. [16] Orlovsky, S.A. Inadequately defined Games. Journal of Computational Mathematics and Mathematical Physics. 1976, V. 16, N 6, pp. 1427-1435. [17] Rastrigin, L. A., Ripa, K.K. Theory of Automatic Random Search. Riga, Zinatne, 1973, p. 342. [18] Sergienko, I. V. The Method of Vector of Descent and Its Application to the Solution of Com binatorial Extremal Problems. Control Systems and Machines. 1976, N. 3, pp. 27-31. [19] Smekhov, B.M., Urinson, Y. M. Optimization Methods for the National Economic Plan. M., Ekonomika, 1976, p. 127. [20] Tabak, T., Kuo, W. Mathematical Programming and Optimal Control. M., Nauka, 1977, p.215. [21] Tikhonov, A.N., Arsenin, A. A. methods of Solving Ill-defined Problems. M., Nauka, 1977, p. 192. [22] Zade. L, Linguistic Variables and their Applications to Approximate Decision Making. M. 1976, p. 165.
BIBLIOGRAPHY
310 BIBLIOGRAPHY for chapter 11 [1] Kantorovich, L. V., Akilov, G. P. Functional Analysis. M., Nauka, 1977.
[2] Kantorovich, L.V., Wulikh, V.Z., Pinsker, A.G. Functional Analysis in Partially Ordered Spaces. M.-L., Gostekhizdat, 1950. [3] Kolbin, V. V. Compactum-continuous Function Spaces. "Control Problems" Transactions. St. Petersburg University Press, 1992, pp. 86-99. [4] Kolbin, V. V. The Generalized Parametric Programming Problem. "Control Problems" Trans actions. St. Petersburg University Press, 1992, pp. 99-108. [5] Lozanovsky, G. Y. On the Functions of Linear Structure Elements. Mathematical Transactions in Higher Schools, 1973, N 1, pp. 45-54. [6] Luxemburg, W. A.J., Zaanen, A.C. Riesz Spaces. VI, Amsterdam, 1971. [7] Nakano, H., Linear Lattices. Detroit, 1966. [8] Wulikh, V. Z. A Bundle in Linear Partially Ordered Spaces and Its Application to the Opera tions Thery. Mathematical Transactions, 22 (64), 1948, N 1, 27-78.